Jan 14 13:48:25.571647 kernel: Linux version 6.12.65-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.1_p20250801 p4) 14.3.1 20250801, GNU ld (Gentoo 2.45 p3) 2.45.0) #1 SMP PREEMPT_DYNAMIC Wed Jan 14 11:12:50 -00 2026 Jan 14 13:48:25.571682 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=166c426371167f765dd2026937f2932948c99d0fb4a3868a9b09e1eb4ef3a9c9 Jan 14 13:48:25.571694 kernel: BIOS-provided physical RAM map: Jan 14 13:48:25.571707 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Jan 14 13:48:25.571715 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Jan 14 13:48:25.571726 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Jan 14 13:48:25.571739 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Jan 14 13:48:25.571748 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Jan 14 13:48:25.571756 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable Jan 14 13:48:25.571765 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Jan 14 13:48:25.571773 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable Jan 14 13:48:25.571913 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Jan 14 13:48:25.571923 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable Jan 14 13:48:25.571932 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Jan 14 13:48:25.571946 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Jan 14 13:48:25.571957 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Jan 14 13:48:25.571970 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce90fff] usable Jan 14 13:48:25.571979 kernel: BIOS-e820: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Jan 14 13:48:25.571988 kernel: BIOS-e820: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Jan 14 13:48:25.571997 kernel: BIOS-e820: [mem 0x000000009ce97000-0x000000009cedbfff] usable Jan 14 13:48:25.572219 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Jan 14 13:48:25.572233 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Jan 14 13:48:25.572242 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Jan 14 13:48:25.572251 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Jan 14 13:48:25.572260 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Jan 14 13:48:25.572269 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Jan 14 13:48:25.572286 kernel: NX (Execute Disable) protection: active Jan 14 13:48:25.572297 kernel: APIC: Static calls initialized Jan 14 13:48:25.572306 kernel: e820: update [mem 0x9b320018-0x9b329c57] usable ==> usable Jan 14 13:48:25.572316 kernel: e820: update [mem 0x9b2e3018-0x9b31fe57] usable ==> usable Jan 14 13:48:25.572325 kernel: extended physical RAM map: Jan 14 13:48:25.572334 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Jan 14 13:48:25.572346 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable Jan 14 13:48:25.572357 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Jan 14 13:48:25.572366 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable Jan 14 13:48:25.572376 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Jan 14 13:48:25.572385 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable Jan 14 13:48:25.572399 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Jan 14 13:48:25.572412 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b2e3017] usable Jan 14 13:48:25.572421 kernel: reserve setup_data: [mem 0x000000009b2e3018-0x000000009b31fe57] usable Jan 14 13:48:25.572436 kernel: reserve setup_data: [mem 0x000000009b31fe58-0x000000009b320017] usable Jan 14 13:48:25.572448 kernel: reserve setup_data: [mem 0x000000009b320018-0x000000009b329c57] usable Jan 14 13:48:25.572460 kernel: reserve setup_data: [mem 0x000000009b329c58-0x000000009bd3efff] usable Jan 14 13:48:25.572473 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Jan 14 13:48:25.572484 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable Jan 14 13:48:25.572493 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Jan 14 13:48:25.572503 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Jan 14 13:48:25.572512 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Jan 14 13:48:25.572524 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce90fff] usable Jan 14 13:48:25.572536 kernel: reserve setup_data: [mem 0x000000009ce91000-0x000000009ce94fff] reserved Jan 14 13:48:25.572658 kernel: reserve setup_data: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS Jan 14 13:48:25.572669 kernel: reserve setup_data: [mem 0x000000009ce97000-0x000000009cedbfff] usable Jan 14 13:48:25.572680 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Jan 14 13:48:25.572691 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Jan 14 13:48:25.572701 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Jan 14 13:48:25.572711 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Jan 14 13:48:25.572722 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Jan 14 13:48:25.572732 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Jan 14 13:48:25.572742 kernel: efi: EFI v2.7 by EDK II Jan 14 13:48:25.572753 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b9e4198 RNG=0x9cb73018 Jan 14 13:48:25.572763 kernel: random: crng init done Jan 14 13:48:25.572905 kernel: efi: Remove mem151: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Jan 14 13:48:25.572917 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Jan 14 13:48:25.572928 kernel: secureboot: Secure boot disabled Jan 14 13:48:25.572938 kernel: SMBIOS 2.8 present. Jan 14 13:48:25.572949 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Jan 14 13:48:25.572959 kernel: DMI: Memory slots populated: 1/1 Jan 14 13:48:25.572970 kernel: Hypervisor detected: KVM Jan 14 13:48:25.572980 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Jan 14 13:48:25.572990 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Jan 14 13:48:25.573001 kernel: kvm-clock: using sched offset of 12786755806 cycles Jan 14 13:48:25.573221 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Jan 14 13:48:25.573337 kernel: tsc: Detected 2445.426 MHz processor Jan 14 13:48:25.573348 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jan 14 13:48:25.573359 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jan 14 13:48:25.573370 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Jan 14 13:48:25.573381 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Jan 14 13:48:25.573392 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jan 14 13:48:25.573402 kernel: Using GB pages for direct mapping Jan 14 13:48:25.573415 kernel: ACPI: Early table checksum verification disabled Jan 14 13:48:25.573426 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Jan 14 13:48:25.573437 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Jan 14 13:48:25.573447 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Jan 14 13:48:25.573458 kernel: ACPI: DSDT 0x000000009CB7A000 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 14 13:48:25.573469 kernel: ACPI: FACS 0x000000009CBDD000 000040 Jan 14 13:48:25.573479 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 14 13:48:25.573493 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 14 13:48:25.573506 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 14 13:48:25.573518 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 14 13:48:25.573529 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Jan 14 13:48:25.573539 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Jan 14 13:48:25.573548 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1b9] Jan 14 13:48:25.573558 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Jan 14 13:48:25.573575 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Jan 14 13:48:25.573588 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Jan 14 13:48:25.573598 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Jan 14 13:48:25.573607 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Jan 14 13:48:25.573617 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Jan 14 13:48:25.573627 kernel: No NUMA configuration found Jan 14 13:48:25.573640 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] Jan 14 13:48:25.573654 kernel: NODE_DATA(0) allocated [mem 0x9ce36dc0-0x9ce3dfff] Jan 14 13:48:25.573668 kernel: Zone ranges: Jan 14 13:48:25.573679 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jan 14 13:48:25.573688 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] Jan 14 13:48:25.573698 kernel: Normal empty Jan 14 13:48:25.573711 kernel: Device empty Jan 14 13:48:25.573724 kernel: Movable zone start for each node Jan 14 13:48:25.573734 kernel: Early memory node ranges Jan 14 13:48:25.573743 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Jan 14 13:48:25.573757 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Jan 14 13:48:25.573768 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Jan 14 13:48:25.573900 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] Jan 14 13:48:25.573912 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] Jan 14 13:48:25.573919 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] Jan 14 13:48:25.573926 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce90fff] Jan 14 13:48:25.573933 kernel: node 0: [mem 0x000000009ce97000-0x000000009cedbfff] Jan 14 13:48:25.573944 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] Jan 14 13:48:25.573952 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jan 14 13:48:25.573973 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Jan 14 13:48:25.573989 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Jan 14 13:48:25.574000 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jan 14 13:48:25.574236 kernel: On node 0, zone DMA: 239 pages in unavailable ranges Jan 14 13:48:25.574252 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Jan 14 13:48:25.574266 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Jan 14 13:48:25.574277 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Jan 14 13:48:25.574288 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges Jan 14 13:48:25.574303 kernel: ACPI: PM-Timer IO Port: 0x608 Jan 14 13:48:25.574314 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Jan 14 13:48:25.574327 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Jan 14 13:48:25.574340 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Jan 14 13:48:25.574356 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Jan 14 13:48:25.574366 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Jan 14 13:48:25.574376 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Jan 14 13:48:25.574390 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Jan 14 13:48:25.574403 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jan 14 13:48:25.574414 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Jan 14 13:48:25.574424 kernel: TSC deadline timer available Jan 14 13:48:25.574439 kernel: CPU topo: Max. logical packages: 1 Jan 14 13:48:25.574452 kernel: CPU topo: Max. logical dies: 1 Jan 14 13:48:25.574465 kernel: CPU topo: Max. dies per package: 1 Jan 14 13:48:25.574475 kernel: CPU topo: Max. threads per core: 1 Jan 14 13:48:25.574485 kernel: CPU topo: Num. cores per package: 4 Jan 14 13:48:25.574496 kernel: CPU topo: Num. threads per package: 4 Jan 14 13:48:25.574508 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Jan 14 13:48:25.574521 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Jan 14 13:48:25.574535 kernel: kvm-guest: KVM setup pv remote TLB flush Jan 14 13:48:25.574546 kernel: kvm-guest: setup PV sched yield Jan 14 13:48:25.574556 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Jan 14 13:48:25.574570 kernel: Booting paravirtualized kernel on KVM Jan 14 13:48:25.574582 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jan 14 13:48:25.574593 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Jan 14 13:48:25.574604 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Jan 14 13:48:25.574619 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Jan 14 13:48:25.574634 kernel: pcpu-alloc: [0] 0 1 2 3 Jan 14 13:48:25.574645 kernel: kvm-guest: PV spinlocks enabled Jan 14 13:48:25.574656 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Jan 14 13:48:25.574668 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=166c426371167f765dd2026937f2932948c99d0fb4a3868a9b09e1eb4ef3a9c9 Jan 14 13:48:25.574679 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jan 14 13:48:25.574692 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 14 13:48:25.574709 kernel: Fallback order for Node 0: 0 Jan 14 13:48:25.574719 kernel: Built 1 zonelists, mobility grouping on. Total pages: 641450 Jan 14 13:48:25.574729 kernel: Policy zone: DMA32 Jan 14 13:48:25.574740 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 14 13:48:25.574753 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Jan 14 13:48:25.574766 kernel: ftrace: allocating 40128 entries in 157 pages Jan 14 13:48:25.574900 kernel: ftrace: allocated 157 pages with 5 groups Jan 14 13:48:25.574919 kernel: Dynamic Preempt: voluntary Jan 14 13:48:25.574929 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 14 13:48:25.574940 kernel: rcu: RCU event tracing is enabled. Jan 14 13:48:25.574953 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Jan 14 13:48:25.574967 kernel: Trampoline variant of Tasks RCU enabled. Jan 14 13:48:25.574978 kernel: Rude variant of Tasks RCU enabled. Jan 14 13:48:25.574988 kernel: Tracing variant of Tasks RCU enabled. Jan 14 13:48:25.575003 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 14 13:48:25.575265 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Jan 14 13:48:25.575277 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 14 13:48:25.575288 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 14 13:48:25.575298 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 14 13:48:25.575310 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Jan 14 13:48:25.575322 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jan 14 13:48:25.575339 kernel: Console: colour dummy device 80x25 Jan 14 13:48:25.575352 kernel: printk: legacy console [ttyS0] enabled Jan 14 13:48:25.575359 kernel: ACPI: Core revision 20240827 Jan 14 13:48:25.575367 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Jan 14 13:48:25.575374 kernel: APIC: Switch to symmetric I/O mode setup Jan 14 13:48:25.575381 kernel: x2apic enabled Jan 14 13:48:25.575389 kernel: APIC: Switched APIC routing to: physical x2apic Jan 14 13:48:25.575396 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Jan 14 13:48:25.575406 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Jan 14 13:48:25.575413 kernel: kvm-guest: setup PV IPIs Jan 14 13:48:25.575421 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jan 14 13:48:25.575428 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x233fd7ba1b0, max_idle_ns: 440795295779 ns Jan 14 13:48:25.575436 kernel: Calibrating delay loop (skipped) preset value.. 4890.85 BogoMIPS (lpj=2445426) Jan 14 13:48:25.575443 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Jan 14 13:48:25.575451 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Jan 14 13:48:25.575460 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Jan 14 13:48:25.575467 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jan 14 13:48:25.575475 kernel: Spectre V2 : Mitigation: Retpolines Jan 14 13:48:25.575482 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Jan 14 13:48:25.575490 kernel: Speculative Store Bypass: Vulnerable Jan 14 13:48:25.575497 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Jan 14 13:48:25.575505 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Jan 14 13:48:25.575515 kernel: active return thunk: srso_alias_return_thunk Jan 14 13:48:25.575522 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Jan 14 13:48:25.575530 kernel: Transient Scheduler Attacks: Forcing mitigation on in a VM Jan 14 13:48:25.575537 kernel: Transient Scheduler Attacks: Vulnerable: Clear CPU buffers attempted, no microcode Jan 14 13:48:25.575545 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jan 14 13:48:25.575552 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jan 14 13:48:25.575561 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jan 14 13:48:25.575569 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jan 14 13:48:25.575576 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Jan 14 13:48:25.575584 kernel: Freeing SMP alternatives memory: 32K Jan 14 13:48:25.575591 kernel: pid_max: default: 32768 minimum: 301 Jan 14 13:48:25.575598 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Jan 14 13:48:25.575605 kernel: landlock: Up and running. Jan 14 13:48:25.575615 kernel: SELinux: Initializing. Jan 14 13:48:25.575622 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 14 13:48:25.575630 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 14 13:48:25.575637 kernel: smpboot: CPU0: AMD EPYC 7763 64-Core Processor (family: 0x19, model: 0x1, stepping: 0x1) Jan 14 13:48:25.575644 kernel: Performance Events: PMU not available due to virtualization, using software events only. Jan 14 13:48:25.575652 kernel: signal: max sigframe size: 1776 Jan 14 13:48:25.575659 kernel: rcu: Hierarchical SRCU implementation. Jan 14 13:48:25.575669 kernel: rcu: Max phase no-delay instances is 400. Jan 14 13:48:25.575676 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Jan 14 13:48:25.575683 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Jan 14 13:48:25.575691 kernel: smp: Bringing up secondary CPUs ... Jan 14 13:48:25.575698 kernel: smpboot: x86: Booting SMP configuration: Jan 14 13:48:25.575705 kernel: .... node #0, CPUs: #1 #2 #3 Jan 14 13:48:25.575713 kernel: smp: Brought up 1 node, 4 CPUs Jan 14 13:48:25.575720 kernel: smpboot: Total of 4 processors activated (19563.40 BogoMIPS) Jan 14 13:48:25.575730 kernel: Memory: 2439048K/2565800K available (14336K kernel code, 2445K rwdata, 31644K rodata, 15536K init, 2500K bss, 120812K reserved, 0K cma-reserved) Jan 14 13:48:25.575737 kernel: devtmpfs: initialized Jan 14 13:48:25.575745 kernel: x86/mm: Memory block size: 128MB Jan 14 13:48:25.575752 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Jan 14 13:48:25.575760 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Jan 14 13:48:25.575767 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) Jan 14 13:48:25.575774 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Jan 14 13:48:25.575903 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce95000-0x9ce96fff] (8192 bytes) Jan 14 13:48:25.575911 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Jan 14 13:48:25.575918 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 14 13:48:25.575926 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Jan 14 13:48:25.575933 kernel: pinctrl core: initialized pinctrl subsystem Jan 14 13:48:25.575940 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 14 13:48:25.575948 kernel: audit: initializing netlink subsys (disabled) Jan 14 13:48:25.575958 kernel: audit: type=2000 audit(1768398491.166:1): state=initialized audit_enabled=0 res=1 Jan 14 13:48:25.575965 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 14 13:48:25.575973 kernel: thermal_sys: Registered thermal governor 'user_space' Jan 14 13:48:25.575980 kernel: cpuidle: using governor menu Jan 14 13:48:25.575988 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 14 13:48:25.575995 kernel: dca service started, version 1.12.1 Jan 14 13:48:25.576002 kernel: PCI: ECAM [mem 0xe0000000-0xefffffff] (base 0xe0000000) for domain 0000 [bus 00-ff] Jan 14 13:48:25.576261 kernel: PCI: Using configuration type 1 for base access Jan 14 13:48:25.576269 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jan 14 13:48:25.576277 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jan 14 13:48:25.576287 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jan 14 13:48:25.576301 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 14 13:48:25.576314 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jan 14 13:48:25.576325 kernel: ACPI: Added _OSI(Module Device) Jan 14 13:48:25.576340 kernel: ACPI: Added _OSI(Processor Device) Jan 14 13:48:25.576351 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 14 13:48:25.576362 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 14 13:48:25.576377 kernel: ACPI: Interpreter enabled Jan 14 13:48:25.576389 kernel: ACPI: PM: (supports S0 S3 S5) Jan 14 13:48:25.576400 kernel: ACPI: Using IOAPIC for interrupt routing Jan 14 13:48:25.576410 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jan 14 13:48:25.598343 kernel: PCI: Using E820 reservations for host bridge windows Jan 14 13:48:25.598363 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Jan 14 13:48:25.598378 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jan 14 13:48:25.600489 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jan 14 13:48:25.600682 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Jan 14 13:48:25.601006 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Jan 14 13:48:25.601238 kernel: PCI host bridge to bus 0000:00 Jan 14 13:48:25.601562 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jan 14 13:48:25.601733 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Jan 14 13:48:25.602251 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jan 14 13:48:25.602448 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Jan 14 13:48:25.602610 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Jan 14 13:48:25.602914 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Jan 14 13:48:25.603319 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jan 14 13:48:25.605319 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Jan 14 13:48:25.605744 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Jan 14 13:48:25.606239 kernel: pci 0000:00:01.0: BAR 0 [mem 0xc0000000-0xc0ffffff pref] Jan 14 13:48:25.606421 kernel: pci 0000:00:01.0: BAR 2 [mem 0xc1044000-0xc1044fff] Jan 14 13:48:25.606589 kernel: pci 0000:00:01.0: ROM [mem 0xffff0000-0xffffffff pref] Jan 14 13:48:25.606909 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jan 14 13:48:25.607334 kernel: pci 0000:00:01.0: pci_fixup_video+0x0/0x100 took 17578 usecs Jan 14 13:48:25.607650 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Jan 14 13:48:25.607979 kernel: pci 0000:00:02.0: BAR 0 [io 0x6100-0x611f] Jan 14 13:48:25.608411 kernel: pci 0000:00:02.0: BAR 1 [mem 0xc1043000-0xc1043fff] Jan 14 13:48:25.608598 kernel: pci 0000:00:02.0: BAR 4 [mem 0x380000000000-0x380000003fff 64bit pref] Jan 14 13:48:25.608937 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Jan 14 13:48:25.609357 kernel: pci 0000:00:03.0: BAR 0 [io 0x6000-0x607f] Jan 14 13:48:25.609565 kernel: pci 0000:00:03.0: BAR 1 [mem 0xc1042000-0xc1042fff] Jan 14 13:48:25.609746 kernel: pci 0000:00:03.0: BAR 4 [mem 0x380000004000-0x380000007fff 64bit pref] Jan 14 13:48:25.610514 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Jan 14 13:48:25.610692 kernel: pci 0000:00:04.0: BAR 0 [io 0x60e0-0x60ff] Jan 14 13:48:25.611007 kernel: pci 0000:00:04.0: BAR 1 [mem 0xc1041000-0xc1041fff] Jan 14 13:48:25.611417 kernel: pci 0000:00:04.0: BAR 4 [mem 0x380000008000-0x38000000bfff 64bit pref] Jan 14 13:48:25.611587 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref] Jan 14 13:48:25.611898 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Jan 14 13:48:25.612357 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Jan 14 13:48:25.612579 kernel: pci 0000:00:1f.0: quirk_ich7_lpc+0x0/0xc0 took 19531 usecs Jan 14 13:48:25.613381 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Jan 14 13:48:25.613609 kernel: pci 0000:00:1f.2: BAR 4 [io 0x60c0-0x60df] Jan 14 13:48:25.613971 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xc1040000-0xc1040fff] Jan 14 13:48:25.614435 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Jan 14 13:48:25.614648 kernel: pci 0000:00:1f.3: BAR 4 [io 0x6080-0x60bf] Jan 14 13:48:25.614670 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Jan 14 13:48:25.614685 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Jan 14 13:48:25.614695 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jan 14 13:48:25.614703 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Jan 14 13:48:25.614717 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Jan 14 13:48:25.614725 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Jan 14 13:48:25.614732 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Jan 14 13:48:25.614740 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Jan 14 13:48:25.614748 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Jan 14 13:48:25.614756 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Jan 14 13:48:25.614764 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Jan 14 13:48:25.614774 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Jan 14 13:48:25.614923 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Jan 14 13:48:25.614938 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Jan 14 13:48:25.614951 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Jan 14 13:48:25.614960 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Jan 14 13:48:25.614968 kernel: iommu: Default domain type: Translated Jan 14 13:48:25.614976 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jan 14 13:48:25.614987 kernel: efivars: Registered efivars operations Jan 14 13:48:25.614996 kernel: PCI: Using ACPI for IRQ routing Jan 14 13:48:25.615004 kernel: PCI: pci_cache_line_size set to 64 bytes Jan 14 13:48:25.615231 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Jan 14 13:48:25.615246 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] Jan 14 13:48:25.615260 kernel: e820: reserve RAM buffer [mem 0x9b2e3018-0x9bffffff] Jan 14 13:48:25.615268 kernel: e820: reserve RAM buffer [mem 0x9b320018-0x9bffffff] Jan 14 13:48:25.615280 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] Jan 14 13:48:25.615289 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] Jan 14 13:48:25.615296 kernel: e820: reserve RAM buffer [mem 0x9ce91000-0x9fffffff] Jan 14 13:48:25.615304 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] Jan 14 13:48:25.615529 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Jan 14 13:48:25.615721 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Jan 14 13:48:25.616317 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jan 14 13:48:25.616346 kernel: vgaarb: loaded Jan 14 13:48:25.616355 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Jan 14 13:48:25.616363 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Jan 14 13:48:25.616371 kernel: clocksource: Switched to clocksource kvm-clock Jan 14 13:48:25.616379 kernel: VFS: Disk quotas dquot_6.6.0 Jan 14 13:48:25.616387 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 14 13:48:25.616401 kernel: pnp: PnP ACPI init Jan 14 13:48:25.616774 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Jan 14 13:48:25.616921 kernel: pnp: PnP ACPI: found 6 devices Jan 14 13:48:25.616936 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jan 14 13:48:25.616950 kernel: NET: Registered PF_INET protocol family Jan 14 13:48:25.616959 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jan 14 13:48:25.616967 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jan 14 13:48:25.617259 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 14 13:48:25.617280 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 14 13:48:25.617289 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jan 14 13:48:25.617297 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jan 14 13:48:25.617305 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 14 13:48:25.617313 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 14 13:48:25.617323 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 14 13:48:25.617342 kernel: NET: Registered PF_XDP protocol family Jan 14 13:48:25.617578 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref]: can't claim; no compatible bridge window Jan 14 13:48:25.617932 kernel: pci 0000:00:04.0: ROM [mem 0x9d000000-0x9d03ffff pref]: assigned Jan 14 13:48:25.618499 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Jan 14 13:48:25.618707 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Jan 14 13:48:25.619269 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Jan 14 13:48:25.619473 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Jan 14 13:48:25.619914 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Jan 14 13:48:25.620374 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Jan 14 13:48:25.620389 kernel: PCI: CLS 0 bytes, default 64 Jan 14 13:48:25.620398 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x233fd7ba1b0, max_idle_ns: 440795295779 ns Jan 14 13:48:25.620410 kernel: Initialise system trusted keyrings Jan 14 13:48:25.620424 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jan 14 13:48:25.620444 kernel: Key type asymmetric registered Jan 14 13:48:25.620457 kernel: Asymmetric key parser 'x509' registered Jan 14 13:48:25.620465 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Jan 14 13:48:25.620473 kernel: io scheduler mq-deadline registered Jan 14 13:48:25.620481 kernel: io scheduler kyber registered Jan 14 13:48:25.620489 kernel: io scheduler bfq registered Jan 14 13:48:25.620503 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jan 14 13:48:25.620521 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Jan 14 13:48:25.620537 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Jan 14 13:48:25.620550 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Jan 14 13:48:25.620562 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 14 13:48:25.620574 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jan 14 13:48:25.620593 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Jan 14 13:48:25.620607 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jan 14 13:48:25.620621 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jan 14 13:48:25.620984 kernel: rtc_cmos 00:04: RTC can wake from S4 Jan 14 13:48:25.621222 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jan 14 13:48:25.621446 kernel: rtc_cmos 00:04: registered as rtc0 Jan 14 13:48:25.621636 kernel: rtc_cmos 00:04: setting system clock to 2026-01-14T13:48:19 UTC (1768398499) Jan 14 13:48:25.621958 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Jan 14 13:48:25.621972 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Jan 14 13:48:25.621980 kernel: efifb: probing for efifb Jan 14 13:48:25.621988 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Jan 14 13:48:25.621996 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Jan 14 13:48:25.622004 kernel: efifb: scrolling: redraw Jan 14 13:48:25.622250 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Jan 14 13:48:25.622261 kernel: Console: switching to colour frame buffer device 160x50 Jan 14 13:48:25.622269 kernel: fb0: EFI VGA frame buffer device Jan 14 13:48:25.622280 kernel: pstore: Using crash dump compression: deflate Jan 14 13:48:25.622288 kernel: pstore: Registered efi_pstore as persistent store backend Jan 14 13:48:25.622296 kernel: NET: Registered PF_INET6 protocol family Jan 14 13:48:25.622303 kernel: Segment Routing with IPv6 Jan 14 13:48:25.622313 kernel: In-situ OAM (IOAM) with IPv6 Jan 14 13:48:25.622321 kernel: NET: Registered PF_PACKET protocol family Jan 14 13:48:25.622329 kernel: Key type dns_resolver registered Jan 14 13:48:25.622337 kernel: IPI shorthand broadcast: enabled Jan 14 13:48:25.622345 kernel: sched_clock: Marking stable (6678093514, 2866517869)->(10645077574, -1100466191) Jan 14 13:48:25.622353 kernel: registered taskstats version 1 Jan 14 13:48:25.622361 kernel: Loading compiled-in X.509 certificates Jan 14 13:48:25.622368 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.65-flatcar: e8d0aa6f955c6f54d5fb15cad90d0ea8c698688e' Jan 14 13:48:25.622378 kernel: Demotion targets for Node 0: null Jan 14 13:48:25.622386 kernel: Key type .fscrypt registered Jan 14 13:48:25.622393 kernel: Key type fscrypt-provisioning registered Jan 14 13:48:25.622401 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 14 13:48:25.622409 kernel: ima: Allocated hash algorithm: sha1 Jan 14 13:48:25.622417 kernel: ima: No architecture policies found Jan 14 13:48:25.622427 kernel: clk: Disabling unused clocks Jan 14 13:48:25.622435 kernel: Freeing unused kernel image (initmem) memory: 15536K Jan 14 13:48:25.622443 kernel: Write protecting the kernel read-only data: 47104k Jan 14 13:48:25.622451 kernel: Freeing unused kernel image (rodata/data gap) memory: 1124K Jan 14 13:48:25.622459 kernel: Run /init as init process Jan 14 13:48:25.622467 kernel: with arguments: Jan 14 13:48:25.622474 kernel: /init Jan 14 13:48:25.622482 kernel: with environment: Jan 14 13:48:25.622604 kernel: HOME=/ Jan 14 13:48:25.622612 kernel: TERM=linux Jan 14 13:48:25.622620 kernel: SCSI subsystem initialized Jan 14 13:48:25.622628 kernel: libata version 3.00 loaded. Jan 14 13:48:25.622956 kernel: ahci 0000:00:1f.2: version 3.0 Jan 14 13:48:25.622971 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Jan 14 13:48:25.623352 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Jan 14 13:48:25.623925 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Jan 14 13:48:25.624332 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Jan 14 13:48:25.624960 kernel: scsi host0: ahci Jan 14 13:48:25.625656 kernel: scsi host1: ahci Jan 14 13:48:25.626480 kernel: scsi host2: ahci Jan 14 13:48:25.627417 kernel: scsi host3: ahci Jan 14 13:48:25.627760 kernel: scsi host4: ahci Jan 14 13:48:25.628482 kernel: scsi host5: ahci Jan 14 13:48:25.628504 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 26 lpm-pol 1 Jan 14 13:48:25.628520 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 26 lpm-pol 1 Jan 14 13:48:25.628531 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 26 lpm-pol 1 Jan 14 13:48:25.628667 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 26 lpm-pol 1 Jan 14 13:48:25.628681 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 26 lpm-pol 1 Jan 14 13:48:25.628692 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 26 lpm-pol 1 Jan 14 13:48:25.628703 kernel: ata1: SATA link down (SStatus 0 SControl 300) Jan 14 13:48:25.628717 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Jan 14 13:48:25.628730 kernel: ata2: SATA link down (SStatus 0 SControl 300) Jan 14 13:48:25.628742 kernel: ata6: SATA link down (SStatus 0 SControl 300) Jan 14 13:48:25.628985 kernel: ata5: SATA link down (SStatus 0 SControl 300) Jan 14 13:48:25.628998 kernel: ata4: SATA link down (SStatus 0 SControl 300) Jan 14 13:48:25.629248 kernel: ata3.00: LPM support broken, forcing max_power Jan 14 13:48:25.629269 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Jan 14 13:48:25.629281 kernel: ata3.00: applying bridge limits Jan 14 13:48:25.629290 kernel: ata3.00: LPM support broken, forcing max_power Jan 14 13:48:25.629299 kernel: ata3.00: configured for UDMA/100 Jan 14 13:48:25.629582 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Jan 14 13:48:25.630703 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Jan 14 13:48:25.644928 kernel: virtio_blk virtio1: [vda] 27000832 512-byte logical blocks (13.8 GB/12.9 GiB) Jan 14 13:48:25.675392 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Jan 14 13:48:25.675443 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jan 14 13:48:25.675461 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Jan 14 13:48:25.675612 kernel: GPT:16515071 != 27000831 Jan 14 13:48:25.675622 kernel: GPT:Alternate GPT header not at the end of the disk. Jan 14 13:48:25.675631 kernel: GPT:16515071 != 27000831 Jan 14 13:48:25.675640 kernel: GPT: Use GNU Parted to correct GPT errors. Jan 14 13:48:25.675649 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 14 13:48:25.675996 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Jan 14 13:48:25.676257 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 14 13:48:25.676422 kernel: device-mapper: uevent: version 1.0.3 Jan 14 13:48:25.676436 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Jan 14 13:48:25.676448 kernel: device-mapper: verity: sha256 using shash "sha256-generic" Jan 14 13:48:25.676464 kernel: raid6: avx2x4 gen() 23277 MB/s Jan 14 13:48:25.676475 kernel: raid6: avx2x2 gen() 16964 MB/s Jan 14 13:48:25.676487 kernel: raid6: avx2x1 gen() 12856 MB/s Jan 14 13:48:25.676498 kernel: raid6: using algorithm avx2x4 gen() 23277 MB/s Jan 14 13:48:25.676629 kernel: raid6: .... xor() 3063 MB/s, rmw enabled Jan 14 13:48:25.676645 kernel: raid6: using avx2x2 recovery algorithm Jan 14 13:48:25.676657 kernel: xor: automatically using best checksumming function avx Jan 14 13:48:25.676668 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 14 13:48:25.676680 kernel: BTRFS: device fsid a2d7d9b8-1cc4-4aa6-91f7-011fd4658df9 devid 1 transid 34 /dev/mapper/usr (253:0) scanned by mount (181) Jan 14 13:48:25.676692 kernel: BTRFS info (device dm-0): first mount of filesystem a2d7d9b8-1cc4-4aa6-91f7-011fd4658df9 Jan 14 13:48:25.676704 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jan 14 13:48:25.676944 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 14 13:48:25.676957 kernel: BTRFS info (device dm-0): enabling free space tree Jan 14 13:48:25.676968 kernel: loop: module loaded Jan 14 13:48:25.676979 kernel: loop0: detected capacity change from 0 to 100536 Jan 14 13:48:25.676991 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jan 14 13:48:25.677004 systemd[1]: Successfully made /usr/ read-only. Jan 14 13:48:25.677238 systemd[1]: systemd 257.9 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jan 14 13:48:25.677361 systemd[1]: Detected virtualization kvm. Jan 14 13:48:25.677373 systemd[1]: Detected architecture x86-64. Jan 14 13:48:25.677385 systemd[1]: Running in initrd. Jan 14 13:48:25.677397 systemd[1]: No hostname configured, using default hostname. Jan 14 13:48:25.677409 systemd[1]: Hostname set to . Jan 14 13:48:25.677424 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Jan 14 13:48:25.677542 systemd[1]: Queued start job for default target initrd.target. Jan 14 13:48:25.677554 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Jan 14 13:48:25.677566 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 14 13:48:25.677577 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 14 13:48:25.677590 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 14 13:48:25.677605 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 14 13:48:25.677970 kernel: hrtimer: interrupt took 4122775 ns Jan 14 13:48:25.677986 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 14 13:48:25.677998 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 14 13:48:25.678230 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 14 13:48:25.678247 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 14 13:48:25.678260 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Jan 14 13:48:25.678387 systemd[1]: Reached target paths.target - Path Units. Jan 14 13:48:25.678400 systemd[1]: Reached target slices.target - Slice Units. Jan 14 13:48:25.678413 systemd[1]: Reached target swap.target - Swaps. Jan 14 13:48:25.678425 systemd[1]: Reached target timers.target - Timer Units. Jan 14 13:48:25.678438 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 14 13:48:25.678452 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 14 13:48:25.678463 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Jan 14 13:48:25.678593 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 14 13:48:25.678605 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Jan 14 13:48:25.678616 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 14 13:48:25.678630 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 14 13:48:25.678644 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 14 13:48:25.678656 systemd[1]: Reached target sockets.target - Socket Units. Jan 14 13:48:25.678667 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jan 14 13:48:25.678909 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 14 13:48:25.678925 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 14 13:48:25.678939 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 14 13:48:25.678952 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Jan 14 13:48:25.678964 systemd[1]: Starting systemd-fsck-usr.service... Jan 14 13:48:25.678976 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 14 13:48:25.679315 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 14 13:48:25.679329 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 14 13:48:25.679341 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 14 13:48:25.679353 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 14 13:48:25.679533 systemd-journald[320]: Collecting audit messages is enabled. Jan 14 13:48:25.679562 systemd[1]: Finished systemd-fsck-usr.service. Jan 14 13:48:25.679576 kernel: audit: type=1130 audit(1768398505.592:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:25.679697 kernel: audit: type=1130 audit(1768398505.643:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:25.679709 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 14 13:48:25.679723 systemd-journald[320]: Journal started Jan 14 13:48:25.679960 systemd-journald[320]: Runtime Journal (/run/log/journal/3f664cc37dea4d3e875d093b3e52a6f3) is 6M, max 48M, 42M free. Jan 14 13:48:25.592000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:25.643000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:25.714598 systemd[1]: Started systemd-journald.service - Journal Service. Jan 14 13:48:25.725000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:25.754581 kernel: audit: type=1130 audit(1768398505.725:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:25.837264 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 14 13:48:25.886952 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 14 13:48:25.925629 kernel: Bridge firewalling registered Jan 14 13:48:25.929534 systemd-modules-load[323]: Inserted module 'br_netfilter' Jan 14 13:48:26.103246 kernel: audit: type=1130 audit(1768398506.022:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:26.022000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:25.998968 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 14 13:48:26.041595 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 14 13:48:26.132731 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 14 13:48:26.207718 kernel: audit: type=1130 audit(1768398506.163:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:26.163000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:26.216761 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 14 13:48:26.285908 kernel: audit: type=1130 audit(1768398506.234:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:26.234000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:26.250900 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 14 13:48:26.292386 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 14 13:48:26.360620 systemd-tmpfiles[333]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Jan 14 13:48:26.389971 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 14 13:48:26.390000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:26.422233 kernel: audit: type=1130 audit(1768398506.390:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:26.447557 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 14 13:48:26.461000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:26.468001 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 14 13:48:26.534231 kernel: audit: type=1130 audit(1768398506.461:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:26.534265 kernel: audit: type=1334 audit(1768398506.463:10): prog-id=6 op=LOAD Jan 14 13:48:26.463000 audit: BPF prog-id=6 op=LOAD Jan 14 13:48:26.548757 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 14 13:48:26.587000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:26.617339 kernel: audit: type=1130 audit(1768398506.587:11): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:26.619468 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 14 13:48:26.658000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:26.671373 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 14 13:48:26.770456 dracut-cmdline[362]: dracut-109 Jan 14 13:48:26.784774 dracut-cmdline[362]: Using kernel command line parameters: SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=166c426371167f765dd2026937f2932948c99d0fb4a3868a9b09e1eb4ef3a9c9 Jan 14 13:48:26.788473 systemd-resolved[353]: Positive Trust Anchors: Jan 14 13:48:26.788484 systemd-resolved[353]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 14 13:48:26.788491 systemd-resolved[353]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Jan 14 13:48:26.788536 systemd-resolved[353]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 14 13:48:27.144399 systemd-resolved[353]: Defaulting to hostname 'linux'. Jan 14 13:48:27.147998 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 14 13:48:27.176000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:27.176613 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 14 13:48:27.857731 kernel: Loading iSCSI transport class v2.0-870. Jan 14 13:48:28.000690 kernel: iscsi: registered transport (tcp) Jan 14 13:48:28.127568 kernel: iscsi: registered transport (qla4xxx) Jan 14 13:48:28.127957 kernel: QLogic iSCSI HBA Driver Jan 14 13:48:28.288248 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 14 13:48:28.351731 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 14 13:48:28.360000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:28.368196 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 14 13:48:28.756561 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 14 13:48:28.781000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:28.785665 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 14 13:48:28.815482 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 14 13:48:28.973773 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 14 13:48:28.989000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:28.993000 audit: BPF prog-id=7 op=LOAD Jan 14 13:48:28.994000 audit: BPF prog-id=8 op=LOAD Jan 14 13:48:28.995786 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 14 13:48:29.126645 systemd-udevd[585]: Using default interface naming scheme 'v257'. Jan 14 13:48:29.167619 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 14 13:48:29.208000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:29.212793 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 14 13:48:29.399689 dracut-pre-trigger[639]: rd.md=0: removing MD RAID activation Jan 14 13:48:29.576263 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 14 13:48:29.594000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:29.597000 audit: BPF prog-id=9 op=LOAD Jan 14 13:48:29.598704 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 14 13:48:29.626604 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 14 13:48:29.655240 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 14 13:48:29.650000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:29.832694 systemd-networkd[727]: lo: Link UP Jan 14 13:48:29.832705 systemd-networkd[727]: lo: Gained carrier Jan 14 13:48:29.854505 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 14 13:48:29.890000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:29.891425 systemd[1]: Reached target network.target - Network. Jan 14 13:48:29.917800 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 14 13:48:29.921000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:29.927524 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 14 13:48:30.093588 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Jan 14 13:48:30.153977 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Jan 14 13:48:30.214812 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Jan 14 13:48:30.419984 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 14 13:48:30.536673 kernel: cryptd: max_cpu_qlen set to 1000 Jan 14 13:48:30.572778 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 14 13:48:30.573574 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 14 13:48:30.679338 kernel: kauditd_printk_skb: 13 callbacks suppressed Jan 14 13:48:30.679516 kernel: audit: type=1131 audit(1768398510.603:25): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:30.603000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:30.603985 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 14 13:48:30.673338 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 14 13:48:30.703352 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 14 13:48:30.750399 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 14 13:48:30.820698 kernel: audit: type=1130 audit(1768398510.768:26): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:30.768000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:30.750669 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 14 13:48:30.901693 kernel: audit: type=1131 audit(1768398510.768:27): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:30.768000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:30.811407 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 14 13:48:30.918213 disk-uuid[772]: Primary Header is updated. Jan 14 13:48:30.918213 disk-uuid[772]: Secondary Entries is updated. Jan 14 13:48:30.918213 disk-uuid[772]: Secondary Header is updated. Jan 14 13:48:31.026361 kernel: AES CTR mode by8 optimization enabled Jan 14 13:48:31.134798 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Jan 14 13:48:31.138953 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 14 13:48:31.195957 kernel: audit: type=1130 audit(1768398511.139:28): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:31.139000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:31.336439 systemd-networkd[727]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 14 13:48:31.336455 systemd-networkd[727]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 14 13:48:31.391216 systemd-networkd[727]: eth0: Link UP Jan 14 13:48:31.498759 kernel: audit: type=1130 audit(1768398511.404:29): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:31.404000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:31.392213 systemd-networkd[727]: eth0: Gained carrier Jan 14 13:48:31.394536 systemd-networkd[727]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 14 13:48:31.396398 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 14 13:48:31.406461 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 14 13:48:31.456293 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 14 13:48:31.469580 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 14 13:48:31.498990 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 14 13:48:31.599738 systemd-networkd[727]: eth0: DHCPv4 address 10.0.0.127/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jan 14 13:48:31.712376 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 14 13:48:31.729000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:31.772359 kernel: audit: type=1130 audit(1768398511.729:30): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:32.058584 disk-uuid[774]: Warning: The kernel is still using the old partition table. Jan 14 13:48:32.058584 disk-uuid[774]: The new table will be used at the next reboot or after you Jan 14 13:48:32.058584 disk-uuid[774]: run partprobe(8) or kpartx(8) Jan 14 13:48:32.058584 disk-uuid[774]: The operation has completed successfully. Jan 14 13:48:32.295961 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 14 13:48:32.299426 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 14 13:48:32.435977 kernel: audit: type=1130 audit(1768398512.346:31): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:32.436739 kernel: audit: type=1131 audit(1768398512.346:32): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:32.346000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:32.346000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:32.356447 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 14 13:48:32.467299 systemd-networkd[727]: eth0: Gained IPv6LL Jan 14 13:48:32.916570 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (870) Jan 14 13:48:32.943449 kernel: BTRFS info (device vda6): first mount of filesystem bc594bac-1fbf-41b0-97ef-4b225e86c0fe Jan 14 13:48:32.943508 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 14 13:48:32.988584 kernel: BTRFS info (device vda6): turning on async discard Jan 14 13:48:32.988666 kernel: BTRFS info (device vda6): enabling free space tree Jan 14 13:48:33.033537 kernel: BTRFS info (device vda6): last unmount of filesystem bc594bac-1fbf-41b0-97ef-4b225e86c0fe Jan 14 13:48:33.050551 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 14 13:48:33.115651 kernel: audit: type=1130 audit(1768398513.069:33): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:33.069000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:33.117499 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 14 13:48:35.570637 ignition[889]: Ignition 2.24.0 Jan 14 13:48:35.570967 ignition[889]: Stage: fetch-offline Jan 14 13:48:35.572468 ignition[889]: no configs at "/usr/lib/ignition/base.d" Jan 14 13:48:35.600491 ignition[889]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 14 13:48:35.616442 ignition[889]: parsed url from cmdline: "" Jan 14 13:48:35.616570 ignition[889]: no config URL provided Jan 14 13:48:35.620517 ignition[889]: reading system config file "/usr/lib/ignition/user.ign" Jan 14 13:48:35.620547 ignition[889]: no config at "/usr/lib/ignition/user.ign" Jan 14 13:48:35.620748 ignition[889]: op(1): [started] loading QEMU firmware config module Jan 14 13:48:35.620756 ignition[889]: op(1): executing: "modprobe" "qemu_fw_cfg" Jan 14 13:48:36.318451 ignition[889]: op(1): [finished] loading QEMU firmware config module Jan 14 13:48:37.813858 ignition[889]: parsing config with SHA512: b2c6a699b7dcaf62b4985ef050fcac4574a0c7ce2a53f4e30d12e2004e6212581257d7fd6fc4464c3375167a16ea6af720e163a1c3680a9812cf7ef18dd29630 Jan 14 13:48:37.983411 unknown[889]: fetched base config from "system" Jan 14 13:48:37.983536 unknown[889]: fetched user config from "qemu" Jan 14 13:48:38.067463 kernel: audit: type=1130 audit(1768398518.015:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:38.015000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:37.985725 ignition[889]: fetch-offline: fetch-offline passed Jan 14 13:48:37.990502 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 14 13:48:37.986282 ignition[889]: Ignition finished successfully Jan 14 13:48:38.017647 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Jan 14 13:48:38.020601 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 14 13:48:38.791655 ignition[900]: Ignition 2.24.0 Jan 14 13:48:38.791795 ignition[900]: Stage: kargs Jan 14 13:48:38.792838 ignition[900]: no configs at "/usr/lib/ignition/base.d" Jan 14 13:48:38.844000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:38.814838 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 14 13:48:38.893390 kernel: audit: type=1130 audit(1768398518.844:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:38.792854 ignition[900]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 14 13:48:38.848761 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 14 13:48:38.799239 ignition[900]: kargs: kargs passed Jan 14 13:48:38.799314 ignition[900]: Ignition finished successfully Jan 14 13:48:39.048358 ignition[908]: Ignition 2.24.0 Jan 14 13:48:39.048373 ignition[908]: Stage: disks Jan 14 13:48:39.048558 ignition[908]: no configs at "/usr/lib/ignition/base.d" Jan 14 13:48:39.048569 ignition[908]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 14 13:48:39.050001 ignition[908]: disks: disks passed Jan 14 13:48:39.050268 ignition[908]: Ignition finished successfully Jan 14 13:48:39.119506 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 14 13:48:39.184466 kernel: audit: type=1130 audit(1768398519.135:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:39.135000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:39.136731 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 14 13:48:39.185594 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 14 13:48:39.212722 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 14 13:48:39.239657 systemd[1]: Reached target sysinit.target - System Initialization. Jan 14 13:48:39.268240 systemd[1]: Reached target basic.target - Basic System. Jan 14 13:48:39.298346 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 14 13:48:39.477739 systemd-fsck[918]: ROOT: clean, 15/456736 files, 38230/456704 blocks Jan 14 13:48:39.492797 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 14 13:48:39.506856 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 14 13:48:39.503000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:39.588859 kernel: audit: type=1130 audit(1768398519.503:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:40.352540 kernel: EXT4-fs (vda9): mounted filesystem 00eaf6ed-0a89-4fef-afb6-3b81d372e1c1 r/w with ordered data mode. Quota mode: none. Jan 14 13:48:40.355531 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 14 13:48:40.383856 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 14 13:48:40.419570 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 14 13:48:40.434821 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 14 13:48:40.455727 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jan 14 13:48:40.455772 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 14 13:48:40.455803 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 14 13:48:40.567808 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 14 13:48:40.585407 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 14 13:48:40.621251 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (927) Jan 14 13:48:40.655630 kernel: BTRFS info (device vda6): first mount of filesystem bc594bac-1fbf-41b0-97ef-4b225e86c0fe Jan 14 13:48:40.655673 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 14 13:48:40.708840 kernel: BTRFS info (device vda6): turning on async discard Jan 14 13:48:40.709005 kernel: BTRFS info (device vda6): enabling free space tree Jan 14 13:48:40.712825 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 14 13:48:41.897835 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 14 13:48:41.984466 kernel: audit: type=1130 audit(1768398521.912:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:41.912000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:41.916575 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 14 13:48:41.959689 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 14 13:48:42.040505 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 14 13:48:42.071616 kernel: BTRFS info (device vda6): last unmount of filesystem bc594bac-1fbf-41b0-97ef-4b225e86c0fe Jan 14 13:48:42.135870 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 14 13:48:42.170000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:42.205357 kernel: audit: type=1130 audit(1768398522.170:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:42.530346 ignition[1026]: INFO : Ignition 2.24.0 Jan 14 13:48:42.530346 ignition[1026]: INFO : Stage: mount Jan 14 13:48:42.555402 ignition[1026]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 14 13:48:42.555402 ignition[1026]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 14 13:48:42.555402 ignition[1026]: INFO : mount: mount passed Jan 14 13:48:42.555402 ignition[1026]: INFO : Ignition finished successfully Jan 14 13:48:42.675243 kernel: audit: type=1130 audit(1768398522.583:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:42.583000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:42.550469 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 14 13:48:42.587266 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 14 13:48:42.706737 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 14 13:48:42.814412 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (1038) Jan 14 13:48:42.842225 kernel: BTRFS info (device vda6): first mount of filesystem bc594bac-1fbf-41b0-97ef-4b225e86c0fe Jan 14 13:48:42.842812 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 14 13:48:42.896724 kernel: BTRFS info (device vda6): turning on async discard Jan 14 13:48:42.898830 kernel: BTRFS info (device vda6): enabling free space tree Jan 14 13:48:42.903321 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 14 13:48:43.342404 ignition[1055]: INFO : Ignition 2.24.0 Jan 14 13:48:43.342404 ignition[1055]: INFO : Stage: files Jan 14 13:48:43.374256 ignition[1055]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 14 13:48:43.374256 ignition[1055]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 14 13:48:43.411326 ignition[1055]: DEBUG : files: compiled without relabeling support, skipping Jan 14 13:48:43.439488 ignition[1055]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 14 13:48:43.471739 ignition[1055]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 14 13:48:43.511432 ignition[1055]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 14 13:48:43.533490 ignition[1055]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 14 13:48:43.554663 ignition[1055]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 14 13:48:43.554663 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Jan 14 13:48:43.554663 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Jan 14 13:48:43.537406 unknown[1055]: wrote ssh authorized keys file for user: core Jan 14 13:48:44.173399 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jan 14 13:48:45.758529 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Jan 14 13:48:45.758529 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jan 14 13:48:45.810933 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jan 14 13:48:45.852501 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jan 14 13:48:45.852501 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jan 14 13:48:45.852501 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 14 13:48:45.852501 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 14 13:48:45.852501 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 14 13:48:45.852501 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 14 13:48:45.852501 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 14 13:48:45.852501 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 14 13:48:45.852501 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.1-x86-64.raw" Jan 14 13:48:45.852501 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.34.1-x86-64.raw" Jan 14 13:48:45.852501 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.1-x86-64.raw" Jan 14 13:48:45.852501 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.34.1-x86-64.raw: attempt #1 Jan 14 13:48:46.416507 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jan 14 13:48:57.328805 ignition[1055]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.34.1-x86-64.raw" Jan 14 13:48:57.328805 ignition[1055]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Jan 14 13:48:57.390878 ignition[1055]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 14 13:48:57.390878 ignition[1055]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 14 13:48:57.390878 ignition[1055]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Jan 14 13:48:57.390878 ignition[1055]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Jan 14 13:48:57.390878 ignition[1055]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 14 13:48:57.390878 ignition[1055]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 14 13:48:57.390878 ignition[1055]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Jan 14 13:48:57.390878 ignition[1055]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Jan 14 13:48:57.817772 ignition[1055]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Jan 14 13:48:59.541863 kernel: clocksource: Long readout interval, skipping watchdog check: cs_nsec: 2130729308 wd_nsec: 2130728258 Jan 14 13:48:59.620863 ignition[1055]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Jan 14 13:48:59.644669 ignition[1055]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Jan 14 13:48:59.644669 ignition[1055]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Jan 14 13:48:59.644669 ignition[1055]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Jan 14 13:48:59.644669 ignition[1055]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 14 13:48:59.644669 ignition[1055]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 14 13:48:59.644669 ignition[1055]: INFO : files: files passed Jan 14 13:48:59.644669 ignition[1055]: INFO : Ignition finished successfully Jan 14 13:48:59.793850 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 14 13:48:59.826000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:59.832745 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 14 13:48:59.918422 kernel: audit: type=1130 audit(1768398539.826:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:59.884820 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 14 13:48:59.960514 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 14 13:49:00.053802 kernel: audit: type=1130 audit(1768398539.980:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:00.053840 kernel: audit: type=1131 audit(1768398539.980:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:59.980000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:59.980000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:48:59.960773 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 14 13:49:00.120700 kernel: audit: type=1130 audit(1768398540.072:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:00.072000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:00.120799 initrd-setup-root-after-ignition[1085]: grep: /sysroot/oem/oem-release: No such file or directory Jan 14 13:49:00.022864 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 14 13:49:00.193872 initrd-setup-root-after-ignition[1087]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 14 13:49:00.193872 initrd-setup-root-after-ignition[1087]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 14 13:49:00.074750 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 14 13:49:00.266943 initrd-setup-root-after-ignition[1090]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 14 13:49:00.159761 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 14 13:49:00.588858 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 14 13:49:00.712815 kernel: audit: type=1130 audit(1768398540.612:45): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:00.712908 kernel: audit: type=1131 audit(1768398540.612:46): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:00.612000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:00.612000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:00.589624 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 14 13:49:00.651632 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 14 13:49:00.728786 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 14 13:49:00.800344 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 14 13:49:00.823506 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 14 13:49:00.972794 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 14 13:49:01.035615 kernel: audit: type=1130 audit(1768398540.973:47): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:00.973000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:00.978497 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 14 13:49:01.108805 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Jan 14 13:49:01.109613 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 14 13:49:01.125824 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 14 13:49:01.188889 systemd[1]: Stopped target timers.target - Timer Units. Jan 14 13:49:01.270782 kernel: audit: type=1131 audit(1768398541.202:48): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:01.202000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:01.201831 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 14 13:49:01.202462 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 14 13:49:01.285976 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 14 13:49:01.304317 systemd[1]: Stopped target basic.target - Basic System. Jan 14 13:49:01.334408 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 14 13:49:01.392992 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 14 13:49:01.433506 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 14 13:49:01.434352 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Jan 14 13:49:01.503454 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 14 13:49:01.522937 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 14 13:49:01.551859 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 14 13:49:01.571781 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 14 13:49:01.628617 systemd[1]: Stopped target swap.target - Swaps. Jan 14 13:49:01.640704 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 14 13:49:01.640946 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 14 13:49:01.682000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:01.694918 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 14 13:49:01.750803 kernel: audit: type=1131 audit(1768398541.682:49): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:01.723987 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 14 13:49:01.775480 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 14 13:49:01.788751 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 14 13:49:01.834007 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 14 13:49:01.834877 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 14 13:49:01.883000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:01.910410 kernel: audit: type=1131 audit(1768398541.883:50): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:01.913278 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 14 13:49:01.914633 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 14 13:49:01.927000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:01.927796 systemd[1]: Stopped target paths.target - Path Units. Jan 14 13:49:01.973416 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 14 13:49:01.977765 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 14 13:49:01.982502 systemd[1]: Stopped target slices.target - Slice Units. Jan 14 13:49:02.021809 systemd[1]: Stopped target sockets.target - Socket Units. Jan 14 13:49:02.043332 systemd[1]: iscsid.socket: Deactivated successfully. Jan 14 13:49:02.043791 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 14 13:49:02.068789 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 14 13:49:02.069580 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 14 13:49:02.089966 systemd[1]: systemd-journald-audit.socket: Deactivated successfully. Jan 14 13:49:02.090599 systemd[1]: Closed systemd-journald-audit.socket - Journal Audit Socket. Jan 14 13:49:02.157000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.172000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.105925 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 14 13:49:02.106633 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 14 13:49:02.250000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.157854 systemd[1]: ignition-files.service: Deactivated successfully. Jan 14 13:49:02.158388 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 14 13:49:02.175717 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 14 13:49:02.214785 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 14 13:49:02.215399 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 14 13:49:02.345000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.252904 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 14 13:49:02.291671 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 14 13:49:02.315794 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 14 13:49:02.345939 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 14 13:49:02.422000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.389857 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 14 13:49:02.436000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.422744 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 14 13:49:02.423473 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 14 13:49:02.451270 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 14 13:49:02.483000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.500902 ignition[1113]: INFO : Ignition 2.24.0 Jan 14 13:49:02.500902 ignition[1113]: INFO : Stage: umount Jan 14 13:49:02.500902 ignition[1113]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 14 13:49:02.500902 ignition[1113]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 14 13:49:02.500902 ignition[1113]: INFO : umount: umount passed Jan 14 13:49:02.500902 ignition[1113]: INFO : Ignition finished successfully Jan 14 13:49:02.513000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.545000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.581000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.453275 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 14 13:49:02.453447 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 14 13:49:02.624000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.648000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup-pre comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.485422 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 14 13:49:02.673000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.485705 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 14 13:49:02.521419 systemd[1]: Stopped target network.target - Network. Jan 14 13:49:02.526757 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 14 13:49:02.526907 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 14 13:49:02.731000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.731000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.546354 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 14 13:49:02.546524 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 14 13:49:02.785000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.797000 audit: BPF prog-id=9 op=UNLOAD Jan 14 13:49:02.582514 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 14 13:49:02.582646 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 14 13:49:02.624920 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 14 13:49:02.625271 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 14 13:49:02.648806 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 14 13:49:02.874000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.648880 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 14 13:49:02.674552 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 14 13:49:02.698309 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 14 13:49:02.711490 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 14 13:49:02.711651 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 14 13:49:02.956000 audit: BPF prog-id=6 op=UNLOAD Jan 14 13:49:02.761448 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 14 13:49:02.761665 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 14 13:49:02.841331 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 14 13:49:02.841838 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 14 13:49:03.048000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.932509 systemd[1]: Stopped target network-pre.target - Preparation for Network. Jan 14 13:49:03.072000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.955500 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 14 13:49:02.955577 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 14 13:49:03.110000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:02.973737 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 14 13:49:03.021976 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 14 13:49:03.022413 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 14 13:49:03.048963 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 14 13:49:03.049379 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 14 13:49:03.072814 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 14 13:49:03.072893 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 14 13:49:03.111887 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 14 13:49:03.251895 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 14 13:49:03.252816 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 14 13:49:03.265733 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 14 13:49:03.264000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:03.265785 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 14 13:49:03.293716 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 14 13:49:03.340000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:03.293769 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 14 13:49:03.316931 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 14 13:49:03.316989 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 14 13:49:03.411659 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 14 13:49:03.412000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:03.411963 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 14 13:49:03.439550 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 14 13:49:03.449000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:03.439679 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 14 13:49:03.479990 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 14 13:49:03.493932 systemd[1]: systemd-network-generator.service: Deactivated successfully. Jan 14 13:49:03.494371 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Jan 14 13:49:03.560000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:03.561843 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 14 13:49:03.599000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:03.562407 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 14 13:49:03.635000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:03.599898 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 14 13:49:03.600426 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 14 13:49:03.683760 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 14 13:49:03.712889 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 14 13:49:03.732000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:03.732000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:03.749462 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 14 13:49:03.755000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:03.749866 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 14 13:49:03.760768 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 14 13:49:03.791943 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 14 13:49:03.877956 systemd[1]: Switching root. Jan 14 13:49:03.957527 systemd-journald[320]: Journal stopped Jan 14 13:49:10.887557 systemd-journald[320]: Received SIGTERM from PID 1 (systemd). Jan 14 13:49:10.887797 kernel: SELinux: policy capability network_peer_controls=1 Jan 14 13:49:10.887819 kernel: SELinux: policy capability open_perms=1 Jan 14 13:49:10.887841 kernel: SELinux: policy capability extended_socket_class=1 Jan 14 13:49:10.887857 kernel: SELinux: policy capability always_check_network=0 Jan 14 13:49:10.887882 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 14 13:49:10.888403 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 14 13:49:10.888422 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jan 14 13:49:10.888439 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 14 13:49:10.888454 kernel: SELinux: policy capability userspace_initial_context=0 Jan 14 13:49:10.888483 systemd[1]: Successfully loaded SELinux policy in 280.225ms. Jan 14 13:49:10.888739 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 21.166ms. Jan 14 13:49:10.888763 systemd[1]: systemd 257.9 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jan 14 13:49:10.888899 systemd[1]: Detected virtualization kvm. Jan 14 13:49:10.888926 systemd[1]: Detected architecture x86-64. Jan 14 13:49:10.888943 systemd[1]: Detected first boot. Jan 14 13:49:10.888959 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Jan 14 13:49:10.888976 kernel: Guest personality initialized and is inactive Jan 14 13:49:10.888991 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Jan 14 13:49:10.889006 kernel: Initialized host personality Jan 14 13:49:10.889520 kernel: NET: Registered PF_VSOCK protocol family Jan 14 13:49:10.889671 zram_generator::config[1157]: No configuration found. Jan 14 13:49:10.889692 systemd[1]: Populated /etc with preset unit settings. Jan 14 13:49:10.889709 kernel: kauditd_printk_skb: 39 callbacks suppressed Jan 14 13:49:10.889725 kernel: audit: type=1334 audit(1768398547.687:90): prog-id=12 op=LOAD Jan 14 13:49:10.889741 kernel: audit: type=1334 audit(1768398547.688:91): prog-id=3 op=UNLOAD Jan 14 13:49:10.890717 kernel: audit: type=1334 audit(1768398547.688:92): prog-id=13 op=LOAD Jan 14 13:49:10.890889 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jan 14 13:49:10.890914 kernel: audit: type=1334 audit(1768398547.688:93): prog-id=14 op=LOAD Jan 14 13:49:10.890934 kernel: audit: type=1334 audit(1768398547.688:94): prog-id=4 op=UNLOAD Jan 14 13:49:10.891396 kernel: audit: type=1334 audit(1768398547.688:95): prog-id=5 op=UNLOAD Jan 14 13:49:10.891422 kernel: audit: type=1131 audit(1768398547.694:96): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:10.891440 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jan 14 13:49:10.891459 kernel: audit: type=1130 audit(1768398547.840:97): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:10.891476 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jan 14 13:49:10.891493 kernel: audit: type=1131 audit(1768398547.840:98): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:10.891636 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jan 14 13:49:10.891656 kernel: audit: type=1334 audit(1768398547.962:99): prog-id=12 op=UNLOAD Jan 14 13:49:10.891672 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jan 14 13:49:10.891689 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jan 14 13:49:10.891706 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jan 14 13:49:10.891723 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jan 14 13:49:10.891866 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jan 14 13:49:10.891887 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jan 14 13:49:10.892400 systemd[1]: Created slice user.slice - User and Session Slice. Jan 14 13:49:10.892549 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 14 13:49:10.892666 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 14 13:49:10.892685 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jan 14 13:49:10.892702 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jan 14 13:49:10.892847 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jan 14 13:49:10.892867 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 14 13:49:10.892888 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jan 14 13:49:10.892913 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 14 13:49:10.892937 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 14 13:49:10.892961 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jan 14 13:49:10.892978 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jan 14 13:49:10.893490 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jan 14 13:49:10.893513 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jan 14 13:49:10.893530 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 14 13:49:10.893547 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 14 13:49:10.893563 systemd[1]: Reached target remote-veritysetup.target - Remote Verity Protected Volumes. Jan 14 13:49:10.893579 systemd[1]: Reached target slices.target - Slice Units. Jan 14 13:49:10.893598 systemd[1]: Reached target swap.target - Swaps. Jan 14 13:49:10.893618 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jan 14 13:49:10.893780 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jan 14 13:49:10.893801 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Jan 14 13:49:10.893819 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Jan 14 13:49:10.893836 systemd[1]: Listening on systemd-mountfsd.socket - DDI File System Mounter Socket. Jan 14 13:49:10.893852 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 14 13:49:10.893873 systemd[1]: Listening on systemd-nsresourced.socket - Namespace Resource Manager Socket. Jan 14 13:49:10.894395 systemd[1]: Listening on systemd-oomd.socket - Userspace Out-Of-Memory (OOM) Killer Socket. Jan 14 13:49:10.894547 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 14 13:49:10.894569 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 14 13:49:10.894586 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jan 14 13:49:10.894603 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jan 14 13:49:10.894619 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jan 14 13:49:10.894635 systemd[1]: Mounting media.mount - External Media Directory... Jan 14 13:49:10.894651 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 14 13:49:10.894792 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jan 14 13:49:10.894811 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jan 14 13:49:10.894828 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jan 14 13:49:10.894847 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jan 14 13:49:10.894978 systemd[1]: Reached target machines.target - Containers. Jan 14 13:49:10.894996 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jan 14 13:49:10.895382 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 14 13:49:10.895407 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 14 13:49:10.895426 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jan 14 13:49:10.895443 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 14 13:49:10.895460 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 14 13:49:10.895476 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 14 13:49:10.895493 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jan 14 13:49:10.895634 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 14 13:49:10.895653 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jan 14 13:49:10.895670 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jan 14 13:49:10.895686 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jan 14 13:49:10.895703 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jan 14 13:49:10.895718 systemd[1]: Stopped systemd-fsck-usr.service. Jan 14 13:49:10.895740 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 14 13:49:10.895870 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 14 13:49:10.895998 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 14 13:49:10.896399 kernel: fuse: init (API version 7.41) Jan 14 13:49:10.896437 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 14 13:49:10.896576 kernel: ACPI: bus type drm_connector registered Jan 14 13:49:10.896595 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jan 14 13:49:10.896613 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Jan 14 13:49:10.896663 systemd-journald[1238]: Collecting audit messages is enabled. Jan 14 13:49:10.896705 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 14 13:49:10.896724 systemd-journald[1238]: Journal started Jan 14 13:49:10.896989 systemd-journald[1238]: Runtime Journal (/run/log/journal/3f664cc37dea4d3e875d093b3e52a6f3) is 6M, max 48M, 42M free. Jan 14 13:49:08.906000 audit[1]: EVENT_LISTENER pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Jan 14 13:49:10.019000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:10.086000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:10.146000 audit: BPF prog-id=14 op=UNLOAD Jan 14 13:49:10.146000 audit: BPF prog-id=13 op=UNLOAD Jan 14 13:49:10.171000 audit: BPF prog-id=15 op=LOAD Jan 14 13:49:10.554000 audit: BPF prog-id=16 op=LOAD Jan 14 13:49:10.555000 audit: BPF prog-id=17 op=LOAD Jan 14 13:49:10.879000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Jan 14 13:49:10.879000 audit[1238]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=5 a1=7ffdd1c7e940 a2=4000 a3=0 items=0 ppid=1 pid=1238 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:10.879000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Jan 14 13:49:07.658930 systemd[1]: Queued start job for default target multi-user.target. Jan 14 13:49:07.690541 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Jan 14 13:49:07.692884 systemd[1]: systemd-journald.service: Deactivated successfully. Jan 14 13:49:07.694683 systemd[1]: systemd-journald.service: Consumed 5.037s CPU time. Jan 14 13:49:10.958993 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 14 13:49:10.980479 systemd[1]: Started systemd-journald.service - Journal Service. Jan 14 13:49:11.002000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.004985 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jan 14 13:49:11.036618 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jan 14 13:49:11.059775 systemd[1]: Mounted media.mount - External Media Directory. Jan 14 13:49:11.091900 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jan 14 13:49:11.113937 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jan 14 13:49:11.131457 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jan 14 13:49:11.321807 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jan 14 13:49:11.338000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.340614 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 14 13:49:11.358000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.365954 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jan 14 13:49:11.366985 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jan 14 13:49:11.383000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.383000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.384399 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 14 13:49:11.384672 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 14 13:49:11.398000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.398000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.400911 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 14 13:49:11.402731 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 14 13:49:11.415000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.415000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.416637 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 14 13:49:11.417379 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 14 13:49:11.431000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.431000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.432978 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jan 14 13:49:11.433830 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jan 14 13:49:11.452000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.452000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.453944 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 14 13:49:11.454669 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 14 13:49:11.475000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.475000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.476732 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 14 13:49:11.494000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.496509 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 14 13:49:11.516000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.518710 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jan 14 13:49:11.539000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.541744 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Jan 14 13:49:11.563000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-load-credentials comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.596746 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 14 13:49:11.615438 systemd[1]: Listening on systemd-importd.socket - Disk Image Download Service Socket. Jan 14 13:49:11.636387 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jan 14 13:49:11.654401 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jan 14 13:49:11.669925 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jan 14 13:49:11.669996 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 14 13:49:11.685907 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Jan 14 13:49:11.704387 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 14 13:49:11.704687 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 14 13:49:11.720548 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jan 14 13:49:11.752007 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jan 14 13:49:11.773448 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 14 13:49:11.781952 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jan 14 13:49:11.797988 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 14 13:49:11.803230 systemd-journald[1238]: Time spent on flushing to /var/log/journal/3f664cc37dea4d3e875d093b3e52a6f3 is 77.757ms for 1211 entries. Jan 14 13:49:11.803230 systemd-journald[1238]: System Journal (/var/log/journal/3f664cc37dea4d3e875d093b3e52a6f3) is 8M, max 163.5M, 155.5M free. Jan 14 13:49:11.971842 systemd-journald[1238]: Received client request to flush runtime journal. Jan 14 13:49:11.924000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:11.801608 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 14 13:49:11.847507 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jan 14 13:49:11.875922 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jan 14 13:49:11.903375 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 14 13:49:11.925533 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jan 14 13:49:11.948663 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jan 14 13:49:11.985487 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jan 14 13:49:12.011000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:12.020929 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jan 14 13:49:12.041000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:12.051649 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jan 14 13:49:12.066844 kernel: loop1: detected capacity change from 0 to 219144 Jan 14 13:49:12.096584 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Jan 14 13:49:12.161861 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 14 13:49:12.187000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:12.251708 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jan 14 13:49:12.259889 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Jan 14 13:49:12.281359 kernel: loop2: detected capacity change from 0 to 50784 Jan 14 13:49:12.300000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:12.335862 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jan 14 13:49:12.348000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:12.356000 audit: BPF prog-id=18 op=LOAD Jan 14 13:49:12.356000 audit: BPF prog-id=19 op=LOAD Jan 14 13:49:12.357000 audit: BPF prog-id=20 op=LOAD Jan 14 13:49:12.361549 systemd[1]: Starting systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer... Jan 14 13:49:12.398005 kernel: loop3: detected capacity change from 0 to 111560 Jan 14 13:49:12.401000 audit: BPF prog-id=21 op=LOAD Jan 14 13:49:12.405650 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 14 13:49:12.431419 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 14 13:49:12.457000 audit: BPF prog-id=22 op=LOAD Jan 14 13:49:12.457000 audit: BPF prog-id=23 op=LOAD Jan 14 13:49:12.457000 audit: BPF prog-id=24 op=LOAD Jan 14 13:49:12.459886 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jan 14 13:49:12.482000 audit: BPF prog-id=25 op=LOAD Jan 14 13:49:12.482000 audit: BPF prog-id=26 op=LOAD Jan 14 13:49:12.482000 audit: BPF prog-id=27 op=LOAD Jan 14 13:49:13.027497 systemd[1]: Starting systemd-nsresourced.service - Namespace Resource Manager... Jan 14 13:49:13.083341 kernel: loop4: detected capacity change from 0 to 219144 Jan 14 13:49:13.115200 kernel: loop5: detected capacity change from 0 to 50784 Jan 14 13:49:13.135452 kernel: loop6: detected capacity change from 0 to 111560 Jan 14 13:49:13.134613 systemd-tmpfiles[1298]: ACLs are not supported, ignoring. Jan 14 13:49:13.134642 systemd-tmpfiles[1298]: ACLs are not supported, ignoring. Jan 14 13:49:13.149815 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 14 13:49:13.155553 (sd-merge)[1302]: Using extensions 'containerd-flatcar.raw', 'docker-flatcar.raw', 'kubernetes.raw'. Jan 14 13:49:13.158000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:13.161753 (sd-merge)[1302]: Merged extensions into '/usr'. Jan 14 13:49:13.166199 kernel: kauditd_printk_skb: 46 callbacks suppressed Jan 14 13:49:13.166390 kernel: audit: type=1130 audit(1768398553.158:144): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:13.193348 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jan 14 13:49:13.205000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:13.215883 systemd[1]: Reload requested from client PID 1277 ('systemd-sysext') (unit systemd-sysext.service)... Jan 14 13:49:13.215899 systemd[1]: Reloading... Jan 14 13:49:13.227163 kernel: audit: type=1130 audit(1768398553.205:145): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:13.247878 systemd-nsresourced[1301]: Not setting up BPF subsystem, as functionality has been disabled at compile time. Jan 14 13:49:13.671192 zram_generator::config[1346]: No configuration found. Jan 14 13:49:13.751803 systemd-oomd[1296]: No swap; memory pressure usage will be degraded Jan 14 13:49:13.790232 systemd-resolved[1297]: Positive Trust Anchors: Jan 14 13:49:13.790329 systemd-resolved[1297]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 14 13:49:13.790336 systemd-resolved[1297]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Jan 14 13:49:13.790380 systemd-resolved[1297]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 14 13:49:14.119934 systemd-resolved[1297]: Defaulting to hostname 'linux'. Jan 14 13:49:14.253563 systemd[1]: Reloading finished in 1036 ms. Jan 14 13:49:14.290837 systemd[1]: Started systemd-nsresourced.service - Namespace Resource Manager. Jan 14 13:49:14.301000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-nsresourced comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:14.302965 systemd[1]: Started systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer. Jan 14 13:49:14.320233 kernel: audit: type=1130 audit(1768398554.301:146): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-nsresourced comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:14.328000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-oomd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:14.329711 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 14 13:49:14.345345 kernel: audit: type=1130 audit(1768398554.328:147): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-oomd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:14.354000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:14.355522 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jan 14 13:49:14.390684 kernel: audit: type=1130 audit(1768398554.354:148): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:14.412000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:14.438645 kernel: audit: type=1130 audit(1768398554.412:149): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:14.530420 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 14 13:49:14.666583 systemd[1]: Starting ensure-sysext.service... Jan 14 13:49:14.681000 audit: BPF prog-id=28 op=LOAD Jan 14 13:49:14.673576 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 14 13:49:14.688194 kernel: audit: type=1334 audit(1768398554.681:150): prog-id=28 op=LOAD Jan 14 13:49:14.682000 audit: BPF prog-id=25 op=UNLOAD Jan 14 13:49:14.696248 kernel: audit: type=1334 audit(1768398554.682:151): prog-id=25 op=UNLOAD Jan 14 13:49:14.682000 audit: BPF prog-id=29 op=LOAD Jan 14 13:49:14.709100 kernel: audit: type=1334 audit(1768398554.682:152): prog-id=29 op=LOAD Jan 14 13:49:14.709177 kernel: audit: type=1334 audit(1768398554.682:153): prog-id=30 op=LOAD Jan 14 13:49:14.682000 audit: BPF prog-id=30 op=LOAD Jan 14 13:49:14.682000 audit: BPF prog-id=26 op=UNLOAD Jan 14 13:49:14.682000 audit: BPF prog-id=27 op=UNLOAD Jan 14 13:49:14.684000 audit: BPF prog-id=31 op=LOAD Jan 14 13:49:14.684000 audit: BPF prog-id=18 op=UNLOAD Jan 14 13:49:14.684000 audit: BPF prog-id=32 op=LOAD Jan 14 13:49:14.684000 audit: BPF prog-id=33 op=LOAD Jan 14 13:49:14.684000 audit: BPF prog-id=19 op=UNLOAD Jan 14 13:49:14.684000 audit: BPF prog-id=20 op=UNLOAD Jan 14 13:49:14.685000 audit: BPF prog-id=34 op=LOAD Jan 14 13:49:14.685000 audit: BPF prog-id=22 op=UNLOAD Jan 14 13:49:14.686000 audit: BPF prog-id=35 op=LOAD Jan 14 13:49:14.686000 audit: BPF prog-id=36 op=LOAD Jan 14 13:49:14.686000 audit: BPF prog-id=23 op=UNLOAD Jan 14 13:49:14.686000 audit: BPF prog-id=24 op=UNLOAD Jan 14 13:49:14.687000 audit: BPF prog-id=37 op=LOAD Jan 14 13:49:14.687000 audit: BPF prog-id=21 op=UNLOAD Jan 14 13:49:14.691000 audit: BPF prog-id=38 op=LOAD Jan 14 13:49:14.691000 audit: BPF prog-id=15 op=UNLOAD Jan 14 13:49:14.691000 audit: BPF prog-id=39 op=LOAD Jan 14 13:49:14.692000 audit: BPF prog-id=40 op=LOAD Jan 14 13:49:14.692000 audit: BPF prog-id=16 op=UNLOAD Jan 14 13:49:14.692000 audit: BPF prog-id=17 op=UNLOAD Jan 14 13:49:14.742748 systemd[1]: Reload requested from client PID 1382 ('systemctl') (unit ensure-sysext.service)... Jan 14 13:49:14.742765 systemd[1]: Reloading... Jan 14 13:49:14.800574 systemd-tmpfiles[1383]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Jan 14 13:49:14.800679 systemd-tmpfiles[1383]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Jan 14 13:49:14.801394 systemd-tmpfiles[1383]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jan 14 13:49:14.803912 systemd-tmpfiles[1383]: ACLs are not supported, ignoring. Jan 14 13:49:14.804139 systemd-tmpfiles[1383]: ACLs are not supported, ignoring. Jan 14 13:49:14.813124 systemd-tmpfiles[1383]: Detected autofs mount point /boot during canonicalization of boot. Jan 14 13:49:14.813139 systemd-tmpfiles[1383]: Skipping /boot Jan 14 13:49:14.839889 systemd-tmpfiles[1383]: Detected autofs mount point /boot during canonicalization of boot. Jan 14 13:49:14.839951 systemd-tmpfiles[1383]: Skipping /boot Jan 14 13:49:14.965456 zram_generator::config[1413]: No configuration found. Jan 14 13:49:15.378687 systemd[1]: Reloading finished in 635 ms. Jan 14 13:49:15.406449 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 14 13:49:15.417000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:15.421439 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jan 14 13:49:15.430000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:15.433000 audit: BPF prog-id=41 op=LOAD Jan 14 13:49:15.434000 audit: BPF prog-id=34 op=UNLOAD Jan 14 13:49:15.434000 audit: BPF prog-id=42 op=LOAD Jan 14 13:49:15.434000 audit: BPF prog-id=43 op=LOAD Jan 14 13:49:15.434000 audit: BPF prog-id=35 op=UNLOAD Jan 14 13:49:15.434000 audit: BPF prog-id=36 op=UNLOAD Jan 14 13:49:15.435000 audit: BPF prog-id=44 op=LOAD Jan 14 13:49:15.435000 audit: BPF prog-id=28 op=UNLOAD Jan 14 13:49:15.435000 audit: BPF prog-id=45 op=LOAD Jan 14 13:49:15.435000 audit: BPF prog-id=46 op=LOAD Jan 14 13:49:15.435000 audit: BPF prog-id=29 op=UNLOAD Jan 14 13:49:15.435000 audit: BPF prog-id=30 op=UNLOAD Jan 14 13:49:15.436000 audit: BPF prog-id=47 op=LOAD Jan 14 13:49:15.436000 audit: BPF prog-id=31 op=UNLOAD Jan 14 13:49:15.436000 audit: BPF prog-id=48 op=LOAD Jan 14 13:49:15.436000 audit: BPF prog-id=49 op=LOAD Jan 14 13:49:15.436000 audit: BPF prog-id=32 op=UNLOAD Jan 14 13:49:15.436000 audit: BPF prog-id=33 op=UNLOAD Jan 14 13:49:15.439000 audit: BPF prog-id=50 op=LOAD Jan 14 13:49:15.439000 audit: BPF prog-id=37 op=UNLOAD Jan 14 13:49:15.441000 audit: BPF prog-id=51 op=LOAD Jan 14 13:49:15.442000 audit: BPF prog-id=38 op=UNLOAD Jan 14 13:49:15.442000 audit: BPF prog-id=52 op=LOAD Jan 14 13:49:15.442000 audit: BPF prog-id=53 op=LOAD Jan 14 13:49:15.442000 audit: BPF prog-id=39 op=UNLOAD Jan 14 13:49:15.442000 audit: BPF prog-id=40 op=UNLOAD Jan 14 13:49:15.482191 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 14 13:49:15.489882 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jan 14 13:49:15.506538 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jan 14 13:49:15.516407 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jan 14 13:49:15.524000 audit: BPF prog-id=8 op=UNLOAD Jan 14 13:49:15.524000 audit: BPF prog-id=7 op=UNLOAD Jan 14 13:49:15.525000 audit: BPF prog-id=54 op=LOAD Jan 14 13:49:15.525000 audit: BPF prog-id=55 op=LOAD Jan 14 13:49:15.528206 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 14 13:49:15.540886 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jan 14 13:49:15.557950 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 14 13:49:15.558351 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 14 13:49:15.566526 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 14 13:49:15.637158 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 14 13:49:15.650000 audit[1465]: SYSTEM_BOOT pid=1465 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Jan 14 13:49:15.697187 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 14 13:49:15.714921 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 14 13:49:15.715941 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 14 13:49:15.724189 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 14 13:49:15.736998 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 14 13:49:15.756986 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 14 13:49:15.773851 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 14 13:49:15.794000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:15.794000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:15.804928 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 14 13:49:15.807822 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 14 13:49:15.820946 systemd-udevd[1464]: Using default interface naming scheme 'v257'. Jan 14 13:49:15.849000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:15.849000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:15.850271 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 14 13:49:15.850627 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 14 13:49:15.873000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Jan 14 13:49:15.873000 audit[1482]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffe820368a0 a2=420 a3=0 items=0 ppid=1454 pid=1482 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:15.873000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 14 13:49:15.885574 augenrules[1482]: No rules Jan 14 13:49:15.895914 systemd[1]: audit-rules.service: Deactivated successfully. Jan 14 13:49:15.896738 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 14 13:49:15.927371 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jan 14 13:49:16.006804 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 14 13:49:16.016979 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jan 14 13:49:16.041218 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jan 14 13:49:16.111577 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 14 13:49:16.116477 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 14 13:49:16.125541 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 14 13:49:16.131581 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 14 13:49:16.157172 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 14 13:49:16.197992 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 14 13:49:16.218742 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 14 13:49:16.229603 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 14 13:49:16.234246 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 14 13:49:16.247930 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 14 13:49:16.408437 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 14 13:49:16.416700 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 14 13:49:16.416836 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 14 13:49:16.419577 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 14 13:49:16.419942 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 14 13:49:16.429619 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 14 13:49:16.430157 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 14 13:49:16.464419 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 14 13:49:16.464772 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 14 13:49:16.486994 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Jan 14 13:49:16.489431 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 14 13:49:16.494778 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 14 13:49:16.507595 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jan 14 13:49:16.527695 systemd[1]: Finished ensure-sysext.service. Jan 14 13:49:16.534672 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 14 13:49:16.535625 augenrules[1508]: /sbin/augenrules: No change Jan 14 13:49:16.542799 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 14 13:49:16.844704 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 14 13:49:16.850375 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jan 14 13:49:16.873000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Jan 14 13:49:16.873000 audit[1545]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffc159c6bb0 a2=420 a3=0 items=0 ppid=1508 pid=1545 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:16.873000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 14 13:49:16.874000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Jan 14 13:49:16.874000 audit[1545]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffc159c9040 a2=420 a3=0 items=0 ppid=1508 pid=1545 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:16.874000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 14 13:49:16.875404 augenrules[1545]: No rules Jan 14 13:49:16.878686 systemd[1]: audit-rules.service: Deactivated successfully. Jan 14 13:49:16.879233 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 14 13:49:17.138274 kernel: mousedev: PS/2 mouse device common for all mice Jan 14 13:49:17.149220 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jan 14 13:49:17.203175 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Jan 14 13:49:17.233168 kernel: ACPI: button: Power Button [PWRF] Jan 14 13:49:17.257349 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Jan 14 13:49:17.285577 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Jan 14 13:49:17.286245 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Jan 14 13:49:18.038741 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jan 14 13:49:18.052241 systemd[1]: Reached target time-set.target - System Time Set. Jan 14 13:49:18.548806 systemd-networkd[1527]: lo: Link UP Jan 14 13:49:18.548884 systemd-networkd[1527]: lo: Gained carrier Jan 14 13:49:18.554468 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 14 13:49:18.554871 systemd-networkd[1527]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 14 13:49:18.554956 systemd-networkd[1527]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 14 13:49:18.575424 systemd[1]: Reached target network.target - Network. Jan 14 13:49:18.580647 systemd-networkd[1527]: eth0: Link UP Jan 14 13:49:18.583238 systemd-networkd[1527]: eth0: Gained carrier Jan 14 13:49:18.583455 systemd-networkd[1527]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 14 13:49:18.598187 systemd-networkd[1527]: eth0: DHCPv4 address 10.0.0.127/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jan 14 13:49:18.599755 systemd-timesyncd[1549]: Network configuration changed, trying to establish connection. Jan 14 13:49:18.604694 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Jan 14 13:49:19.900760 systemd-timesyncd[1549]: Contacted time server 10.0.0.1:123 (10.0.0.1). Jan 14 13:49:19.900895 systemd-timesyncd[1549]: Initial clock synchronization to Wed 2026-01-14 13:49:19.900456 UTC. Jan 14 13:49:19.900941 systemd-resolved[1297]: Clock change detected. Flushing caches. Jan 14 13:49:19.917760 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jan 14 13:49:19.981352 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 14 13:49:20.232391 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 14 13:49:20.233048 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 14 13:49:20.247301 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 14 13:49:20.282260 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Jan 14 13:49:20.706916 kernel: kvm_amd: TSC scaling supported Jan 14 13:49:20.707126 kernel: kvm_amd: Nested Virtualization enabled Jan 14 13:49:20.707151 kernel: kvm_amd: Nested Paging enabled Jan 14 13:49:20.715904 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Jan 14 13:49:20.715963 kernel: kvm_amd: PMU virtualization is disabled Jan 14 13:49:20.738982 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 14 13:49:20.753115 ldconfig[1456]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jan 14 13:49:20.781439 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jan 14 13:49:20.797252 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jan 14 13:49:20.900955 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jan 14 13:49:20.912088 systemd[1]: Reached target sysinit.target - System Initialization. Jan 14 13:49:20.920719 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jan 14 13:49:20.929205 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jan 14 13:49:20.937314 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Jan 14 13:49:20.945303 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jan 14 13:49:20.955346 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jan 14 13:49:20.966072 systemd[1]: Started systemd-sysupdate-reboot.timer - Reboot Automatically After System Update. Jan 14 13:49:20.975159 systemd[1]: Started systemd-sysupdate.timer - Automatic System Update. Jan 14 13:49:20.982704 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jan 14 13:49:20.991188 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jan 14 13:49:20.991299 systemd[1]: Reached target paths.target - Path Units. Jan 14 13:49:20.997957 systemd[1]: Reached target timers.target - Timer Units. Jan 14 13:49:21.008421 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jan 14 13:49:21.019374 systemd[1]: Starting docker.socket - Docker Socket for the API... Jan 14 13:49:21.034146 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Jan 14 13:49:21.043267 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Jan 14 13:49:21.052318 systemd[1]: Reached target ssh-access.target - SSH Access Available. Jan 14 13:49:21.078236 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jan 14 13:49:21.085808 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Jan 14 13:49:21.095498 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jan 14 13:49:21.106780 kernel: EDAC MC: Ver: 3.0.0 Jan 14 13:49:21.107972 systemd[1]: Reached target sockets.target - Socket Units. Jan 14 13:49:21.114969 systemd[1]: Reached target basic.target - Basic System. Jan 14 13:49:21.121465 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jan 14 13:49:21.121901 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jan 14 13:49:21.124336 systemd[1]: Starting containerd.service - containerd container runtime... Jan 14 13:49:21.134351 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jan 14 13:49:21.143803 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jan 14 13:49:21.154692 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jan 14 13:49:21.165804 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jan 14 13:49:21.172338 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jan 14 13:49:21.175117 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Jan 14 13:49:21.188659 jq[1601]: false Jan 14 13:49:21.198918 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jan 14 13:49:21.212005 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jan 14 13:49:21.217379 google_oslogin_nss_cache[1603]: oslogin_cache_refresh[1603]: Refreshing passwd entry cache Jan 14 13:49:21.217346 oslogin_cache_refresh[1603]: Refreshing passwd entry cache Jan 14 13:49:21.223510 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jan 14 13:49:21.237495 extend-filesystems[1602]: Found /dev/vda6 Jan 14 13:49:21.237733 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jan 14 13:49:21.245242 oslogin_cache_refresh[1603]: Failure getting users, quitting Jan 14 13:49:21.252783 google_oslogin_nss_cache[1603]: oslogin_cache_refresh[1603]: Failure getting users, quitting Jan 14 13:49:21.252783 google_oslogin_nss_cache[1603]: oslogin_cache_refresh[1603]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Jan 14 13:49:21.252783 google_oslogin_nss_cache[1603]: oslogin_cache_refresh[1603]: Refreshing group entry cache Jan 14 13:49:21.245264 oslogin_cache_refresh[1603]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Jan 14 13:49:21.245320 oslogin_cache_refresh[1603]: Refreshing group entry cache Jan 14 13:49:21.262655 extend-filesystems[1602]: Found /dev/vda9 Jan 14 13:49:21.261788 systemd[1]: Starting systemd-logind.service - User Login Management... Jan 14 13:49:21.274091 extend-filesystems[1602]: Checking size of /dev/vda9 Jan 14 13:49:21.269170 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jan 14 13:49:21.280984 oslogin_cache_refresh[1603]: Failure getting groups, quitting Jan 14 13:49:21.300245 google_oslogin_nss_cache[1603]: oslogin_cache_refresh[1603]: Failure getting groups, quitting Jan 14 13:49:21.300245 google_oslogin_nss_cache[1603]: oslogin_cache_refresh[1603]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Jan 14 13:49:21.270111 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jan 14 13:49:21.281002 oslogin_cache_refresh[1603]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Jan 14 13:49:21.271342 systemd[1]: Starting update-engine.service - Update Engine... Jan 14 13:49:21.283119 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jan 14 13:49:21.307761 extend-filesystems[1602]: Resized partition /dev/vda9 Jan 14 13:49:21.350320 kernel: EXT4-fs (vda9): resizing filesystem from 456704 to 1784827 blocks Jan 14 13:49:21.307725 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jan 14 13:49:21.350434 extend-filesystems[1628]: resize2fs 1.47.3 (8-Jul-2025) Jan 14 13:49:21.308301 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jan 14 13:49:21.370810 jq[1622]: true Jan 14 13:49:21.308686 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jan 14 13:49:21.309073 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Jan 14 13:49:21.309305 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Jan 14 13:49:21.322698 systemd[1]: motdgen.service: Deactivated successfully. Jan 14 13:49:21.323937 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jan 14 13:49:21.327086 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jan 14 13:49:21.327366 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jan 14 13:49:21.372153 systemd-networkd[1527]: eth0: Gained IPv6LL Jan 14 13:49:21.401987 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jan 14 13:49:21.432420 systemd[1]: Reached target network-online.target - Network is Online. Jan 14 13:49:21.444762 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Jan 14 13:49:21.447894 jq[1633]: true Jan 14 13:49:21.455149 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 13:49:21.466016 update_engine[1620]: I20260114 13:49:21.465340 1620 main.cc:92] Flatcar Update Engine starting Jan 14 13:49:21.473039 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jan 14 13:49:21.493414 tar[1631]: linux-amd64/LICENSE Jan 14 13:49:21.493414 tar[1631]: linux-amd64/helm Jan 14 13:49:21.522657 kernel: EXT4-fs (vda9): resized filesystem to 1784827 Jan 14 13:49:21.564045 extend-filesystems[1628]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Jan 14 13:49:21.564045 extend-filesystems[1628]: old_desc_blocks = 1, new_desc_blocks = 1 Jan 14 13:49:21.564045 extend-filesystems[1628]: The filesystem on /dev/vda9 is now 1784827 (4k) blocks long. Jan 14 13:49:21.603004 extend-filesystems[1602]: Resized filesystem in /dev/vda9 Jan 14 13:49:21.565198 systemd[1]: extend-filesystems.service: Deactivated successfully. Jan 14 13:49:21.565736 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jan 14 13:49:21.582407 systemd-logind[1617]: Watching system buttons on /dev/input/event2 (Power Button) Jan 14 13:49:21.582442 systemd-logind[1617]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jan 14 13:49:21.592319 systemd-logind[1617]: New seat seat0. Jan 14 13:49:21.603426 systemd[1]: Started systemd-logind.service - User Login Management. Jan 14 13:49:21.622784 sshd_keygen[1629]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jan 14 13:49:21.658802 dbus-daemon[1599]: [system] SELinux support is enabled Jan 14 13:49:21.659492 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jan 14 13:49:21.669197 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jan 14 13:49:21.669233 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jan 14 13:49:21.679752 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jan 14 13:49:21.679772 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jan 14 13:49:21.687797 dbus-daemon[1599]: [system] Successfully activated service 'org.freedesktop.systemd1' Jan 14 13:49:21.691991 update_engine[1620]: I20260114 13:49:21.690073 1620 update_check_scheduler.cc:74] Next update check in 7m44s Jan 14 13:49:21.691805 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jan 14 13:49:21.701510 bash[1688]: Updated "/home/core/.ssh/authorized_keys" Jan 14 13:49:21.706722 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jan 14 13:49:21.718731 systemd[1]: coreos-metadata.service: Deactivated successfully. Jan 14 13:49:21.719135 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Jan 14 13:49:21.733802 systemd[1]: Started update-engine.service - Update Engine. Jan 14 13:49:21.742062 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jan 14 13:49:21.744147 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Jan 14 13:49:21.750349 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jan 14 13:49:21.762168 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jan 14 13:49:21.779083 systemd[1]: Starting issuegen.service - Generate /run/issue... Jan 14 13:49:21.811995 systemd[1]: issuegen.service: Deactivated successfully. Jan 14 13:49:21.812408 systemd[1]: Finished issuegen.service - Generate /run/issue. Jan 14 13:49:21.830409 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jan 14 13:49:21.880714 locksmithd[1701]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jan 14 13:49:21.881471 containerd[1635]: time="2026-01-14T13:49:21Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Jan 14 13:49:21.885639 containerd[1635]: time="2026-01-14T13:49:21.885610317Z" level=info msg="starting containerd" revision=fcd43222d6b07379a4be9786bda52438f0dd16a1 version=v2.1.5 Jan 14 13:49:21.889017 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jan 14 13:49:21.900685 containerd[1635]: time="2026-01-14T13:49:21.900472809Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="8.806µs" Jan 14 13:49:21.900930 containerd[1635]: time="2026-01-14T13:49:21.900747592Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Jan 14 13:49:21.901056 containerd[1635]: time="2026-01-14T13:49:21.901032323Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Jan 14 13:49:21.901123 containerd[1635]: time="2026-01-14T13:49:21.901105790Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Jan 14 13:49:21.901317 containerd[1635]: time="2026-01-14T13:49:21.901298109Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Jan 14 13:49:21.901382 containerd[1635]: time="2026-01-14T13:49:21.901367790Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jan 14 13:49:21.901501 containerd[1635]: time="2026-01-14T13:49:21.901484828Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jan 14 13:49:21.901707 containerd[1635]: time="2026-01-14T13:49:21.901691223Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jan 14 13:49:21.902068 containerd[1635]: time="2026-01-14T13:49:21.902047358Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jan 14 13:49:21.902126 containerd[1635]: time="2026-01-14T13:49:21.902114924Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jan 14 13:49:21.902172 containerd[1635]: time="2026-01-14T13:49:21.902158775Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jan 14 13:49:21.902209 containerd[1635]: time="2026-01-14T13:49:21.902199472Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Jan 14 13:49:21.902676 containerd[1635]: time="2026-01-14T13:49:21.902495724Z" level=info msg="skip loading plugin" error="EROFS unsupported, please `modprobe erofs`: skip plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Jan 14 13:49:21.902736 containerd[1635]: time="2026-01-14T13:49:21.902722278Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Jan 14 13:49:21.902972 containerd[1635]: time="2026-01-14T13:49:21.902951966Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Jan 14 13:49:21.903237 systemd[1]: Started getty@tty1.service - Getty on tty1. Jan 14 13:49:21.904242 containerd[1635]: time="2026-01-14T13:49:21.904223179Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jan 14 13:49:21.904322 containerd[1635]: time="2026-01-14T13:49:21.904308689Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jan 14 13:49:21.904367 containerd[1635]: time="2026-01-14T13:49:21.904353953Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Jan 14 13:49:21.904442 containerd[1635]: time="2026-01-14T13:49:21.904427690Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Jan 14 13:49:21.904997 containerd[1635]: time="2026-01-14T13:49:21.904978770Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Jan 14 13:49:21.905110 containerd[1635]: time="2026-01-14T13:49:21.905095117Z" level=info msg="metadata content store policy set" policy=shared Jan 14 13:49:21.920656 containerd[1635]: time="2026-01-14T13:49:21.916738690Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Jan 14 13:49:21.920656 containerd[1635]: time="2026-01-14T13:49:21.916900221Z" level=info msg="loading plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Jan 14 13:49:21.920656 containerd[1635]: time="2026-01-14T13:49:21.916981122Z" level=info msg="skip loading plugin" error="could not find mkfs.erofs: exec: \"mkfs.erofs\": executable file not found in $PATH: skip plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Jan 14 13:49:21.920656 containerd[1635]: time="2026-01-14T13:49:21.916994697Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Jan 14 13:49:21.920656 containerd[1635]: time="2026-01-14T13:49:21.917007592Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Jan 14 13:49:21.920656 containerd[1635]: time="2026-01-14T13:49:21.917021648Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Jan 14 13:49:21.920656 containerd[1635]: time="2026-01-14T13:49:21.917111265Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Jan 14 13:49:21.920656 containerd[1635]: time="2026-01-14T13:49:21.917122677Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Jan 14 13:49:21.920656 containerd[1635]: time="2026-01-14T13:49:21.917133957Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Jan 14 13:49:21.920656 containerd[1635]: time="2026-01-14T13:49:21.917146411Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Jan 14 13:49:21.920656 containerd[1635]: time="2026-01-14T13:49:21.917161068Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Jan 14 13:49:21.920656 containerd[1635]: time="2026-01-14T13:49:21.917171688Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Jan 14 13:49:21.920656 containerd[1635]: time="2026-01-14T13:49:21.917184421Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Jan 14 13:49:21.920656 containerd[1635]: time="2026-01-14T13:49:21.917198528Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Jan 14 13:49:21.917450 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917343048Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917372303Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917389644Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917402889Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917417446Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917431553Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917446070Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917463633Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917482268Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917500702Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917644971Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917680898Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917735550Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917752161Z" level=info msg="Start snapshots syncer" Jan 14 13:49:21.921217 containerd[1635]: time="2026-01-14T13:49:21.917928671Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Jan 14 13:49:21.923738 containerd[1635]: time="2026-01-14T13:49:21.918466815Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"cgroupWritable\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"\",\"binDirs\":[\"/opt/cni/bin\"],\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogLineSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Jan 14 13:49:21.924187 containerd[1635]: time="2026-01-14T13:49:21.924167987Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Jan 14 13:49:21.924276 containerd[1635]: time="2026-01-14T13:49:21.924263264Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Jan 14 13:49:21.924428 containerd[1635]: time="2026-01-14T13:49:21.924411220Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Jan 14 13:49:21.924484 containerd[1635]: time="2026-01-14T13:49:21.924473237Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Jan 14 13:49:21.924651 containerd[1635]: time="2026-01-14T13:49:21.924636832Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Jan 14 13:49:21.924701 containerd[1635]: time="2026-01-14T13:49:21.924690923Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Jan 14 13:49:21.924751 containerd[1635]: time="2026-01-14T13:49:21.924732350Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Jan 14 13:49:21.924939 containerd[1635]: time="2026-01-14T13:49:21.924916233Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Jan 14 13:49:21.925019 containerd[1635]: time="2026-01-14T13:49:21.925002384Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Jan 14 13:49:21.925089 containerd[1635]: time="2026-01-14T13:49:21.925074108Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Jan 14 13:49:21.925156 containerd[1635]: time="2026-01-14T13:49:21.925141133Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Jan 14 13:49:21.925248 containerd[1635]: time="2026-01-14T13:49:21.925229408Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jan 14 13:49:21.925324 containerd[1635]: time="2026-01-14T13:49:21.925308705Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jan 14 13:49:21.925384 containerd[1635]: time="2026-01-14T13:49:21.925371903Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jan 14 13:49:21.925444 containerd[1635]: time="2026-01-14T13:49:21.925431986Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jan 14 13:49:21.925484 containerd[1635]: time="2026-01-14T13:49:21.925473884Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Jan 14 13:49:21.925662 containerd[1635]: time="2026-01-14T13:49:21.925646767Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Jan 14 13:49:21.925714 containerd[1635]: time="2026-01-14T13:49:21.925704665Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Jan 14 13:49:21.925754 containerd[1635]: time="2026-01-14T13:49:21.925744920Z" level=info msg="runtime interface created" Jan 14 13:49:21.925787 containerd[1635]: time="2026-01-14T13:49:21.925778353Z" level=info msg="created NRI interface" Jan 14 13:49:21.925920 containerd[1635]: time="2026-01-14T13:49:21.925827834Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Jan 14 13:49:21.925979 containerd[1635]: time="2026-01-14T13:49:21.925968628Z" level=info msg="Connect containerd service" Jan 14 13:49:21.926032 containerd[1635]: time="2026-01-14T13:49:21.926022277Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jan 14 13:49:21.927445 containerd[1635]: time="2026-01-14T13:49:21.927424134Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 14 13:49:21.928124 systemd[1]: Reached target getty.target - Login Prompts. Jan 14 13:49:22.067508 containerd[1635]: time="2026-01-14T13:49:22.067287799Z" level=info msg="Start subscribing containerd event" Jan 14 13:49:22.068922 containerd[1635]: time="2026-01-14T13:49:22.067992244Z" level=info msg="Start recovering state" Jan 14 13:49:22.069431 containerd[1635]: time="2026-01-14T13:49:22.069263266Z" level=info msg="Start event monitor" Jan 14 13:49:22.069816 containerd[1635]: time="2026-01-14T13:49:22.069665427Z" level=info msg="Start cni network conf syncer for default" Jan 14 13:49:22.070349 containerd[1635]: time="2026-01-14T13:49:22.070242284Z" level=info msg="Start streaming server" Jan 14 13:49:22.070722 containerd[1635]: time="2026-01-14T13:49:22.070503091Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jan 14 13:49:22.073213 containerd[1635]: time="2026-01-14T13:49:22.071305809Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Jan 14 13:49:22.073727 containerd[1635]: time="2026-01-14T13:49:22.073495015Z" level=info msg="runtime interface starting up..." Jan 14 13:49:22.073727 containerd[1635]: time="2026-01-14T13:49:22.073716718Z" level=info msg="starting plugins..." Jan 14 13:49:22.073811 containerd[1635]: time="2026-01-14T13:49:22.073741144Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Jan 14 13:49:22.074022 containerd[1635]: time="2026-01-14T13:49:22.072972048Z" level=info msg=serving... address=/run/containerd/containerd.sock Jan 14 13:49:22.074195 systemd[1]: Started containerd.service - containerd container runtime. Jan 14 13:49:22.077910 containerd[1635]: time="2026-01-14T13:49:22.077774042Z" level=info msg="containerd successfully booted in 0.198958s" Jan 14 13:49:22.213260 tar[1631]: linux-amd64/README.md Jan 14 13:49:22.258058 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jan 14 13:49:22.901661 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 13:49:22.915018 systemd[1]: Reached target multi-user.target - Multi-User System. Jan 14 13:49:22.927379 (kubelet)[1739]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 14 13:49:22.927990 systemd[1]: Startup finished in 9.967s (kernel) + 41.021s (initrd) + 17.422s (userspace) = 1min 8.412s. Jan 14 13:49:23.701585 kubelet[1739]: E0114 13:49:23.701358 1739 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 14 13:49:23.706439 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 14 13:49:23.707031 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 14 13:49:23.708272 systemd[1]: kubelet.service: Consumed 1.180s CPU time, 258.5M memory peak. Jan 14 13:49:30.674382 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jan 14 13:49:30.685101 systemd[1]: Started sshd@0-10.0.0.127:22-10.0.0.1:35792.service - OpenSSH per-connection server daemon (10.0.0.1:35792). Jan 14 13:49:31.613492 sshd[1753]: Accepted publickey for core from 10.0.0.1 port 35792 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:49:31.652393 sshd-session[1753]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:49:31.796838 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jan 14 13:49:31.815079 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jan 14 13:49:31.865185 systemd-logind[1617]: New session 1 of user core. Jan 14 13:49:32.175418 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jan 14 13:49:32.212175 systemd[1]: Starting user@500.service - User Manager for UID 500... Jan 14 13:49:32.314313 (systemd)[1759]: pam_unix(systemd-user:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:49:32.365783 systemd-logind[1617]: New session 2 of user core. Jan 14 13:49:33.238905 systemd[1759]: Queued start job for default target default.target. Jan 14 13:49:33.266333 systemd[1759]: Created slice app.slice - User Application Slice. Jan 14 13:49:33.266382 systemd[1759]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of User's Temporary Directories. Jan 14 13:49:33.266401 systemd[1759]: Reached target paths.target - Paths. Jan 14 13:49:33.266477 systemd[1759]: Reached target timers.target - Timers. Jan 14 13:49:33.275275 systemd[1759]: Starting dbus.socket - D-Bus User Message Bus Socket... Jan 14 13:49:33.280837 systemd[1759]: Starting systemd-tmpfiles-setup.service - Create User Files and Directories... Jan 14 13:49:33.455307 systemd[1759]: Finished systemd-tmpfiles-setup.service - Create User Files and Directories. Jan 14 13:49:33.472333 systemd[1759]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jan 14 13:49:33.472867 systemd[1759]: Reached target sockets.target - Sockets. Jan 14 13:49:33.472927 systemd[1759]: Reached target basic.target - Basic System. Jan 14 13:49:33.473133 systemd[1759]: Reached target default.target - Main User Target. Jan 14 13:49:33.473188 systemd[1759]: Startup finished in 1.009s. Jan 14 13:49:33.481298 systemd[1]: Started user@500.service - User Manager for UID 500. Jan 14 13:49:33.494392 systemd[1]: Started session-1.scope - Session 1 of User core. Jan 14 13:49:33.617437 systemd[1]: Started sshd@1-10.0.0.127:22-10.0.0.1:35794.service - OpenSSH per-connection server daemon (10.0.0.1:35794). Jan 14 13:49:33.855849 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jan 14 13:49:33.896401 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 13:49:34.013293 sshd[1773]: Accepted publickey for core from 10.0.0.1 port 35794 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:49:34.029486 sshd-session[1773]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:49:34.100153 systemd-logind[1617]: New session 3 of user core. Jan 14 13:49:34.113282 systemd[1]: Started session-3.scope - Session 3 of User core. Jan 14 13:49:34.288408 sshd[1780]: Connection closed by 10.0.0.1 port 35794 Jan 14 13:49:34.289400 sshd-session[1773]: pam_unix(sshd:session): session closed for user core Jan 14 13:49:34.325500 systemd[1]: sshd@1-10.0.0.127:22-10.0.0.1:35794.service: Deactivated successfully. Jan 14 13:49:34.331104 systemd[1]: session-3.scope: Deactivated successfully. Jan 14 13:49:34.346899 systemd-logind[1617]: Session 3 logged out. Waiting for processes to exit. Jan 14 13:49:34.368148 systemd[1]: Started sshd@2-10.0.0.127:22-10.0.0.1:35796.service - OpenSSH per-connection server daemon (10.0.0.1:35796). Jan 14 13:49:34.375248 systemd-logind[1617]: Removed session 3. Jan 14 13:49:34.711408 sshd[1786]: Accepted publickey for core from 10.0.0.1 port 35796 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:49:34.727502 sshd-session[1786]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:49:34.786801 systemd-logind[1617]: New session 4 of user core. Jan 14 13:49:34.803926 systemd[1]: Started session-4.scope - Session 4 of User core. Jan 14 13:49:34.807270 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 13:49:34.889871 (kubelet)[1794]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 14 13:49:34.943427 sshd[1796]: Connection closed by 10.0.0.1 port 35796 Jan 14 13:49:34.942137 sshd-session[1786]: pam_unix(sshd:session): session closed for user core Jan 14 13:49:34.993172 systemd[1]: Started sshd@3-10.0.0.127:22-10.0.0.1:49206.service - OpenSSH per-connection server daemon (10.0.0.1:49206). Jan 14 13:49:34.994481 systemd[1]: sshd@2-10.0.0.127:22-10.0.0.1:35796.service: Deactivated successfully. Jan 14 13:49:35.007916 systemd[1]: session-4.scope: Deactivated successfully. Jan 14 13:49:35.031413 systemd-logind[1617]: Session 4 logged out. Waiting for processes to exit. Jan 14 13:49:35.051296 systemd-logind[1617]: Removed session 4. Jan 14 13:49:35.421263 sshd[1804]: Accepted publickey for core from 10.0.0.1 port 49206 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:49:35.427276 sshd-session[1804]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:49:35.478166 systemd-logind[1617]: New session 5 of user core. Jan 14 13:49:35.493782 systemd[1]: Started session-5.scope - Session 5 of User core. Jan 14 13:49:35.503488 kubelet[1794]: E0114 13:49:35.501265 1794 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 14 13:49:35.513424 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 14 13:49:35.514152 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 14 13:49:35.515316 systemd[1]: kubelet.service: Consumed 728ms CPU time, 110.4M memory peak. Jan 14 13:49:35.677475 sshd[1814]: Connection closed by 10.0.0.1 port 49206 Jan 14 13:49:35.678483 sshd-session[1804]: pam_unix(sshd:session): session closed for user core Jan 14 13:49:35.711191 systemd[1]: sshd@3-10.0.0.127:22-10.0.0.1:49206.service: Deactivated successfully. Jan 14 13:49:35.722487 systemd[1]: session-5.scope: Deactivated successfully. Jan 14 13:49:35.743862 systemd-logind[1617]: Session 5 logged out. Waiting for processes to exit. Jan 14 13:49:35.748286 systemd[1]: Started sshd@4-10.0.0.127:22-10.0.0.1:49208.service - OpenSSH per-connection server daemon (10.0.0.1:49208). Jan 14 13:49:35.760214 systemd-logind[1617]: Removed session 5. Jan 14 13:49:36.100468 sshd[1820]: Accepted publickey for core from 10.0.0.1 port 49208 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:49:36.106946 sshd-session[1820]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:49:36.174383 systemd-logind[1617]: New session 6 of user core. Jan 14 13:49:36.202452 systemd[1]: Started session-6.scope - Session 6 of User core. Jan 14 13:49:36.419229 sudo[1825]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jan 14 13:49:36.422861 sudo[1825]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 14 13:49:36.488422 sudo[1825]: pam_unix(sudo:session): session closed for user root Jan 14 13:49:36.521495 sshd[1824]: Connection closed by 10.0.0.1 port 49208 Jan 14 13:49:36.522414 sshd-session[1820]: pam_unix(sshd:session): session closed for user core Jan 14 13:49:36.561912 systemd[1]: Started sshd@5-10.0.0.127:22-10.0.0.1:49222.service - OpenSSH per-connection server daemon (10.0.0.1:49222). Jan 14 13:49:36.572184 systemd[1]: sshd@4-10.0.0.127:22-10.0.0.1:49208.service: Deactivated successfully. Jan 14 13:49:36.586921 systemd[1]: session-6.scope: Deactivated successfully. Jan 14 13:49:36.603429 systemd-logind[1617]: Session 6 logged out. Waiting for processes to exit. Jan 14 13:49:36.627284 systemd-logind[1617]: Removed session 6. Jan 14 13:49:36.850824 sshd[1829]: Accepted publickey for core from 10.0.0.1 port 49222 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:49:36.867400 sshd-session[1829]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:49:36.942783 systemd-logind[1617]: New session 7 of user core. Jan 14 13:49:36.959343 systemd[1]: Started session-7.scope - Session 7 of User core. Jan 14 13:49:37.163387 sudo[1838]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jan 14 13:49:37.164814 sudo[1838]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 14 13:49:37.232256 sudo[1838]: pam_unix(sudo:session): session closed for user root Jan 14 13:49:37.295370 sudo[1837]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Jan 14 13:49:37.296844 sudo[1837]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 14 13:49:37.427962 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 14 13:49:37.985000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Jan 14 13:49:38.002278 augenrules[1862]: No rules Jan 14 13:49:38.019173 kernel: kauditd_printk_skb: 68 callbacks suppressed Jan 14 13:49:38.019249 kernel: audit: type=1305 audit(1768398577.985:216): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Jan 14 13:49:38.017977 systemd[1]: audit-rules.service: Deactivated successfully. Jan 14 13:49:38.020206 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 14 13:49:38.054381 sudo[1837]: pam_unix(sudo:session): session closed for user root Jan 14 13:49:38.080503 sshd[1836]: Connection closed by 10.0.0.1 port 49222 Jan 14 13:49:38.088267 sshd-session[1829]: pam_unix(sshd:session): session closed for user core Jan 14 13:49:38.113157 kernel: audit: type=1300 audit(1768398577.985:216): arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffd62b11630 a2=420 a3=0 items=0 ppid=1843 pid=1862 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:37.985000 audit[1862]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffd62b11630 a2=420 a3=0 items=0 ppid=1843 pid=1862 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:38.208465 kernel: audit: type=1327 audit(1768398577.985:216): proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 14 13:49:37.985000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 14 13:49:38.257909 kernel: audit: type=1130 audit(1768398578.024:217): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:38.024000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:38.290256 systemd[1]: sshd@5-10.0.0.127:22-10.0.0.1:49222.service: Deactivated successfully. Jan 14 13:49:38.024000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:38.321473 systemd[1]: session-7.scope: Deactivated successfully. Jan 14 13:49:38.342773 systemd-logind[1617]: Session 7 logged out. Waiting for processes to exit. Jan 14 13:49:38.348354 kernel: audit: type=1131 audit(1768398578.024:218): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:38.352372 kernel: audit: type=1106 audit(1768398578.051:219): pid=1837 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 13:49:38.051000 audit[1837]: USER_END pid=1837 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 13:49:38.358965 systemd[1]: Started sshd@6-10.0.0.127:22-10.0.0.1:49234.service - OpenSSH per-connection server daemon (10.0.0.1:49234). Jan 14 13:49:38.377474 systemd-logind[1617]: Removed session 7. Jan 14 13:49:38.399384 kernel: audit: type=1104 audit(1768398578.058:220): pid=1837 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 13:49:38.058000 audit[1837]: CRED_DISP pid=1837 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 13:49:38.456440 kernel: audit: type=1106 audit(1768398578.109:221): pid=1829 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:49:38.456828 kernel: audit: type=1104 audit(1768398578.109:222): pid=1829 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:49:38.456862 kernel: audit: type=1131 audit(1768398578.295:223): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@5-10.0.0.127:22-10.0.0.1:49222 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:38.109000 audit[1829]: USER_END pid=1829 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:49:38.109000 audit[1829]: CRED_DISP pid=1829 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:49:38.295000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@5-10.0.0.127:22-10.0.0.1:49222 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:38.362000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-10.0.0.127:22-10.0.0.1:49234 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:39.051405 sshd[1871]: Accepted publickey for core from 10.0.0.1 port 49234 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:49:39.047000 audit[1871]: USER_ACCT pid=1871 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:49:39.063000 audit[1871]: CRED_ACQ pid=1871 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:49:39.066000 audit[1871]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fffd8e84b60 a2=3 a3=0 items=0 ppid=1 pid=1871 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=8 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:39.066000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:49:39.068831 sshd-session[1871]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:49:39.112265 systemd-logind[1617]: New session 8 of user core. Jan 14 13:49:39.393479 systemd[1]: Started session-8.scope - Session 8 of User core. Jan 14 13:49:39.439000 audit[1871]: USER_START pid=1871 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:49:39.448000 audit[1875]: CRED_ACQ pid=1875 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:49:39.523000 audit[1876]: USER_ACCT pid=1876 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_unix,pam_faillock acct="core" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 13:49:39.527000 audit[1876]: CRED_REFR pid=1876 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 13:49:39.527291 sudo[1876]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jan 14 13:49:39.528486 sudo[1876]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 14 13:49:39.527000 audit[1876]: USER_START pid=1876 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 13:49:41.641879 systemd[1]: Starting docker.service - Docker Application Container Engine... Jan 14 13:49:41.706468 (dockerd)[1898]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jan 14 13:49:43.907490 dockerd[1898]: time="2026-01-14T13:49:43.904979956Z" level=info msg="Starting up" Jan 14 13:49:43.935979 dockerd[1898]: time="2026-01-14T13:49:43.934277470Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Jan 14 13:49:44.544868 dockerd[1898]: time="2026-01-14T13:49:44.543420760Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Jan 14 13:49:45.385470 dockerd[1898]: time="2026-01-14T13:49:45.382254959Z" level=info msg="Loading containers: start." Jan 14 13:49:45.705359 kernel: Initializing XFRM netlink socket Jan 14 13:49:45.719947 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jan 14 13:49:45.796465 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 13:49:50.468909 kernel: kauditd_printk_skb: 11 callbacks suppressed Jan 14 13:49:50.485958 kernel: audit: type=1325 audit(1768398590.430:233): table=nat:2 family=2 entries=2 op=nft_register_chain pid=1956 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:50.430000 audit[1956]: NETFILTER_CFG table=nat:2 family=2 entries=2 op=nft_register_chain pid=1956 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:50.430000 audit[1956]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffe64a3f290 a2=0 a3=0 items=0 ppid=1898 pid=1956 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:50.430000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Jan 14 13:49:50.631488 kernel: audit: type=1300 audit(1768398590.430:233): arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffe64a3f290 a2=0 a3=0 items=0 ppid=1898 pid=1956 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:50.632481 kernel: audit: type=1327 audit(1768398590.430:233): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Jan 14 13:49:50.632843 kernel: audit: type=1325 audit(1768398590.467:234): table=filter:3 family=2 entries=2 op=nft_register_chain pid=1958 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:50.467000 audit[1958]: NETFILTER_CFG table=filter:3 family=2 entries=2 op=nft_register_chain pid=1958 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:50.467000 audit[1958]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffc72264530 a2=0 a3=0 items=0 ppid=1898 pid=1958 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:50.759003 kernel: audit: type=1300 audit(1768398590.467:234): arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffc72264530 a2=0 a3=0 items=0 ppid=1898 pid=1958 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:50.759965 kernel: audit: type=1327 audit(1768398590.467:234): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Jan 14 13:49:50.467000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Jan 14 13:49:50.792975 kernel: audit: type=1325 audit(1768398590.518:235): table=filter:4 family=2 entries=1 op=nft_register_chain pid=1960 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:50.518000 audit[1960]: NETFILTER_CFG table=filter:4 family=2 entries=1 op=nft_register_chain pid=1960 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:50.833046 kernel: audit: type=1300 audit(1768398590.518:235): arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd7d5b9430 a2=0 a3=0 items=0 ppid=1898 pid=1960 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:50.518000 audit[1960]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd7d5b9430 a2=0 a3=0 items=0 ppid=1898 pid=1960 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:50.518000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Jan 14 13:49:50.968505 kernel: audit: type=1327 audit(1768398590.518:235): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Jan 14 13:49:50.972421 kernel: audit: type=1325 audit(1768398590.579:236): table=filter:5 family=2 entries=1 op=nft_register_chain pid=1962 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:50.579000 audit[1962]: NETFILTER_CFG table=filter:5 family=2 entries=1 op=nft_register_chain pid=1962 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:50.579000 audit[1962]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff18646890 a2=0 a3=0 items=0 ppid=1898 pid=1962 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:50.579000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D425249444745 Jan 14 13:49:50.626000 audit[1964]: NETFILTER_CFG table=filter:6 family=2 entries=1 op=nft_register_chain pid=1964 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:50.626000 audit[1964]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffc46f0f4f0 a2=0 a3=0 items=0 ppid=1898 pid=1964 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:50.626000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D4354 Jan 14 13:49:50.663000 audit[1966]: NETFILTER_CFG table=filter:7 family=2 entries=1 op=nft_register_chain pid=1966 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:50.663000 audit[1966]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7fffa2088560 a2=0 a3=0 items=0 ppid=1898 pid=1966 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:50.663000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 14 13:49:50.701000 audit[1968]: NETFILTER_CFG table=filter:8 family=2 entries=1 op=nft_register_chain pid=1968 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:50.701000 audit[1968]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffe90202f30 a2=0 a3=0 items=0 ppid=1898 pid=1968 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:50.701000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 14 13:49:50.751000 audit[1970]: NETFILTER_CFG table=nat:9 family=2 entries=2 op=nft_register_chain pid=1970 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:50.751000 audit[1970]: SYSCALL arch=c000003e syscall=46 success=yes exit=384 a0=3 a1=7ffe4af231c0 a2=0 a3=0 items=0 ppid=1898 pid=1970 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:50.751000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Jan 14 13:49:51.226000 audit[1977]: NETFILTER_CFG table=nat:10 family=2 entries=2 op=nft_register_chain pid=1977 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:51.226000 audit[1977]: SYSCALL arch=c000003e syscall=46 success=yes exit=472 a0=3 a1=7ffe7a537160 a2=0 a3=0 items=0 ppid=1898 pid=1977 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:51.226000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003132372E302E302E302F38 Jan 14 13:49:51.263811 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 13:49:51.265000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:51.273000 audit[1979]: NETFILTER_CFG table=filter:11 family=2 entries=2 op=nft_register_chain pid=1979 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:51.273000 audit[1979]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffe77d4c300 a2=0 a3=0 items=0 ppid=1898 pid=1979 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:51.273000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D464F5257415244 Jan 14 13:49:51.335000 audit[1983]: NETFILTER_CFG table=filter:12 family=2 entries=1 op=nft_register_rule pid=1983 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:51.335000 audit[1983]: SYSCALL arch=c000003e syscall=46 success=yes exit=236 a0=3 a1=7ffd968dea90 a2=0 a3=0 items=0 ppid=1898 pid=1983 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:51.335000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D425249444745 Jan 14 13:49:51.358943 (kubelet)[1980]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 14 13:49:51.368000 audit[1985]: NETFILTER_CFG table=filter:13 family=2 entries=1 op=nft_register_rule pid=1985 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:51.368000 audit[1985]: SYSCALL arch=c000003e syscall=46 success=yes exit=248 a0=3 a1=7ffc1985d740 a2=0 a3=0 items=0 ppid=1898 pid=1985 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:51.368000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 14 13:49:51.404000 audit[1987]: NETFILTER_CFG table=filter:14 family=2 entries=1 op=nft_register_rule pid=1987 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:51.404000 audit[1987]: SYSCALL arch=c000003e syscall=46 success=yes exit=232 a0=3 a1=7fffdf71d180 a2=0 a3=0 items=0 ppid=1898 pid=1987 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:51.404000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D4354 Jan 14 13:49:52.239000 audit[2022]: NETFILTER_CFG table=nat:15 family=10 entries=2 op=nft_register_chain pid=2022 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:52.239000 audit[2022]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffcb2c99bb0 a2=0 a3=0 items=0 ppid=1898 pid=2022 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:52.239000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Jan 14 13:49:52.293000 audit[2024]: NETFILTER_CFG table=filter:16 family=10 entries=2 op=nft_register_chain pid=2024 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:52.293000 audit[2024]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffe0a962580 a2=0 a3=0 items=0 ppid=1898 pid=2024 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:52.293000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Jan 14 13:49:52.376000 audit[2027]: NETFILTER_CFG table=filter:17 family=10 entries=1 op=nft_register_chain pid=2027 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:52.376000 audit[2027]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd66702f90 a2=0 a3=0 items=0 ppid=1898 pid=2027 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:52.376000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Jan 14 13:49:52.437000 audit[2029]: NETFILTER_CFG table=filter:18 family=10 entries=1 op=nft_register_chain pid=2029 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:52.437000 audit[2029]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffef652e6d0 a2=0 a3=0 items=0 ppid=1898 pid=2029 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:52.437000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D425249444745 Jan 14 13:49:52.509000 audit[2031]: NETFILTER_CFG table=filter:19 family=10 entries=1 op=nft_register_chain pid=2031 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:52.509000 audit[2031]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffc9f0d5650 a2=0 a3=0 items=0 ppid=1898 pid=2031 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:52.509000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D4354 Jan 14 13:49:52.561000 audit[2033]: NETFILTER_CFG table=filter:20 family=10 entries=1 op=nft_register_chain pid=2033 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:52.561000 audit[2033]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffcf2040890 a2=0 a3=0 items=0 ppid=1898 pid=2033 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:52.561000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 14 13:49:52.597000 audit[2035]: NETFILTER_CFG table=filter:21 family=10 entries=1 op=nft_register_chain pid=2035 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:52.597000 audit[2035]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffc7838d6b0 a2=0 a3=0 items=0 ppid=1898 pid=2035 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:52.597000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 14 13:49:52.640000 audit[2037]: NETFILTER_CFG table=nat:22 family=10 entries=2 op=nft_register_chain pid=2037 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:52.640000 audit[2037]: SYSCALL arch=c000003e syscall=46 success=yes exit=384 a0=3 a1=7ffd7ed291e0 a2=0 a3=0 items=0 ppid=1898 pid=2037 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:52.640000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Jan 14 13:49:52.681000 audit[2039]: NETFILTER_CFG table=nat:23 family=10 entries=2 op=nft_register_chain pid=2039 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:52.681000 audit[2039]: SYSCALL arch=c000003e syscall=46 success=yes exit=484 a0=3 a1=7ffcef7b7bd0 a2=0 a3=0 items=0 ppid=1898 pid=2039 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:52.681000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003A3A312F313238 Jan 14 13:49:52.730000 audit[2041]: NETFILTER_CFG table=filter:24 family=10 entries=2 op=nft_register_chain pid=2041 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:52.730000 audit[2041]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffc937fc290 a2=0 a3=0 items=0 ppid=1898 pid=2041 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:52.730000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D464F5257415244 Jan 14 13:49:52.765000 audit[2043]: NETFILTER_CFG table=filter:25 family=10 entries=1 op=nft_register_rule pid=2043 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:52.765000 audit[2043]: SYSCALL arch=c000003e syscall=46 success=yes exit=236 a0=3 a1=7ffd72e6aaf0 a2=0 a3=0 items=0 ppid=1898 pid=2043 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:52.765000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D425249444745 Jan 14 13:49:52.813000 audit[2045]: NETFILTER_CFG table=filter:26 family=10 entries=1 op=nft_register_rule pid=2045 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:52.813000 audit[2045]: SYSCALL arch=c000003e syscall=46 success=yes exit=248 a0=3 a1=7fff71b18c80 a2=0 a3=0 items=0 ppid=1898 pid=2045 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:52.813000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 14 13:49:52.853000 audit[2047]: NETFILTER_CFG table=filter:27 family=10 entries=1 op=nft_register_rule pid=2047 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:52.853000 audit[2047]: SYSCALL arch=c000003e syscall=46 success=yes exit=232 a0=3 a1=7ffdc14c6880 a2=0 a3=0 items=0 ppid=1898 pid=2047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:52.853000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D4354 Jan 14 13:49:52.945000 audit[2052]: NETFILTER_CFG table=filter:28 family=2 entries=1 op=nft_register_chain pid=2052 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:52.945000 audit[2052]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffe5e6b5290 a2=0 a3=0 items=0 ppid=1898 pid=2052 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:52.945000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Jan 14 13:49:53.023000 audit[2054]: NETFILTER_CFG table=filter:29 family=2 entries=1 op=nft_register_rule pid=2054 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:53.023000 audit[2054]: SYSCALL arch=c000003e syscall=46 success=yes exit=212 a0=3 a1=7ffe464cff60 a2=0 a3=0 items=0 ppid=1898 pid=2054 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:53.023000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E Jan 14 13:49:53.108000 audit[2056]: NETFILTER_CFG table=filter:30 family=2 entries=1 op=nft_register_rule pid=2056 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:53.108000 audit[2056]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7ffeea584390 a2=0 a3=0 items=0 ppid=1898 pid=2056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:53.108000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Jan 14 13:49:53.142000 audit[2058]: NETFILTER_CFG table=filter:31 family=10 entries=1 op=nft_register_chain pid=2058 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:53.142000 audit[2058]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffe26d9e040 a2=0 a3=0 items=0 ppid=1898 pid=2058 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:53.142000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Jan 14 13:49:53.173000 audit[2060]: NETFILTER_CFG table=filter:32 family=10 entries=1 op=nft_register_rule pid=2060 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:53.173000 audit[2060]: SYSCALL arch=c000003e syscall=46 success=yes exit=212 a0=3 a1=7ffc72eb8df0 a2=0 a3=0 items=0 ppid=1898 pid=2060 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:53.173000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E Jan 14 13:49:53.195000 audit[2062]: NETFILTER_CFG table=filter:33 family=10 entries=1 op=nft_register_rule pid=2062 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:49:53.195000 audit[2062]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7ffd791ce630 a2=0 a3=0 items=0 ppid=1898 pid=2062 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:53.195000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Jan 14 13:49:53.344099 kubelet[1980]: E0114 13:49:53.339819 1980 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 14 13:49:53.358111 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 14 13:49:53.359427 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 14 13:49:53.364000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 13:49:53.368345 systemd[1]: kubelet.service: Consumed 4.906s CPU time, 110.3M memory peak. Jan 14 13:49:53.386000 audit[2068]: NETFILTER_CFG table=nat:34 family=2 entries=2 op=nft_register_chain pid=2068 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:53.386000 audit[2068]: SYSCALL arch=c000003e syscall=46 success=yes exit=520 a0=3 a1=7fffb6cbba00 a2=0 a3=0 items=0 ppid=1898 pid=2068 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:53.386000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900504F5354524F5554494E47002D73003137322E31372E302E302F31360000002D6F00646F636B657230002D6A004D415351554552414445 Jan 14 13:49:53.414000 audit[2070]: NETFILTER_CFG table=nat:35 family=2 entries=1 op=nft_register_rule pid=2070 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:53.414000 audit[2070]: SYSCALL arch=c000003e syscall=46 success=yes exit=288 a0=3 a1=7ffe7af4ba00 a2=0 a3=0 items=0 ppid=1898 pid=2070 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:53.414000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900444F434B4552002D6900646F636B657230002D6A0052455455524E Jan 14 13:49:53.536000 audit[2078]: NETFILTER_CFG table=filter:36 family=2 entries=1 op=nft_register_rule pid=2078 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:53.536000 audit[2078]: SYSCALL arch=c000003e syscall=46 success=yes exit=300 a0=3 a1=7ffd240d5e30 a2=0 a3=0 items=0 ppid=1898 pid=2078 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:53.536000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D464F5257415244002D6900646F636B657230002D6A00414343455054 Jan 14 13:49:53.644000 audit[2084]: NETFILTER_CFG table=filter:37 family=2 entries=1 op=nft_register_rule pid=2084 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:53.644000 audit[2084]: SYSCALL arch=c000003e syscall=46 success=yes exit=376 a0=3 a1=7ffd7f4c3a30 a2=0 a3=0 items=0 ppid=1898 pid=2084 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:53.644000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45520000002D6900646F636B657230002D6F00646F636B657230002D6A0044524F50 Jan 14 13:49:53.669000 audit[2086]: NETFILTER_CFG table=filter:38 family=2 entries=1 op=nft_register_rule pid=2086 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:53.669000 audit[2086]: SYSCALL arch=c000003e syscall=46 success=yes exit=512 a0=3 a1=7ffe86123610 a2=0 a3=0 items=0 ppid=1898 pid=2086 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:53.669000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D4354002D6F00646F636B657230002D6D00636F6E6E747261636B002D2D637473746174650052454C415445442C45535441424C4953484544002D6A00414343455054 Jan 14 13:49:53.703000 audit[2088]: NETFILTER_CFG table=filter:39 family=2 entries=1 op=nft_register_rule pid=2088 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:53.703000 audit[2088]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffd971e4fb0 a2=0 a3=0 items=0 ppid=1898 pid=2088 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:53.703000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D425249444745002D6F00646F636B657230002D6A00444F434B4552 Jan 14 13:49:53.729000 audit[2090]: NETFILTER_CFG table=filter:40 family=2 entries=1 op=nft_register_rule pid=2090 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:53.729000 audit[2090]: SYSCALL arch=c000003e syscall=46 success=yes exit=428 a0=3 a1=7ffdd5f03230 a2=0 a3=0 items=0 ppid=1898 pid=2090 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:53.729000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D49534F4C4154494F4E2D53544147452D31002D6900646F636B6572300000002D6F00646F636B657230002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 14 13:49:53.757000 audit[2092]: NETFILTER_CFG table=filter:41 family=2 entries=1 op=nft_register_rule pid=2092 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:49:53.757000 audit[2092]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffd7e5c4b80 a2=0 a3=0 items=0 ppid=1898 pid=2092 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:49:53.757000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4900444F434B45522D49534F4C4154494F4E2D53544147452D32002D6F00646F636B657230002D6A0044524F50 Jan 14 13:49:53.762500 systemd-networkd[1527]: docker0: Link UP Jan 14 13:49:53.790386 dockerd[1898]: time="2026-01-14T13:49:53.789907065Z" level=info msg="Loading containers: done." Jan 14 13:49:53.988075 dockerd[1898]: time="2026-01-14T13:49:53.986077320Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jan 14 13:49:53.988075 dockerd[1898]: time="2026-01-14T13:49:53.987006768Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Jan 14 13:49:53.988075 dockerd[1898]: time="2026-01-14T13:49:53.987304575Z" level=info msg="Initializing buildkit" Jan 14 13:49:54.538097 dockerd[1898]: time="2026-01-14T13:49:54.536498035Z" level=info msg="Completed buildkit initialization" Jan 14 13:49:54.590000 dockerd[1898]: time="2026-01-14T13:49:54.588016868Z" level=info msg="Daemon has completed initialization" Jan 14 13:49:54.597387 dockerd[1898]: time="2026-01-14T13:49:54.592972058Z" level=info msg="API listen on /run/docker.sock" Jan 14 13:49:54.600490 systemd[1]: Started docker.service - Docker Application Container Engine. Jan 14 13:49:54.602000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=docker comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:49:58.260386 containerd[1635]: time="2026-01-14T13:49:58.259335103Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.3\"" Jan 14 13:50:01.105893 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount112912108.mount: Deactivated successfully. Jan 14 13:50:03.615323 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Jan 14 13:50:03.637829 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 13:50:04.723000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:50:04.723441 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 13:50:04.738337 kernel: kauditd_printk_skb: 113 callbacks suppressed Jan 14 13:50:04.738962 kernel: audit: type=1130 audit(1768398604.723:276): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:50:04.863361 (kubelet)[2204]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 14 13:50:06.117924 kubelet[2204]: E0114 13:50:06.117773 2204 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 14 13:50:06.129444 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 14 13:50:06.130920 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 14 13:50:06.133000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 13:50:06.134501 systemd[1]: kubelet.service: Consumed 2.384s CPU time, 109.8M memory peak. Jan 14 13:50:06.194053 kernel: audit: type=1131 audit(1768398606.133:277): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 13:50:06.567416 update_engine[1620]: I20260114 13:50:06.553061 1620 update_attempter.cc:509] Updating boot flags... Jan 14 13:50:12.641998 containerd[1635]: time="2026-01-14T13:50:12.640389529Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.34.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:12.648033 containerd[1635]: time="2026-01-14T13:50:12.646309671Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.34.3: active requests=0, bytes read=27025159" Jan 14 13:50:12.652451 containerd[1635]: time="2026-01-14T13:50:12.651908374Z" level=info msg="ImageCreate event name:\"sha256:aa27095f5619377172f3d59289ccb2ba567ebea93a736d1705be068b2c030b0c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:12.672959 containerd[1635]: time="2026-01-14T13:50:12.670375300Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:5af1030676ceca025742ef5e73a504d11b59be0e5551cdb8c9cf0d3c1231b460\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:12.675260 containerd[1635]: time="2026-01-14T13:50:12.674335216Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.34.3\" with image id \"sha256:aa27095f5619377172f3d59289ccb2ba567ebea93a736d1705be068b2c030b0c\", repo tag \"registry.k8s.io/kube-apiserver:v1.34.3\", repo digest \"registry.k8s.io/kube-apiserver@sha256:5af1030676ceca025742ef5e73a504d11b59be0e5551cdb8c9cf0d3c1231b460\", size \"27064672\" in 14.411357331s" Jan 14 13:50:12.675260 containerd[1635]: time="2026-01-14T13:50:12.674965266Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.34.3\" returns image reference \"sha256:aa27095f5619377172f3d59289ccb2ba567ebea93a736d1705be068b2c030b0c\"" Jan 14 13:50:12.686054 containerd[1635]: time="2026-01-14T13:50:12.685931964Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.3\"" Jan 14 13:50:16.401991 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Jan 14 13:50:16.447860 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 13:50:17.722010 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 13:50:17.722000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:50:17.789482 kernel: audit: type=1130 audit(1768398617.722:278): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:50:17.817405 (kubelet)[2241]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 14 13:50:18.440448 kubelet[2241]: E0114 13:50:18.439029 2241 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 14 13:50:18.470987 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 14 13:50:18.473971 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 14 13:50:18.475000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 13:50:18.476430 systemd[1]: kubelet.service: Consumed 1.460s CPU time, 110.8M memory peak. Jan 14 13:50:18.531039 kernel: audit: type=1131 audit(1768398618.475:279): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 13:50:21.836968 containerd[1635]: time="2026-01-14T13:50:21.836035246Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.34.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:21.841234 containerd[1635]: time="2026-01-14T13:50:21.840828368Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.34.3: active requests=0, bytes read=21154285" Jan 14 13:50:21.845402 containerd[1635]: time="2026-01-14T13:50:21.845347752Z" level=info msg="ImageCreate event name:\"sha256:5826b25d990d7d314d236c8d128f43e443583891f5cdffa7bf8bca50ae9e0942\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:21.863251 containerd[1635]: time="2026-01-14T13:50:21.863027412Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:716a210d31ee5e27053ea0e1a3a3deb4910791a85ba4b1120410b5a4cbcf1954\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:21.867437 containerd[1635]: time="2026-01-14T13:50:21.867406440Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.34.3\" with image id \"sha256:5826b25d990d7d314d236c8d128f43e443583891f5cdffa7bf8bca50ae9e0942\", repo tag \"registry.k8s.io/kube-controller-manager:v1.34.3\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:716a210d31ee5e27053ea0e1a3a3deb4910791a85ba4b1120410b5a4cbcf1954\", size \"22819474\" in 9.180510506s" Jan 14 13:50:21.867869 containerd[1635]: time="2026-01-14T13:50:21.867507849Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.34.3\" returns image reference \"sha256:5826b25d990d7d314d236c8d128f43e443583891f5cdffa7bf8bca50ae9e0942\"" Jan 14 13:50:21.897350 containerd[1635]: time="2026-01-14T13:50:21.896293699Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.3\"" Jan 14 13:50:26.205368 containerd[1635]: time="2026-01-14T13:50:26.203441237Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.34.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:26.212288 containerd[1635]: time="2026-01-14T13:50:26.210468939Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.34.3: active requests=0, bytes read=15717792" Jan 14 13:50:26.221415 containerd[1635]: time="2026-01-14T13:50:26.220908389Z" level=info msg="ImageCreate event name:\"sha256:aec12dadf56dd45659a682b94571f115a1be02ee4a262b3b5176394f5c030c78\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:26.231504 containerd[1635]: time="2026-01-14T13:50:26.231304989Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:f9a9bc7948fd804ef02255fe82ac2e85d2a66534bae2fe1348c14849260a1fe2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:26.236465 containerd[1635]: time="2026-01-14T13:50:26.236246791Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.34.3\" with image id \"sha256:aec12dadf56dd45659a682b94571f115a1be02ee4a262b3b5176394f5c030c78\", repo tag \"registry.k8s.io/kube-scheduler:v1.34.3\", repo digest \"registry.k8s.io/kube-scheduler@sha256:f9a9bc7948fd804ef02255fe82ac2e85d2a66534bae2fe1348c14849260a1fe2\", size \"17382979\" in 4.339588023s" Jan 14 13:50:26.236465 containerd[1635]: time="2026-01-14T13:50:26.236457734Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.34.3\" returns image reference \"sha256:aec12dadf56dd45659a682b94571f115a1be02ee4a262b3b5176394f5c030c78\"" Jan 14 13:50:26.244979 containerd[1635]: time="2026-01-14T13:50:26.244072330Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.3\"" Jan 14 13:50:28.606338 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Jan 14 13:50:28.617293 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 13:50:29.167409 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 13:50:29.167000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:50:29.220960 kernel: audit: type=1130 audit(1768398629.167:280): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:50:29.229486 (kubelet)[2266]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 14 13:50:29.603035 kubelet[2266]: E0114 13:50:29.601461 2266 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 14 13:50:29.611372 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 14 13:50:29.612873 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 14 13:50:29.614000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 13:50:29.616464 systemd[1]: kubelet.service: Consumed 726ms CPU time, 110.3M memory peak. Jan 14 13:50:29.664854 kernel: audit: type=1131 audit(1768398629.614:281): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 13:50:29.679909 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3236257311.mount: Deactivated successfully. Jan 14 13:50:32.565737 containerd[1635]: time="2026-01-14T13:50:32.564874899Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.34.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:32.569143 containerd[1635]: time="2026-01-14T13:50:32.569103051Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.34.3: active requests=0, bytes read=25961571" Jan 14 13:50:32.577418 containerd[1635]: time="2026-01-14T13:50:32.577351187Z" level=info msg="ImageCreate event name:\"sha256:36eef8e07bdd6abdc2bbf44041e49480fe499a3cedb0ae054b50daa1a35cf691\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:32.585074 containerd[1635]: time="2026-01-14T13:50:32.584444190Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:7298ab89a103523d02ff4f49bedf9359710af61df92efdc07bac873064f03ed6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:32.587094 containerd[1635]: time="2026-01-14T13:50:32.586474166Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.34.3\" with image id \"sha256:36eef8e07bdd6abdc2bbf44041e49480fe499a3cedb0ae054b50daa1a35cf691\", repo tag \"registry.k8s.io/kube-proxy:v1.34.3\", repo digest \"registry.k8s.io/kube-proxy@sha256:7298ab89a103523d02ff4f49bedf9359710af61df92efdc07bac873064f03ed6\", size \"25964312\" in 6.342189922s" Jan 14 13:50:32.588011 containerd[1635]: time="2026-01-14T13:50:32.587839030Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.34.3\" returns image reference \"sha256:36eef8e07bdd6abdc2bbf44041e49480fe499a3cedb0ae054b50daa1a35cf691\"" Jan 14 13:50:32.593132 containerd[1635]: time="2026-01-14T13:50:32.592353847Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\"" Jan 14 13:50:33.941879 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4038290836.mount: Deactivated successfully. Jan 14 13:50:39.857023 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Jan 14 13:50:39.869413 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 13:50:40.553941 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 13:50:40.555000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:50:40.611992 containerd[1635]: time="2026-01-14T13:50:40.610072046Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:40.617858 kernel: audit: type=1130 audit(1768398640.555:282): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:50:40.617961 containerd[1635]: time="2026-01-14T13:50:40.617121734Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.1: active requests=0, bytes read=22363711" Jan 14 13:50:40.627872 containerd[1635]: time="2026-01-14T13:50:40.627104793Z" level=info msg="ImageCreate event name:\"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:40.629147 (kubelet)[2339]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 14 13:50:40.653897 containerd[1635]: time="2026-01-14T13:50:40.650925592Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:40.653897 containerd[1635]: time="2026-01-14T13:50:40.653469313Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.1\" with image id \"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:e8c262566636e6bc340ece6473b0eed193cad045384401529721ddbe6463d31c\", size \"22384805\" in 8.06049482s" Jan 14 13:50:40.661048 containerd[1635]: time="2026-01-14T13:50:40.653508987Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.1\" returns image reference \"sha256:52546a367cc9e0d924aa3b190596a9167fa6e53245023b5b5baf0f07e5443969\"" Jan 14 13:50:40.667963 containerd[1635]: time="2026-01-14T13:50:40.666859335Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\"" Jan 14 13:50:41.034178 kubelet[2339]: E0114 13:50:41.033913 2339 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 14 13:50:41.045467 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 14 13:50:41.047056 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 14 13:50:41.049000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 13:50:41.050029 systemd[1]: kubelet.service: Consumed 879ms CPU time, 110.7M memory peak. Jan 14 13:50:41.097147 kernel: audit: type=1131 audit(1768398641.049:283): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 13:50:41.495038 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2016875939.mount: Deactivated successfully. Jan 14 13:50:41.539938 containerd[1635]: time="2026-01-14T13:50:41.539878980Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:41.549008 containerd[1635]: time="2026-01-14T13:50:41.548698009Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10.1: active requests=0, bytes read=316649" Jan 14 13:50:41.558939 containerd[1635]: time="2026-01-14T13:50:41.558443186Z" level=info msg="ImageCreate event name:\"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:41.566818 containerd[1635]: time="2026-01-14T13:50:41.566493972Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:41.578428 containerd[1635]: time="2026-01-14T13:50:41.575902411Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10.1\" with image id \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\", repo tag \"registry.k8s.io/pause:3.10.1\", repo digest \"registry.k8s.io/pause@sha256:278fb9dbcca9518083ad1e11276933a2e96f23de604a3a08cc3c80002767d24c\", size \"320448\" in 908.833514ms" Jan 14 13:50:41.578428 containerd[1635]: time="2026-01-14T13:50:41.576087687Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10.1\" returns image reference \"sha256:cd073f4c5f6a8e9dc6f3125ba00cf60819cae95c1ec84a1f146ee4a9cf9e803f\"" Jan 14 13:50:41.586965 containerd[1635]: time="2026-01-14T13:50:41.584887840Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.4-0\"" Jan 14 13:50:42.768055 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount12337088.mount: Deactivated successfully. Jan 14 13:50:51.110200 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 7. Jan 14 13:50:51.120889 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 13:50:51.601200 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 13:50:51.601000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:50:51.648793 kernel: audit: type=1130 audit(1768398651.601:284): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:50:51.664923 (kubelet)[2411]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 14 13:50:52.172148 kubelet[2411]: E0114 13:50:52.169501 2411 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 14 13:50:52.181234 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 14 13:50:52.183175 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 14 13:50:52.185198 systemd[1]: kubelet.service: Consumed 875ms CPU time, 112.5M memory peak. Jan 14 13:50:52.185000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 13:50:52.241870 kernel: audit: type=1131 audit(1768398652.185:285): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 13:50:56.611951 containerd[1635]: time="2026-01-14T13:50:56.611888119Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.6.4-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:56.615870 containerd[1635]: time="2026-01-14T13:50:56.615840499Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.6.4-0: active requests=0, bytes read=74052490" Jan 14 13:50:56.622961 containerd[1635]: time="2026-01-14T13:50:56.622267882Z" level=info msg="ImageCreate event name:\"sha256:5f1f5298c888daa46c4409ff4cefe5ca9d16e479419f94cdb5f5d5563dac0115\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:56.631952 containerd[1635]: time="2026-01-14T13:50:56.631919320Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:e36c081683425b5b3bc1425bc508b37e7107bb65dfa9367bf5a80125d431fa19\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:50:56.633302 containerd[1635]: time="2026-01-14T13:50:56.633183935Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.6.4-0\" with image id \"sha256:5f1f5298c888daa46c4409ff4cefe5ca9d16e479419f94cdb5f5d5563dac0115\", repo tag \"registry.k8s.io/etcd:3.6.4-0\", repo digest \"registry.k8s.io/etcd@sha256:e36c081683425b5b3bc1425bc508b37e7107bb65dfa9367bf5a80125d431fa19\", size \"74311308\" in 15.04826698s" Jan 14 13:50:56.633302 containerd[1635]: time="2026-01-14T13:50:56.633217288Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.6.4-0\" returns image reference \"sha256:5f1f5298c888daa46c4409ff4cefe5ca9d16e479419f94cdb5f5d5563dac0115\"" Jan 14 13:51:02.356935 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 8. Jan 14 13:51:02.364035 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 13:51:03.008496 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 13:51:03.008000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:51:03.062016 kernel: audit: type=1130 audit(1768398663.008:286): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:51:03.089904 (kubelet)[2454]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 14 13:51:03.639093 kubelet[2454]: E0114 13:51:03.637996 2454 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 14 13:51:03.646118 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 14 13:51:03.647069 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 14 13:51:03.647000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 13:51:03.649336 systemd[1]: kubelet.service: Consumed 1.143s CPU time, 110.5M memory peak. Jan 14 13:51:03.702053 kernel: audit: type=1131 audit(1768398663.647:287): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 13:51:07.207972 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 13:51:07.207000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:51:07.209296 systemd[1]: kubelet.service: Consumed 1.143s CPU time, 110.5M memory peak. Jan 14 13:51:07.218911 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 13:51:07.207000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:51:07.268893 kernel: audit: type=1130 audit(1768398667.207:288): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:51:07.268960 kernel: audit: type=1131 audit(1768398667.207:289): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:51:07.415439 systemd[1]: Reload requested from client PID 2469 ('systemctl') (unit session-8.scope)... Jan 14 13:51:07.416311 systemd[1]: Reloading... Jan 14 13:51:07.714856 zram_generator::config[2511]: No configuration found. Jan 14 13:51:08.680218 systemd[1]: Reloading finished in 1262 ms. Jan 14 13:51:08.805000 audit: BPF prog-id=61 op=LOAD Jan 14 13:51:08.807000 audit: BPF prog-id=57 op=UNLOAD Jan 14 13:51:08.828969 kernel: audit: type=1334 audit(1768398668.805:290): prog-id=61 op=LOAD Jan 14 13:51:08.829031 kernel: audit: type=1334 audit(1768398668.807:291): prog-id=57 op=UNLOAD Jan 14 13:51:08.809000 audit: BPF prog-id=62 op=LOAD Jan 14 13:51:08.809000 audit: BPF prog-id=50 op=UNLOAD Jan 14 13:51:08.813000 audit: BPF prog-id=63 op=LOAD Jan 14 13:51:08.829984 kernel: audit: type=1334 audit(1768398668.809:292): prog-id=62 op=LOAD Jan 14 13:51:08.830021 kernel: audit: type=1334 audit(1768398668.809:293): prog-id=50 op=UNLOAD Jan 14 13:51:08.830251 kernel: audit: type=1334 audit(1768398668.813:294): prog-id=63 op=LOAD Jan 14 13:51:08.813000 audit: BPF prog-id=47 op=UNLOAD Jan 14 13:51:08.830896 kernel: audit: type=1334 audit(1768398668.813:295): prog-id=47 op=UNLOAD Jan 14 13:51:08.830936 kernel: audit: type=1334 audit(1768398668.814:296): prog-id=64 op=LOAD Jan 14 13:51:08.830972 kernel: audit: type=1334 audit(1768398668.814:297): prog-id=65 op=LOAD Jan 14 13:51:08.831000 kernel: audit: type=1334 audit(1768398668.814:298): prog-id=48 op=UNLOAD Jan 14 13:51:08.831026 kernel: audit: type=1334 audit(1768398668.816:299): prog-id=49 op=UNLOAD Jan 14 13:51:08.814000 audit: BPF prog-id=64 op=LOAD Jan 14 13:51:08.814000 audit: BPF prog-id=65 op=LOAD Jan 14 13:51:08.814000 audit: BPF prog-id=48 op=UNLOAD Jan 14 13:51:08.816000 audit: BPF prog-id=49 op=UNLOAD Jan 14 13:51:08.822000 audit: BPF prog-id=66 op=LOAD Jan 14 13:51:08.822000 audit: BPF prog-id=51 op=UNLOAD Jan 14 13:51:08.822000 audit: BPF prog-id=67 op=LOAD Jan 14 13:51:08.822000 audit: BPF prog-id=68 op=LOAD Jan 14 13:51:08.822000 audit: BPF prog-id=52 op=UNLOAD Jan 14 13:51:08.822000 audit: BPF prog-id=53 op=UNLOAD Jan 14 13:51:08.826000 audit: BPF prog-id=69 op=LOAD Jan 14 13:51:08.826000 audit: BPF prog-id=56 op=UNLOAD Jan 14 13:51:08.826000 audit: BPF prog-id=70 op=LOAD Jan 14 13:51:08.826000 audit: BPF prog-id=41 op=UNLOAD Jan 14 13:51:08.826000 audit: BPF prog-id=71 op=LOAD Jan 14 13:51:08.826000 audit: BPF prog-id=72 op=LOAD Jan 14 13:51:08.826000 audit: BPF prog-id=42 op=UNLOAD Jan 14 13:51:08.829000 audit: BPF prog-id=43 op=UNLOAD Jan 14 13:51:08.830000 audit: BPF prog-id=73 op=LOAD Jan 14 13:51:08.830000 audit: BPF prog-id=44 op=UNLOAD Jan 14 13:51:08.830000 audit: BPF prog-id=74 op=LOAD Jan 14 13:51:08.830000 audit: BPF prog-id=75 op=LOAD Jan 14 13:51:08.830000 audit: BPF prog-id=45 op=UNLOAD Jan 14 13:51:08.830000 audit: BPF prog-id=46 op=UNLOAD Jan 14 13:51:08.835000 audit: BPF prog-id=76 op=LOAD Jan 14 13:51:08.835000 audit: BPF prog-id=58 op=UNLOAD Jan 14 13:51:08.835000 audit: BPF prog-id=77 op=LOAD Jan 14 13:51:08.835000 audit: BPF prog-id=78 op=LOAD Jan 14 13:51:08.835000 audit: BPF prog-id=59 op=UNLOAD Jan 14 13:51:08.835000 audit: BPF prog-id=60 op=UNLOAD Jan 14 13:51:08.841000 audit: BPF prog-id=79 op=LOAD Jan 14 13:51:08.841000 audit: BPF prog-id=80 op=LOAD Jan 14 13:51:08.841000 audit: BPF prog-id=54 op=UNLOAD Jan 14 13:51:08.841000 audit: BPF prog-id=55 op=UNLOAD Jan 14 13:51:09.000415 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jan 14 13:51:09.001496 systemd[1]: kubelet.service: Failed with result 'signal'. Jan 14 13:51:09.002000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 13:51:09.003330 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 13:51:09.004371 systemd[1]: kubelet.service: Consumed 591ms CPU time, 98.2M memory peak. Jan 14 13:51:09.017288 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 13:51:09.747261 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 13:51:09.747000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:51:09.822415 (kubelet)[2561]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 14 13:51:10.495257 kubelet[2561]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jan 14 13:51:10.495257 kubelet[2561]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 14 13:51:10.495257 kubelet[2561]: I0114 13:51:10.494503 2561 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 14 13:51:11.096492 kubelet[2561]: I0114 13:51:11.095807 2561 server.go:529] "Kubelet version" kubeletVersion="v1.34.1" Jan 14 13:51:11.096492 kubelet[2561]: I0114 13:51:11.095979 2561 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 14 13:51:11.098444 kubelet[2561]: I0114 13:51:11.097921 2561 watchdog_linux.go:95] "Systemd watchdog is not enabled" Jan 14 13:51:11.098444 kubelet[2561]: I0114 13:51:11.098232 2561 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jan 14 13:51:11.099498 kubelet[2561]: I0114 13:51:11.099292 2561 server.go:956] "Client rotation is on, will bootstrap in background" Jan 14 13:51:11.213400 kubelet[2561]: E0114 13:51:11.212477 2561 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.127:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.127:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jan 14 13:51:11.218198 kubelet[2561]: I0114 13:51:11.216229 2561 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 14 13:51:11.262404 kubelet[2561]: I0114 13:51:11.260382 2561 server.go:1423] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 14 13:51:11.291474 kubelet[2561]: I0114 13:51:11.290925 2561 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Jan 14 13:51:11.300835 kubelet[2561]: I0114 13:51:11.298238 2561 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 14 13:51:11.304227 kubelet[2561]: I0114 13:51:11.300344 2561 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 14 13:51:11.304227 kubelet[2561]: I0114 13:51:11.303507 2561 topology_manager.go:138] "Creating topology manager with none policy" Jan 14 13:51:11.304227 kubelet[2561]: I0114 13:51:11.303850 2561 container_manager_linux.go:306] "Creating device plugin manager" Jan 14 13:51:11.304227 kubelet[2561]: I0114 13:51:11.303957 2561 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Jan 14 13:51:11.331360 kubelet[2561]: I0114 13:51:11.330342 2561 state_mem.go:36] "Initialized new in-memory state store" Jan 14 13:51:11.334689 kubelet[2561]: I0114 13:51:11.334329 2561 kubelet.go:475] "Attempting to sync node with API server" Jan 14 13:51:11.335008 kubelet[2561]: I0114 13:51:11.334849 2561 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 14 13:51:11.335008 kubelet[2561]: I0114 13:51:11.334883 2561 kubelet.go:387] "Adding apiserver pod source" Jan 14 13:51:11.335008 kubelet[2561]: I0114 13:51:11.334920 2561 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 14 13:51:11.337302 kubelet[2561]: E0114 13:51:11.336011 2561 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.127:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.127:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 14 13:51:11.341008 kubelet[2561]: E0114 13:51:11.340254 2561 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.127:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.127:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 14 13:51:11.364402 kubelet[2561]: I0114 13:51:11.364376 2561 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v2.1.5" apiVersion="v1" Jan 14 13:51:11.368419 kubelet[2561]: I0114 13:51:11.365870 2561 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Jan 14 13:51:11.368419 kubelet[2561]: I0114 13:51:11.365906 2561 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Jan 14 13:51:11.368419 kubelet[2561]: W0114 13:51:11.365967 2561 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jan 14 13:51:11.427491 kubelet[2561]: I0114 13:51:11.426399 2561 server.go:1262] "Started kubelet" Jan 14 13:51:11.432970 kubelet[2561]: I0114 13:51:11.429962 2561 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 14 13:51:11.432970 kubelet[2561]: I0114 13:51:11.430240 2561 server_v1.go:49] "podresources" method="list" useActivePods=true Jan 14 13:51:11.432970 kubelet[2561]: I0114 13:51:11.430897 2561 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 14 13:51:11.432970 kubelet[2561]: I0114 13:51:11.432002 2561 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Jan 14 13:51:11.434796 kubelet[2561]: I0114 13:51:11.434382 2561 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 14 13:51:11.444469 kubelet[2561]: I0114 13:51:11.444430 2561 server.go:310] "Adding debug handlers to kubelet server" Jan 14 13:51:11.451836 kubelet[2561]: I0114 13:51:11.450434 2561 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 14 13:51:11.459426 kubelet[2561]: E0114 13:51:11.459004 2561 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 14 13:51:11.462851 kubelet[2561]: E0114 13:51:11.449425 2561 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.127:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.127:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.188a9d38e79450c8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-01-14 13:51:11.425327304 +0000 UTC m=+1.546292266,LastTimestamp:2026-01-14 13:51:11.425327304 +0000 UTC m=+1.546292266,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jan 14 13:51:11.463443 kubelet[2561]: I0114 13:51:11.459923 2561 volume_manager.go:313] "Starting Kubelet Volume Manager" Jan 14 13:51:11.463919 kubelet[2561]: I0114 13:51:11.460245 2561 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 14 13:51:11.464318 kubelet[2561]: E0114 13:51:11.461237 2561 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.127:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.127:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 14 13:51:11.464371 kubelet[2561]: E0114 13:51:11.461919 2561 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.127:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.127:6443: connect: connection refused" interval="200ms" Jan 14 13:51:11.464967 kubelet[2561]: I0114 13:51:11.464951 2561 reconciler.go:29] "Reconciler: start to sync state" Jan 14 13:51:11.467031 kubelet[2561]: I0114 13:51:11.466220 2561 factory.go:223] Registration of the systemd container factory successfully Jan 14 13:51:11.467353 kubelet[2561]: I0114 13:51:11.467333 2561 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 14 13:51:11.472932 kubelet[2561]: E0114 13:51:11.472911 2561 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 14 13:51:11.482335 kubelet[2561]: I0114 13:51:11.481389 2561 factory.go:223] Registration of the containerd container factory successfully Jan 14 13:51:11.536000 audit[2582]: NETFILTER_CFG table=mangle:42 family=10 entries=2 op=nft_register_chain pid=2582 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:11.536000 audit[2582]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffdd8aeb170 a2=0 a3=0 items=0 ppid=2561 pid=2582 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:11.536000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Jan 14 13:51:11.542255 kubelet[2561]: I0114 13:51:11.539921 2561 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Jan 14 13:51:11.558000 audit[2584]: NETFILTER_CFG table=mangle:43 family=2 entries=2 op=nft_register_chain pid=2584 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:11.558000 audit[2584]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7fffd25b40d0 a2=0 a3=0 items=0 ppid=2561 pid=2584 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:11.558000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Jan 14 13:51:11.560000 audit[2585]: NETFILTER_CFG table=mangle:44 family=10 entries=1 op=nft_register_chain pid=2585 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:11.560000 audit[2585]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffce539f2a0 a2=0 a3=0 items=0 ppid=2561 pid=2585 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:11.560000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Jan 14 13:51:11.563989 kubelet[2561]: E0114 13:51:11.563956 2561 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 14 13:51:11.582311 kubelet[2561]: I0114 13:51:11.581271 2561 cpu_manager.go:221] "Starting CPU manager" policy="none" Jan 14 13:51:11.582311 kubelet[2561]: I0114 13:51:11.581451 2561 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jan 14 13:51:11.582311 kubelet[2561]: I0114 13:51:11.581477 2561 state_mem.go:36] "Initialized new in-memory state store" Jan 14 13:51:11.583000 audit[2587]: NETFILTER_CFG table=nat:45 family=10 entries=1 op=nft_register_chain pid=2587 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:11.583000 audit[2587]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffdda6889f0 a2=0 a3=0 items=0 ppid=2561 pid=2587 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:11.583000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Jan 14 13:51:11.592000 audit[2588]: NETFILTER_CFG table=filter:46 family=2 entries=1 op=nft_register_chain pid=2588 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:11.592000 audit[2588]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fffe67c4cd0 a2=0 a3=0 items=0 ppid=2561 pid=2588 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:11.592000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D4649524557414C4C002D740066696C746572 Jan 14 13:51:11.598888 kubelet[2561]: I0114 13:51:11.598868 2561 policy_none.go:49] "None policy: Start" Jan 14 13:51:11.599021 kubelet[2561]: I0114 13:51:11.599006 2561 memory_manager.go:187] "Starting memorymanager" policy="None" Jan 14 13:51:11.599270 kubelet[2561]: I0114 13:51:11.599258 2561 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Jan 14 13:51:11.605000 audit[2589]: NETFILTER_CFG table=filter:47 family=10 entries=1 op=nft_register_chain pid=2589 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:11.605000 audit[2589]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe28978d20 a2=0 a3=0 items=0 ppid=2561 pid=2589 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:11.605000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Jan 14 13:51:11.613318 kubelet[2561]: I0114 13:51:11.611372 2561 policy_none.go:47] "Start" Jan 14 13:51:11.626000 audit[2591]: NETFILTER_CFG table=filter:48 family=2 entries=2 op=nft_register_chain pid=2591 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:11.626000 audit[2591]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffdd32874e0 a2=0 a3=0 items=0 ppid=2561 pid=2591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:11.626000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 14 13:51:11.666268 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jan 14 13:51:11.667432 kubelet[2561]: E0114 13:51:11.665502 2561 kubelet_node_status.go:404] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 14 13:51:11.670965 kubelet[2561]: E0114 13:51:11.669242 2561 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.127:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.127:6443: connect: connection refused" interval="400ms" Jan 14 13:51:11.678000 audit[2593]: NETFILTER_CFG table=filter:49 family=2 entries=2 op=nft_register_chain pid=2593 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:11.678000 audit[2593]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffc82f0cc50 a2=0 a3=0 items=0 ppid=2561 pid=2593 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:11.678000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 14 13:51:11.715895 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jan 14 13:51:11.737259 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jan 14 13:51:11.760405 kubelet[2561]: E0114 13:51:11.760010 2561 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Jan 14 13:51:11.760493 kubelet[2561]: I0114 13:51:11.760475 2561 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 14 13:51:11.760893 kubelet[2561]: I0114 13:51:11.760490 2561 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 14 13:51:11.763343 kubelet[2561]: I0114 13:51:11.761319 2561 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 14 13:51:11.766000 audit[2597]: NETFILTER_CFG table=filter:50 family=2 entries=1 op=nft_register_rule pid=2597 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:11.766000 audit[2597]: SYSCALL arch=c000003e syscall=46 success=yes exit=924 a0=3 a1=7ffc844322e0 a2=0 a3=0 items=0 ppid=2561 pid=2597 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:11.766000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D41004B5542452D4649524557414C4C002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E7400626C6F636B20696E636F6D696E67206C6F63616C6E657420636F6E6E656374696F6E73002D2D647374003132372E302E302E302F380000002D2D737263003132372E Jan 14 13:51:11.770773 kubelet[2561]: I0114 13:51:11.769990 2561 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Jan 14 13:51:11.770773 kubelet[2561]: I0114 13:51:11.770007 2561 status_manager.go:244] "Starting to sync pod status with apiserver" Jan 14 13:51:11.770773 kubelet[2561]: I0114 13:51:11.770200 2561 kubelet.go:2427] "Starting kubelet main sync loop" Jan 14 13:51:11.770773 kubelet[2561]: E0114 13:51:11.770258 2561 kubelet.go:2451] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Jan 14 13:51:11.778364 kubelet[2561]: E0114 13:51:11.776416 2561 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.127:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.127:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 14 13:51:11.778364 kubelet[2561]: E0114 13:51:11.777336 2561 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jan 14 13:51:11.778364 kubelet[2561]: E0114 13:51:11.777373 2561 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jan 14 13:51:11.799000 audit[2598]: NETFILTER_CFG table=mangle:51 family=2 entries=1 op=nft_register_chain pid=2598 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:11.799000 audit[2598]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffcfb7d3db0 a2=0 a3=0 items=0 ppid=2561 pid=2598 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:11.799000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Jan 14 13:51:11.817000 audit[2599]: NETFILTER_CFG table=nat:52 family=2 entries=1 op=nft_register_chain pid=2599 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:11.817000 audit[2599]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffde2b4c290 a2=0 a3=0 items=0 ppid=2561 pid=2599 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:11.817000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Jan 14 13:51:11.856000 audit[2600]: NETFILTER_CFG table=filter:53 family=2 entries=1 op=nft_register_chain pid=2600 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:11.856000 audit[2600]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fffa251e840 a2=0 a3=0 items=0 ppid=2561 pid=2600 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:11.856000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Jan 14 13:51:11.872476 kubelet[2561]: I0114 13:51:11.872451 2561 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 14 13:51:11.877261 kubelet[2561]: E0114 13:51:11.875976 2561 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.127:6443/api/v1/nodes\": dial tcp 10.0.0.127:6443: connect: connection refused" node="localhost" Jan 14 13:51:11.936325 systemd[1]: Created slice kubepods-burstable-pod0ae88e709ded7aa62d05a5e6bfe67ee8.slice - libcontainer container kubepods-burstable-pod0ae88e709ded7aa62d05a5e6bfe67ee8.slice. Jan 14 13:51:11.974478 kubelet[2561]: I0114 13:51:11.973976 2561 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0ae88e709ded7aa62d05a5e6bfe67ee8-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"0ae88e709ded7aa62d05a5e6bfe67ee8\") " pod="kube-system/kube-apiserver-localhost" Jan 14 13:51:11.974478 kubelet[2561]: I0114 13:51:11.974026 2561 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 14 13:51:11.974478 kubelet[2561]: I0114 13:51:11.974249 2561 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/07ca0cbf79ad6ba9473d8e9f7715e571-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"07ca0cbf79ad6ba9473d8e9f7715e571\") " pod="kube-system/kube-scheduler-localhost" Jan 14 13:51:11.974478 kubelet[2561]: I0114 13:51:11.974273 2561 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 14 13:51:11.974478 kubelet[2561]: I0114 13:51:11.974294 2561 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 14 13:51:11.975282 kubelet[2561]: I0114 13:51:11.974312 2561 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 14 13:51:11.975282 kubelet[2561]: I0114 13:51:11.974331 2561 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 14 13:51:11.975282 kubelet[2561]: I0114 13:51:11.974350 2561 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0ae88e709ded7aa62d05a5e6bfe67ee8-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"0ae88e709ded7aa62d05a5e6bfe67ee8\") " pod="kube-system/kube-apiserver-localhost" Jan 14 13:51:11.975282 kubelet[2561]: I0114 13:51:11.974370 2561 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0ae88e709ded7aa62d05a5e6bfe67ee8-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"0ae88e709ded7aa62d05a5e6bfe67ee8\") " pod="kube-system/kube-apiserver-localhost" Jan 14 13:51:12.011987 kubelet[2561]: E0114 13:51:12.010451 2561 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 14 13:51:12.042899 systemd[1]: Created slice kubepods-burstable-pod5bbfee13ce9e07281eca876a0b8067f2.slice - libcontainer container kubepods-burstable-pod5bbfee13ce9e07281eca876a0b8067f2.slice. Jan 14 13:51:12.056387 kubelet[2561]: E0114 13:51:12.055465 2561 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 14 13:51:12.073473 systemd[1]: Created slice kubepods-burstable-pod07ca0cbf79ad6ba9473d8e9f7715e571.slice - libcontainer container kubepods-burstable-pod07ca0cbf79ad6ba9473d8e9f7715e571.slice. Jan 14 13:51:12.076464 kubelet[2561]: E0114 13:51:12.074279 2561 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.127:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.127:6443: connect: connection refused" interval="800ms" Jan 14 13:51:12.083950 kubelet[2561]: E0114 13:51:12.083006 2561 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 14 13:51:12.091422 kubelet[2561]: I0114 13:51:12.089932 2561 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 14 13:51:12.091422 kubelet[2561]: E0114 13:51:12.091241 2561 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.127:6443/api/v1/nodes\": dial tcp 10.0.0.127:6443: connect: connection refused" node="localhost" Jan 14 13:51:12.101428 kubelet[2561]: E0114 13:51:12.101200 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:12.106346 containerd[1635]: time="2026-01-14T13:51:12.105025650Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:07ca0cbf79ad6ba9473d8e9f7715e571,Namespace:kube-system,Attempt:0,}" Jan 14 13:51:12.333396 kubelet[2561]: E0114 13:51:12.331978 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:12.338269 containerd[1635]: time="2026-01-14T13:51:12.337361091Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:0ae88e709ded7aa62d05a5e6bfe67ee8,Namespace:kube-system,Attempt:0,}" Jan 14 13:51:12.373971 kubelet[2561]: E0114 13:51:12.373301 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:12.375990 containerd[1635]: time="2026-01-14T13:51:12.375243112Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:5bbfee13ce9e07281eca876a0b8067f2,Namespace:kube-system,Attempt:0,}" Jan 14 13:51:12.439394 kubelet[2561]: E0114 13:51:12.435501 2561 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.127:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.127:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 14 13:51:12.491351 kubelet[2561]: E0114 13:51:12.489299 2561 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.127:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.127:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 14 13:51:12.498960 kubelet[2561]: I0114 13:51:12.498929 2561 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 14 13:51:12.504325 kubelet[2561]: E0114 13:51:12.502398 2561 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.127:6443/api/v1/nodes\": dial tcp 10.0.0.127:6443: connect: connection refused" node="localhost" Jan 14 13:51:12.546271 kubelet[2561]: E0114 13:51:12.545303 2561 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.127:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.127:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 14 13:51:12.862201 kubelet[2561]: E0114 13:51:12.861376 2561 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.127:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.127:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 14 13:51:12.877249 kubelet[2561]: E0114 13:51:12.876973 2561 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.127:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.127:6443: connect: connection refused" interval="1.6s" Jan 14 13:51:13.032494 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1448269198.mount: Deactivated successfully. Jan 14 13:51:13.068991 containerd[1635]: time="2026-01-14T13:51:13.068948824Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 14 13:51:13.101844 containerd[1635]: time="2026-01-14T13:51:13.100932469Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Jan 14 13:51:13.115415 containerd[1635]: time="2026-01-14T13:51:13.115319259Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 14 13:51:13.134015 containerd[1635]: time="2026-01-14T13:51:13.133980500Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 14 13:51:13.145889 containerd[1635]: time="2026-01-14T13:51:13.145862268Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 14 13:51:13.152858 containerd[1635]: time="2026-01-14T13:51:13.150397496Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Jan 14 13:51:13.158453 containerd[1635]: time="2026-01-14T13:51:13.158253107Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 14 13:51:13.165451 containerd[1635]: time="2026-01-14T13:51:13.164928492Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Jan 14 13:51:13.169309 containerd[1635]: time="2026-01-14T13:51:13.168328402Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 1.032429769s" Jan 14 13:51:13.175314 containerd[1635]: time="2026-01-14T13:51:13.174429242Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 784.013668ms" Jan 14 13:51:13.179328 containerd[1635]: time="2026-01-14T13:51:13.178224007Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 821.049369ms" Jan 14 13:51:13.325503 kubelet[2561]: I0114 13:51:13.323356 2561 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 14 13:51:13.332230 kubelet[2561]: E0114 13:51:13.331317 2561 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.127:6443/api/v1/nodes\": dial tcp 10.0.0.127:6443: connect: connection refused" node="localhost" Jan 14 13:51:13.332230 kubelet[2561]: E0114 13:51:13.327352 2561 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.127:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.127:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jan 14 13:51:13.408027 containerd[1635]: time="2026-01-14T13:51:13.407280935Z" level=info msg="connecting to shim 5bd93046c39914e08bb5000cc0feb5d5ccf8119de267e33b33a927b606f0e5d7" address="unix:///run/containerd/s/7dfc3cad69f86f6bf167f54bea4381a570774b6f0ec575734e9cfe33dc73fca5" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:51:13.500359 containerd[1635]: time="2026-01-14T13:51:13.499419573Z" level=info msg="connecting to shim c59a3421ff93fce8b9907588684abd2d9bf9a807ca00f20328bb0f8787fa6bd8" address="unix:///run/containerd/s/f91bef982f6bee161afc95d69feb5d346a07d69fd623dc8913f86e1f20392587" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:51:13.506385 containerd[1635]: time="2026-01-14T13:51:13.505376392Z" level=info msg="connecting to shim 3d83a2899dc76060d3f888287ea06aaed91617051fcaee87bb8e0a6f1c2596b7" address="unix:///run/containerd/s/9fa59e9daa0e02101064411e3bcabf6efbadb8f3004e28a801246bb95b64cd7f" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:51:13.635194 systemd[1]: Started cri-containerd-5bd93046c39914e08bb5000cc0feb5d5ccf8119de267e33b33a927b606f0e5d7.scope - libcontainer container 5bd93046c39914e08bb5000cc0feb5d5ccf8119de267e33b33a927b606f0e5d7. Jan 14 13:51:13.729880 systemd[1]: Started cri-containerd-3d83a2899dc76060d3f888287ea06aaed91617051fcaee87bb8e0a6f1c2596b7.scope - libcontainer container 3d83a2899dc76060d3f888287ea06aaed91617051fcaee87bb8e0a6f1c2596b7. Jan 14 13:51:13.748000 audit: BPF prog-id=81 op=LOAD Jan 14 13:51:13.751000 audit: BPF prog-id=82 op=LOAD Jan 14 13:51:13.751000 audit[2639]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001b0238 a2=98 a3=0 items=0 ppid=2614 pid=2639 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:13.751000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562643933303436633339393134653038626235303030636330666562 Jan 14 13:51:13.752000 audit: BPF prog-id=82 op=UNLOAD Jan 14 13:51:13.752000 audit[2639]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2614 pid=2639 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:13.752000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562643933303436633339393134653038626235303030636330666562 Jan 14 13:51:13.753000 audit: BPF prog-id=83 op=LOAD Jan 14 13:51:13.753000 audit[2639]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001b0488 a2=98 a3=0 items=0 ppid=2614 pid=2639 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:13.753000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562643933303436633339393134653038626235303030636330666562 Jan 14 13:51:13.753000 audit: BPF prog-id=84 op=LOAD Jan 14 13:51:13.753000 audit[2639]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001b0218 a2=98 a3=0 items=0 ppid=2614 pid=2639 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:13.753000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562643933303436633339393134653038626235303030636330666562 Jan 14 13:51:13.753000 audit: BPF prog-id=84 op=UNLOAD Jan 14 13:51:13.753000 audit[2639]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2614 pid=2639 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:13.753000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562643933303436633339393134653038626235303030636330666562 Jan 14 13:51:13.754000 audit: BPF prog-id=83 op=UNLOAD Jan 14 13:51:13.754000 audit[2639]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2614 pid=2639 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:13.754000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562643933303436633339393134653038626235303030636330666562 Jan 14 13:51:13.754000 audit: BPF prog-id=85 op=LOAD Jan 14 13:51:13.754000 audit[2639]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001b06e8 a2=98 a3=0 items=0 ppid=2614 pid=2639 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:13.754000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562643933303436633339393134653038626235303030636330666562 Jan 14 13:51:13.816211 systemd[1]: Started cri-containerd-c59a3421ff93fce8b9907588684abd2d9bf9a807ca00f20328bb0f8787fa6bd8.scope - libcontainer container c59a3421ff93fce8b9907588684abd2d9bf9a807ca00f20328bb0f8787fa6bd8. Jan 14 13:51:13.851900 kernel: kauditd_printk_skb: 90 callbacks suppressed Jan 14 13:51:13.852232 kernel: audit: type=1334 audit(1768398673.830:352): prog-id=86 op=LOAD Jan 14 13:51:13.830000 audit: BPF prog-id=86 op=LOAD Jan 14 13:51:13.877000 audit: BPF prog-id=87 op=LOAD Jan 14 13:51:13.903456 kernel: audit: type=1334 audit(1768398673.877:353): prog-id=87 op=LOAD Jan 14 13:51:13.877000 audit[2669]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8238 a2=98 a3=0 items=0 ppid=2644 pid=2669 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:13.994390 kernel: audit: type=1300 audit(1768398673.877:353): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8238 a2=98 a3=0 items=0 ppid=2644 pid=2669 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:13.877000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364383361323839396463373630363064336638383832383765613036 Jan 14 13:51:14.053007 containerd[1635]: time="2026-01-14T13:51:14.052484715Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:5bbfee13ce9e07281eca876a0b8067f2,Namespace:kube-system,Attempt:0,} returns sandbox id \"5bd93046c39914e08bb5000cc0feb5d5ccf8119de267e33b33a927b606f0e5d7\"" Jan 14 13:51:14.059271 kubelet[2561]: E0114 13:51:14.057971 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:14.080949 kernel: audit: type=1327 audit(1768398673.877:353): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364383361323839396463373630363064336638383832383765613036 Jan 14 13:51:14.081240 kubelet[2561]: E0114 13:51:14.080002 2561 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.127:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.127:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.188a9d38e79450c8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-01-14 13:51:11.425327304 +0000 UTC m=+1.546292266,LastTimestamp:2026-01-14 13:51:11.425327304 +0000 UTC m=+1.546292266,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jan 14 13:51:13.877000 audit: BPF prog-id=87 op=UNLOAD Jan 14 13:51:14.088989 containerd[1635]: time="2026-01-14T13:51:14.086003139Z" level=info msg="CreateContainer within sandbox \"5bd93046c39914e08bb5000cc0feb5d5ccf8119de267e33b33a927b606f0e5d7\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jan 14 13:51:14.105911 kernel: audit: type=1334 audit(1768398673.877:354): prog-id=87 op=UNLOAD Jan 14 13:51:14.106004 kernel: audit: type=1300 audit(1768398673.877:354): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2644 pid=2669 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:13.877000 audit[2669]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2644 pid=2669 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.144851 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2515785486.mount: Deactivated successfully. Jan 14 13:51:13.877000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364383361323839396463373630363064336638383832383765613036 Jan 14 13:51:14.238951 kernel: audit: type=1327 audit(1768398673.877:354): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364383361323839396463373630363064336638383832383765613036 Jan 14 13:51:14.255858 kernel: audit: type=1334 audit(1768398673.877:355): prog-id=88 op=LOAD Jan 14 13:51:14.255975 kernel: audit: type=1300 audit(1768398673.877:355): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8488 a2=98 a3=0 items=0 ppid=2644 pid=2669 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:13.877000 audit: BPF prog-id=88 op=LOAD Jan 14 13:51:13.877000 audit[2669]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8488 a2=98 a3=0 items=0 ppid=2644 pid=2669 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.255417 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1898114917.mount: Deactivated successfully. Jan 14 13:51:14.301912 containerd[1635]: time="2026-01-14T13:51:14.300239974Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:0ae88e709ded7aa62d05a5e6bfe67ee8,Namespace:kube-system,Attempt:0,} returns sandbox id \"3d83a2899dc76060d3f888287ea06aaed91617051fcaee87bb8e0a6f1c2596b7\"" Jan 14 13:51:14.306408 kubelet[2561]: E0114 13:51:14.304341 2561 reflector.go:205] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.127:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.127:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 14 13:51:14.309721 kubelet[2561]: E0114 13:51:14.308871 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:14.320781 kernel: audit: type=1327 audit(1768398673.877:355): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364383361323839396463373630363064336638383832383765613036 Jan 14 13:51:13.877000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364383361323839396463373630363064336638383832383765613036 Jan 14 13:51:14.365844 containerd[1635]: time="2026-01-14T13:51:14.325484742Z" level=info msg="CreateContainer within sandbox \"3d83a2899dc76060d3f888287ea06aaed91617051fcaee87bb8e0a6f1c2596b7\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jan 14 13:51:13.878000 audit: BPF prog-id=89 op=LOAD Jan 14 13:51:13.878000 audit[2669]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a8218 a2=98 a3=0 items=0 ppid=2644 pid=2669 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:13.878000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364383361323839396463373630363064336638383832383765613036 Jan 14 13:51:13.878000 audit: BPF prog-id=89 op=UNLOAD Jan 14 13:51:13.878000 audit[2669]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2644 pid=2669 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:13.878000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364383361323839396463373630363064336638383832383765613036 Jan 14 13:51:13.878000 audit: BPF prog-id=88 op=UNLOAD Jan 14 13:51:13.878000 audit[2669]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2644 pid=2669 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.377448 containerd[1635]: time="2026-01-14T13:51:14.377414373Z" level=info msg="Container 879199bd2c873bd40e2ea98edd46d6ad2ba88fddec93daee501324de418d7271: CDI devices from CRI Config.CDIDevices: []" Jan 14 13:51:13.878000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364383361323839396463373630363064336638383832383765613036 Jan 14 13:51:13.878000 audit: BPF prog-id=90 op=LOAD Jan 14 13:51:13.878000 audit[2669]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a86e8 a2=98 a3=0 items=0 ppid=2644 pid=2669 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:13.878000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364383361323839396463373630363064336638383832383765613036 Jan 14 13:51:14.032000 audit: BPF prog-id=91 op=LOAD Jan 14 13:51:14.036000 audit: BPF prog-id=92 op=LOAD Jan 14 13:51:14.036000 audit[2683]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=2643 pid=2683 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.036000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6335396133343231666639336663653862393930373538383638346162 Jan 14 13:51:14.036000 audit: BPF prog-id=92 op=UNLOAD Jan 14 13:51:14.036000 audit[2683]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2643 pid=2683 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.036000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6335396133343231666639336663653862393930373538383638346162 Jan 14 13:51:14.037000 audit: BPF prog-id=93 op=LOAD Jan 14 13:51:14.037000 audit[2683]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=2643 pid=2683 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.037000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6335396133343231666639336663653862393930373538383638346162 Jan 14 13:51:14.037000 audit: BPF prog-id=94 op=LOAD Jan 14 13:51:14.037000 audit[2683]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=2643 pid=2683 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.037000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6335396133343231666639336663653862393930373538383638346162 Jan 14 13:51:14.037000 audit: BPF prog-id=94 op=UNLOAD Jan 14 13:51:14.037000 audit[2683]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2643 pid=2683 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.037000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6335396133343231666639336663653862393930373538383638346162 Jan 14 13:51:14.038000 audit: BPF prog-id=93 op=UNLOAD Jan 14 13:51:14.038000 audit[2683]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2643 pid=2683 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.038000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6335396133343231666639336663653862393930373538383638346162 Jan 14 13:51:14.038000 audit: BPF prog-id=95 op=LOAD Jan 14 13:51:14.038000 audit[2683]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=2643 pid=2683 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.038000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6335396133343231666639336663653862393930373538383638346162 Jan 14 13:51:14.407163 containerd[1635]: time="2026-01-14T13:51:14.406986149Z" level=info msg="Container a2a31fefccc68746a0a498c985c68a5bb14d1c954628de619071990dfa8d6e77: CDI devices from CRI Config.CDIDevices: []" Jan 14 13:51:14.409505 containerd[1635]: time="2026-01-14T13:51:14.406994750Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:07ca0cbf79ad6ba9473d8e9f7715e571,Namespace:kube-system,Attempt:0,} returns sandbox id \"c59a3421ff93fce8b9907588684abd2d9bf9a807ca00f20328bb0f8787fa6bd8\"" Jan 14 13:51:14.411319 kubelet[2561]: E0114 13:51:14.411293 2561 reflector.go:205] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.127:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.127:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 14 13:51:14.412015 kubelet[2561]: E0114 13:51:14.411988 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:14.417484 containerd[1635]: time="2026-01-14T13:51:14.417227375Z" level=info msg="CreateContainer within sandbox \"5bd93046c39914e08bb5000cc0feb5d5ccf8119de267e33b33a927b606f0e5d7\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"879199bd2c873bd40e2ea98edd46d6ad2ba88fddec93daee501324de418d7271\"" Jan 14 13:51:14.420374 containerd[1635]: time="2026-01-14T13:51:14.419962646Z" level=info msg="StartContainer for \"879199bd2c873bd40e2ea98edd46d6ad2ba88fddec93daee501324de418d7271\"" Jan 14 13:51:14.424144 containerd[1635]: time="2026-01-14T13:51:14.423986043Z" level=info msg="CreateContainer within sandbox \"c59a3421ff93fce8b9907588684abd2d9bf9a807ca00f20328bb0f8787fa6bd8\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jan 14 13:51:14.425280 containerd[1635]: time="2026-01-14T13:51:14.425251231Z" level=info msg="connecting to shim 879199bd2c873bd40e2ea98edd46d6ad2ba88fddec93daee501324de418d7271" address="unix:///run/containerd/s/7dfc3cad69f86f6bf167f54bea4381a570774b6f0ec575734e9cfe33dc73fca5" protocol=ttrpc version=3 Jan 14 13:51:14.437422 containerd[1635]: time="2026-01-14T13:51:14.437242900Z" level=info msg="CreateContainer within sandbox \"3d83a2899dc76060d3f888287ea06aaed91617051fcaee87bb8e0a6f1c2596b7\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"a2a31fefccc68746a0a498c985c68a5bb14d1c954628de619071990dfa8d6e77\"" Jan 14 13:51:14.440476 containerd[1635]: time="2026-01-14T13:51:14.439014313Z" level=info msg="StartContainer for \"a2a31fefccc68746a0a498c985c68a5bb14d1c954628de619071990dfa8d6e77\"" Jan 14 13:51:14.443798 containerd[1635]: time="2026-01-14T13:51:14.443212520Z" level=info msg="connecting to shim a2a31fefccc68746a0a498c985c68a5bb14d1c954628de619071990dfa8d6e77" address="unix:///run/containerd/s/9fa59e9daa0e02101064411e3bcabf6efbadb8f3004e28a801246bb95b64cd7f" protocol=ttrpc version=3 Jan 14 13:51:14.477464 containerd[1635]: time="2026-01-14T13:51:14.477334121Z" level=info msg="Container 21849cc5c8f0225e0d33197a473a59ee58d338e97f21578910457a68414390ef: CDI devices from CRI Config.CDIDevices: []" Jan 14 13:51:14.481328 kubelet[2561]: E0114 13:51:14.481289 2561 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.127:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.127:6443: connect: connection refused" interval="3.2s" Jan 14 13:51:14.510273 containerd[1635]: time="2026-01-14T13:51:14.509920230Z" level=info msg="CreateContainer within sandbox \"c59a3421ff93fce8b9907588684abd2d9bf9a807ca00f20328bb0f8787fa6bd8\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"21849cc5c8f0225e0d33197a473a59ee58d338e97f21578910457a68414390ef\"" Jan 14 13:51:14.511416 containerd[1635]: time="2026-01-14T13:51:14.511238505Z" level=info msg="StartContainer for \"21849cc5c8f0225e0d33197a473a59ee58d338e97f21578910457a68414390ef\"" Jan 14 13:51:14.513774 containerd[1635]: time="2026-01-14T13:51:14.513280078Z" level=info msg="connecting to shim 21849cc5c8f0225e0d33197a473a59ee58d338e97f21578910457a68414390ef" address="unix:///run/containerd/s/f91bef982f6bee161afc95d69feb5d346a07d69fd623dc8913f86e1f20392587" protocol=ttrpc version=3 Jan 14 13:51:14.553920 systemd[1]: Started cri-containerd-a2a31fefccc68746a0a498c985c68a5bb14d1c954628de619071990dfa8d6e77.scope - libcontainer container a2a31fefccc68746a0a498c985c68a5bb14d1c954628de619071990dfa8d6e77. Jan 14 13:51:14.584757 systemd[1]: Started cri-containerd-879199bd2c873bd40e2ea98edd46d6ad2ba88fddec93daee501324de418d7271.scope - libcontainer container 879199bd2c873bd40e2ea98edd46d6ad2ba88fddec93daee501324de418d7271. Jan 14 13:51:14.611200 kubelet[2561]: E0114 13:51:14.611152 2561 reflector.go:205] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.127:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.127:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 14 13:51:14.647477 systemd[1]: Started cri-containerd-21849cc5c8f0225e0d33197a473a59ee58d338e97f21578910457a68414390ef.scope - libcontainer container 21849cc5c8f0225e0d33197a473a59ee58d338e97f21578910457a68414390ef. Jan 14 13:51:14.660000 audit: BPF prog-id=96 op=LOAD Jan 14 13:51:14.665000 audit: BPF prog-id=97 op=LOAD Jan 14 13:51:14.666000 audit: BPF prog-id=98 op=LOAD Jan 14 13:51:14.665000 audit[2745]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=2644 pid=2745 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.665000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6132613331666566636363363837343661306134393863393835633638 Jan 14 13:51:14.667000 audit: BPF prog-id=97 op=UNLOAD Jan 14 13:51:14.667000 audit[2745]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2644 pid=2745 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.667000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6132613331666566636363363837343661306134393863393835633638 Jan 14 13:51:14.668000 audit: BPF prog-id=99 op=LOAD Jan 14 13:51:14.668000 audit[2750]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=2614 pid=2750 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.668000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837393139396264326338373362643430653265613938656464343664 Jan 14 13:51:14.668000 audit: BPF prog-id=99 op=UNLOAD Jan 14 13:51:14.668000 audit[2750]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2614 pid=2750 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.668000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837393139396264326338373362643430653265613938656464343664 Jan 14 13:51:14.671000 audit: BPF prog-id=100 op=LOAD Jan 14 13:51:14.671000 audit[2750]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=2614 pid=2750 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.671000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837393139396264326338373362643430653265613938656464343664 Jan 14 13:51:14.671000 audit: BPF prog-id=101 op=LOAD Jan 14 13:51:14.671000 audit[2750]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=2614 pid=2750 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.671000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837393139396264326338373362643430653265613938656464343664 Jan 14 13:51:14.671000 audit: BPF prog-id=101 op=UNLOAD Jan 14 13:51:14.671000 audit[2750]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2614 pid=2750 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.671000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837393139396264326338373362643430653265613938656464343664 Jan 14 13:51:14.671000 audit: BPF prog-id=100 op=UNLOAD Jan 14 13:51:14.671000 audit[2750]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2614 pid=2750 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.671000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837393139396264326338373362643430653265613938656464343664 Jan 14 13:51:14.671000 audit: BPF prog-id=102 op=LOAD Jan 14 13:51:14.671000 audit: BPF prog-id=103 op=LOAD Jan 14 13:51:14.671000 audit[2750]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=2614 pid=2750 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.671000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837393139396264326338373362643430653265613938656464343664 Jan 14 13:51:14.671000 audit[2745]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=2644 pid=2745 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.671000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6132613331666566636363363837343661306134393863393835633638 Jan 14 13:51:14.672000 audit: BPF prog-id=104 op=LOAD Jan 14 13:51:14.672000 audit[2745]: SYSCALL arch=c000003e syscall=321 success=yes exit=24 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=2644 pid=2745 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.672000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6132613331666566636363363837343661306134393863393835633638 Jan 14 13:51:14.674000 audit: BPF prog-id=104 op=UNLOAD Jan 14 13:51:14.674000 audit[2745]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=18 a1=0 a2=0 a3=0 items=0 ppid=2644 pid=2745 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.674000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6132613331666566636363363837343661306134393863393835633638 Jan 14 13:51:14.674000 audit: BPF prog-id=102 op=UNLOAD Jan 14 13:51:14.674000 audit[2745]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2644 pid=2745 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.674000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6132613331666566636363363837343661306134393863393835633638 Jan 14 13:51:14.679000 audit: BPF prog-id=105 op=LOAD Jan 14 13:51:14.679000 audit[2745]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=2644 pid=2745 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.679000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6132613331666566636363363837343661306134393863393835633638 Jan 14 13:51:14.718000 audit: BPF prog-id=106 op=LOAD Jan 14 13:51:14.720000 audit: BPF prog-id=107 op=LOAD Jan 14 13:51:14.720000 audit[2762]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=2643 pid=2762 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.720000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231383439636335633866303232356530643333313937613437336135 Jan 14 13:51:14.720000 audit: BPF prog-id=107 op=UNLOAD Jan 14 13:51:14.720000 audit[2762]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2643 pid=2762 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.720000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231383439636335633866303232356530643333313937613437336135 Jan 14 13:51:14.722000 audit: BPF prog-id=108 op=LOAD Jan 14 13:51:14.722000 audit[2762]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=2643 pid=2762 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.722000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231383439636335633866303232356530643333313937613437336135 Jan 14 13:51:14.722000 audit: BPF prog-id=109 op=LOAD Jan 14 13:51:14.722000 audit[2762]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=2643 pid=2762 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.722000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231383439636335633866303232356530643333313937613437336135 Jan 14 13:51:14.722000 audit: BPF prog-id=109 op=UNLOAD Jan 14 13:51:14.722000 audit[2762]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2643 pid=2762 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.722000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231383439636335633866303232356530643333313937613437336135 Jan 14 13:51:14.722000 audit: BPF prog-id=108 op=UNLOAD Jan 14 13:51:14.722000 audit[2762]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2643 pid=2762 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.722000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231383439636335633866303232356530643333313937613437336135 Jan 14 13:51:14.722000 audit: BPF prog-id=110 op=LOAD Jan 14 13:51:14.722000 audit[2762]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=2643 pid=2762 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:14.722000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231383439636335633866303232356530643333313937613437336135 Jan 14 13:51:14.890304 containerd[1635]: time="2026-01-14T13:51:14.890259293Z" level=info msg="StartContainer for \"879199bd2c873bd40e2ea98edd46d6ad2ba88fddec93daee501324de418d7271\" returns successfully" Jan 14 13:51:14.944199 containerd[1635]: time="2026-01-14T13:51:14.943943624Z" level=info msg="StartContainer for \"a2a31fefccc68746a0a498c985c68a5bb14d1c954628de619071990dfa8d6e77\" returns successfully" Jan 14 13:51:14.946186 kubelet[2561]: I0114 13:51:14.946001 2561 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 14 13:51:14.951186 kubelet[2561]: E0114 13:51:14.950990 2561 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.127:6443/api/v1/nodes\": dial tcp 10.0.0.127:6443: connect: connection refused" node="localhost" Jan 14 13:51:15.010470 containerd[1635]: time="2026-01-14T13:51:15.010265582Z" level=info msg="StartContainer for \"21849cc5c8f0225e0d33197a473a59ee58d338e97f21578910457a68414390ef\" returns successfully" Jan 14 13:51:15.365926 kubelet[2561]: E0114 13:51:15.365343 2561 reflector.go:205] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.127:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.127:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 14 13:51:15.921766 kubelet[2561]: E0114 13:51:15.920474 2561 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 14 13:51:15.921766 kubelet[2561]: E0114 13:51:15.921215 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:15.966968 kubelet[2561]: E0114 13:51:15.966480 2561 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 14 13:51:15.967424 kubelet[2561]: E0114 13:51:15.967406 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:15.990295 kubelet[2561]: E0114 13:51:15.990264 2561 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 14 13:51:15.990824 kubelet[2561]: E0114 13:51:15.990804 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:16.998930 kubelet[2561]: E0114 13:51:16.998487 2561 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 14 13:51:17.001948 kubelet[2561]: E0114 13:51:17.001358 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:17.001948 kubelet[2561]: E0114 13:51:17.000391 2561 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 14 13:51:17.001948 kubelet[2561]: E0114 13:51:17.001755 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:17.001948 kubelet[2561]: E0114 13:51:17.000156 2561 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 14 13:51:17.001948 kubelet[2561]: E0114 13:51:17.001886 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:18.008453 kubelet[2561]: E0114 13:51:18.007399 2561 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 14 13:51:18.011452 kubelet[2561]: E0114 13:51:18.010340 2561 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 14 13:51:18.012890 kubelet[2561]: E0114 13:51:18.012770 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:18.012890 kubelet[2561]: E0114 13:51:18.011956 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:18.158431 kubelet[2561]: I0114 13:51:18.158355 2561 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 14 13:51:19.015110 kubelet[2561]: E0114 13:51:19.014938 2561 kubelet.go:3215] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 14 13:51:19.015475 kubelet[2561]: E0114 13:51:19.015205 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:19.851164 kubelet[2561]: E0114 13:51:19.851129 2561 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Jan 14 13:51:20.050879 kubelet[2561]: I0114 13:51:20.050504 2561 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Jan 14 13:51:20.063210 kubelet[2561]: I0114 13:51:20.061824 2561 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Jan 14 13:51:20.089122 kubelet[2561]: E0114 13:51:20.088921 2561 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Jan 14 13:51:20.089846 kubelet[2561]: I0114 13:51:20.089827 2561 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Jan 14 13:51:20.095319 kubelet[2561]: E0114 13:51:20.094938 2561 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Jan 14 13:51:20.095319 kubelet[2561]: I0114 13:51:20.094959 2561 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Jan 14 13:51:20.103496 kubelet[2561]: E0114 13:51:20.101425 2561 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Jan 14 13:51:20.381909 kubelet[2561]: I0114 13:51:20.381262 2561 apiserver.go:52] "Watching apiserver" Jan 14 13:51:20.465480 kubelet[2561]: I0114 13:51:20.465160 2561 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 14 13:51:21.118894 kubelet[2561]: I0114 13:51:21.118287 2561 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Jan 14 13:51:21.149329 kubelet[2561]: E0114 13:51:21.149298 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:22.032937 kubelet[2561]: E0114 13:51:22.032895 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:27.095828 systemd[1]: Reload requested from client PID 2860 ('systemctl') (unit session-8.scope)... Jan 14 13:51:27.095857 systemd[1]: Reloading... Jan 14 13:51:27.150147 kubelet[2561]: I0114 13:51:27.149918 2561 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Jan 14 13:51:28.453098 kubelet[2561]: I0114 13:51:28.452321 2561 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=7.452263006 podStartE2EDuration="7.452263006s" podCreationTimestamp="2026-01-14 13:51:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-14 13:51:21.940921128 +0000 UTC m=+12.061886068" watchObservedRunningTime="2026-01-14 13:51:28.452263006 +0000 UTC m=+18.573227947" Jan 14 13:51:28.546493 kubelet[2561]: E0114 13:51:28.473052 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:29.159432 kubelet[2561]: E0114 13:51:29.154441 2561 kubelet.go:2617] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.334s" Jan 14 13:51:29.204263 kubelet[2561]: I0114 13:51:29.204187 2561 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Jan 14 13:51:29.404812 kubelet[2561]: E0114 13:51:29.402912 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:29.707425 kubelet[2561]: I0114 13:51:29.700490 2561 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=2.700463543 podStartE2EDuration="2.700463543s" podCreationTimestamp="2026-01-14 13:51:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-14 13:51:29.472222358 +0000 UTC m=+19.593187329" watchObservedRunningTime="2026-01-14 13:51:29.700463543 +0000 UTC m=+19.821428485" Jan 14 13:51:29.795479 kubelet[2561]: E0114 13:51:29.793801 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:29.826047 kubelet[2561]: E0114 13:51:29.817262 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:29.839218 zram_generator::config[2907]: No configuration found. Jan 14 13:51:31.173879 kubelet[2561]: E0114 13:51:31.172794 2561 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:31.282101 kubelet[2561]: I0114 13:51:31.273345 2561 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=2.273322614 podStartE2EDuration="2.273322614s" podCreationTimestamp="2026-01-14 13:51:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-14 13:51:29.74913729 +0000 UTC m=+19.870102251" watchObservedRunningTime="2026-01-14 13:51:31.273322614 +0000 UTC m=+21.394287556" Jan 14 13:51:31.320871 systemd[1]: Reloading finished in 4203 ms. Jan 14 13:51:31.487215 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 13:51:31.525146 systemd[1]: kubelet.service: Deactivated successfully. Jan 14 13:51:31.527494 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 13:51:31.529000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:51:31.530442 systemd[1]: kubelet.service: Consumed 7.697s CPU time, 130.8M memory peak. Jan 14 13:51:31.541900 kernel: kauditd_printk_skb: 100 callbacks suppressed Jan 14 13:51:31.552137 kernel: audit: type=1131 audit(1768398691.529:392): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:51:31.555849 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 13:51:31.585136 kernel: audit: type=1334 audit(1768398691.557:393): prog-id=111 op=LOAD Jan 14 13:51:31.557000 audit: BPF prog-id=111 op=LOAD Jan 14 13:51:31.558000 audit: BPF prog-id=62 op=UNLOAD Jan 14 13:51:31.617260 kernel: audit: type=1334 audit(1768398691.558:394): prog-id=62 op=UNLOAD Jan 14 13:51:31.619243 kernel: audit: type=1334 audit(1768398691.571:395): prog-id=112 op=LOAD Jan 14 13:51:31.571000 audit: BPF prog-id=112 op=LOAD Jan 14 13:51:31.572000 audit: BPF prog-id=61 op=UNLOAD Jan 14 13:51:31.577000 audit: BPF prog-id=113 op=LOAD Jan 14 13:51:31.673445 kernel: audit: type=1334 audit(1768398691.572:396): prog-id=61 op=UNLOAD Jan 14 13:51:31.680841 kernel: audit: type=1334 audit(1768398691.577:397): prog-id=113 op=LOAD Jan 14 13:51:31.680889 kernel: audit: type=1334 audit(1768398691.577:398): prog-id=73 op=UNLOAD Jan 14 13:51:31.577000 audit: BPF prog-id=73 op=UNLOAD Jan 14 13:51:31.686777 kernel: audit: type=1334 audit(1768398691.577:399): prog-id=114 op=LOAD Jan 14 13:51:31.577000 audit: BPF prog-id=114 op=LOAD Jan 14 13:51:31.577000 audit: BPF prog-id=115 op=LOAD Jan 14 13:51:31.713433 kernel: audit: type=1334 audit(1768398691.577:400): prog-id=115 op=LOAD Jan 14 13:51:31.714084 kernel: audit: type=1334 audit(1768398691.577:401): prog-id=74 op=UNLOAD Jan 14 13:51:31.577000 audit: BPF prog-id=74 op=UNLOAD Jan 14 13:51:31.577000 audit: BPF prog-id=75 op=UNLOAD Jan 14 13:51:31.587000 audit: BPF prog-id=116 op=LOAD Jan 14 13:51:31.587000 audit: BPF prog-id=76 op=UNLOAD Jan 14 13:51:31.588000 audit: BPF prog-id=117 op=LOAD Jan 14 13:51:31.588000 audit: BPF prog-id=118 op=LOAD Jan 14 13:51:31.588000 audit: BPF prog-id=77 op=UNLOAD Jan 14 13:51:31.588000 audit: BPF prog-id=78 op=UNLOAD Jan 14 13:51:31.591000 audit: BPF prog-id=119 op=LOAD Jan 14 13:51:31.591000 audit: BPF prog-id=63 op=UNLOAD Jan 14 13:51:31.591000 audit: BPF prog-id=120 op=LOAD Jan 14 13:51:31.592000 audit: BPF prog-id=121 op=LOAD Jan 14 13:51:31.592000 audit: BPF prog-id=64 op=UNLOAD Jan 14 13:51:31.592000 audit: BPF prog-id=65 op=UNLOAD Jan 14 13:51:31.595000 audit: BPF prog-id=122 op=LOAD Jan 14 13:51:31.595000 audit: BPF prog-id=69 op=UNLOAD Jan 14 13:51:31.601000 audit: BPF prog-id=123 op=LOAD Jan 14 13:51:31.601000 audit: BPF prog-id=66 op=UNLOAD Jan 14 13:51:31.601000 audit: BPF prog-id=124 op=LOAD Jan 14 13:51:31.601000 audit: BPF prog-id=125 op=LOAD Jan 14 13:51:31.601000 audit: BPF prog-id=67 op=UNLOAD Jan 14 13:51:31.601000 audit: BPF prog-id=68 op=UNLOAD Jan 14 13:51:31.602000 audit: BPF prog-id=126 op=LOAD Jan 14 13:51:31.602000 audit: BPF prog-id=127 op=LOAD Jan 14 13:51:31.602000 audit: BPF prog-id=79 op=UNLOAD Jan 14 13:51:31.602000 audit: BPF prog-id=80 op=UNLOAD Jan 14 13:51:31.604000 audit: BPF prog-id=128 op=LOAD Jan 14 13:51:31.604000 audit: BPF prog-id=70 op=UNLOAD Jan 14 13:51:31.604000 audit: BPF prog-id=129 op=LOAD Jan 14 13:51:31.604000 audit: BPF prog-id=130 op=LOAD Jan 14 13:51:31.604000 audit: BPF prog-id=71 op=UNLOAD Jan 14 13:51:31.604000 audit: BPF prog-id=72 op=UNLOAD Jan 14 13:51:32.629000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:51:32.629286 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 13:51:32.681913 (kubelet)[2951]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 14 13:51:34.298233 kubelet[2951]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jan 14 13:51:34.298233 kubelet[2951]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 14 13:51:34.408418 kubelet[2951]: I0114 13:51:34.299188 2951 server.go:213] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 14 13:51:34.796482 kubelet[2951]: I0114 13:51:34.790272 2951 server.go:529] "Kubelet version" kubeletVersion="v1.34.1" Jan 14 13:51:34.796482 kubelet[2951]: I0114 13:51:34.793286 2951 server.go:531] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 14 13:51:34.820243 kubelet[2951]: I0114 13:51:34.797866 2951 watchdog_linux.go:95] "Systemd watchdog is not enabled" Jan 14 13:51:34.820243 kubelet[2951]: I0114 13:51:34.798145 2951 watchdog_linux.go:137] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jan 14 13:51:34.820243 kubelet[2951]: I0114 13:51:34.814464 2951 server.go:956] "Client rotation is on, will bootstrap in background" Jan 14 13:51:35.026387 kubelet[2951]: I0114 13:51:35.025298 2951 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Jan 14 13:51:35.122820 kubelet[2951]: I0114 13:51:35.120280 2951 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 14 13:51:35.803304 kubelet[2951]: I0114 13:51:35.802137 2951 server.go:1423] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 14 13:51:36.082771 kubelet[2951]: I0114 13:51:36.076417 2951 server.go:781] "--cgroups-per-qos enabled, but --cgroup-root was not specified. Defaulting to /" Jan 14 13:51:36.098168 kubelet[2951]: I0114 13:51:36.093082 2951 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 14 13:51:36.118331 kubelet[2951]: I0114 13:51:36.104785 2951 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 14 13:51:36.128781 kubelet[2951]: I0114 13:51:36.121091 2951 topology_manager.go:138] "Creating topology manager with none policy" Jan 14 13:51:36.128781 kubelet[2951]: I0114 13:51:36.121924 2951 container_manager_linux.go:306] "Creating device plugin manager" Jan 14 13:51:36.132092 kubelet[2951]: I0114 13:51:36.131380 2951 container_manager_linux.go:315] "Creating Dynamic Resource Allocation (DRA) manager" Jan 14 13:51:36.155394 kubelet[2951]: I0114 13:51:36.154917 2951 state_mem.go:36] "Initialized new in-memory state store" Jan 14 13:51:36.180453 kubelet[2951]: I0114 13:51:36.180142 2951 kubelet.go:475] "Attempting to sync node with API server" Jan 14 13:51:36.184448 kubelet[2951]: I0114 13:51:36.182168 2951 kubelet.go:376] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 14 13:51:36.184448 kubelet[2951]: I0114 13:51:36.183425 2951 kubelet.go:387] "Adding apiserver pod source" Jan 14 13:51:36.184448 kubelet[2951]: I0114 13:51:36.184108 2951 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 14 13:51:36.293837 kubelet[2951]: I0114 13:51:36.290846 2951 kuberuntime_manager.go:291] "Container runtime initialized" containerRuntime="containerd" version="v2.1.5" apiVersion="v1" Jan 14 13:51:36.355780 kubelet[2951]: I0114 13:51:36.355227 2951 kubelet.go:940] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Jan 14 13:51:36.355780 kubelet[2951]: I0114 13:51:36.355290 2951 kubelet.go:964] "Not starting PodCertificateRequest manager because we are in static kubelet mode or the PodCertificateProjection feature gate is disabled" Jan 14 13:51:36.589253 kubelet[2951]: I0114 13:51:36.586400 2951 server.go:1262] "Started kubelet" Jan 14 13:51:36.594246 kubelet[2951]: I0114 13:51:36.592330 2951 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Jan 14 13:51:36.631100 kubelet[2951]: I0114 13:51:36.608307 2951 ratelimit.go:56] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 14 13:51:36.678465 kubelet[2951]: I0114 13:51:36.675497 2951 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 14 13:51:36.719123 kubelet[2951]: I0114 13:51:36.718116 2951 volume_manager.go:313] "Starting Kubelet Volume Manager" Jan 14 13:51:36.733930 kubelet[2951]: I0114 13:51:36.732343 2951 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 14 13:51:36.756894 kubelet[2951]: I0114 13:51:36.753496 2951 server_v1.go:49] "podresources" method="list" useActivePods=true Jan 14 13:51:36.809467 kubelet[2951]: I0114 13:51:36.757222 2951 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 14 13:51:36.809467 kubelet[2951]: I0114 13:51:36.781255 2951 server.go:249] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 14 13:51:36.809467 kubelet[2951]: I0114 13:51:36.809796 2951 reconciler.go:29] "Reconciler: start to sync state" Jan 14 13:51:36.809467 kubelet[2951]: I0114 13:51:36.810499 2951 server.go:310] "Adding debug handlers to kubelet server" Jan 14 13:51:37.216923 kubelet[2951]: I0114 13:51:37.216297 2951 apiserver.go:52] "Watching apiserver" Jan 14 13:51:37.245929 kubelet[2951]: I0114 13:51:37.203349 2951 factory.go:223] Registration of the systemd container factory successfully Jan 14 13:51:37.281139 kubelet[2951]: I0114 13:51:37.253195 2951 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 14 13:51:37.402792 kubelet[2951]: E0114 13:51:37.402400 2951 kubelet.go:1615] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 14 13:51:37.554427 kubelet[2951]: I0114 13:51:37.531064 2951 factory.go:223] Registration of the containerd container factory successfully Jan 14 13:51:37.786443 kubelet[2951]: I0114 13:51:37.783412 2951 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv4" Jan 14 13:51:37.886179 kubelet[2951]: I0114 13:51:37.847129 2951 kubelet_network_linux.go:54] "Initialized iptables rules." protocol="IPv6" Jan 14 13:51:38.244127 kubelet[2951]: I0114 13:51:38.154405 2951 status_manager.go:244] "Starting to sync pod status with apiserver" Jan 14 13:51:38.393901 kubelet[2951]: I0114 13:51:38.377374 2951 kubelet.go:2427] "Starting kubelet main sync loop" Jan 14 13:51:38.393901 kubelet[2951]: E0114 13:51:38.391863 2951 kubelet.go:2451] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 14 13:51:38.796476 kubelet[2951]: E0114 13:51:38.607371 2951 kubelet.go:2451] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 14 13:51:38.840380 kubelet[2951]: E0114 13:51:38.829372 2951 kubelet.go:2451] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 14 13:51:39.252212 kubelet[2951]: E0114 13:51:39.250925 2951 kubelet.go:2451] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jan 14 13:51:40.092798 kubelet[2951]: E0114 13:51:40.091190 2951 kubelet.go:2451] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jan 14 13:51:41.541375 kubelet[2951]: I0114 13:51:41.540376 2951 cpu_manager.go:221] "Starting CPU manager" policy="none" Jan 14 13:51:41.541375 kubelet[2951]: I0114 13:51:41.540505 2951 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jan 14 13:51:41.541375 kubelet[2951]: I0114 13:51:41.541246 2951 state_mem.go:36] "Initialized new in-memory state store" Jan 14 13:51:41.558425 kubelet[2951]: I0114 13:51:41.554111 2951 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jan 14 13:51:41.558425 kubelet[2951]: I0114 13:51:41.554274 2951 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jan 14 13:51:41.574067 kubelet[2951]: I0114 13:51:41.562506 2951 policy_none.go:49] "None policy: Start" Jan 14 13:51:41.574067 kubelet[2951]: I0114 13:51:41.571771 2951 memory_manager.go:187] "Starting memorymanager" policy="None" Jan 14 13:51:41.582818 kubelet[2951]: I0114 13:51:41.574350 2951 state_mem.go:36] "Initializing new in-memory state store" logger="Memory Manager state checkpoint" Jan 14 13:51:41.582818 kubelet[2951]: I0114 13:51:41.579896 2951 state_mem.go:77] "Updated machine memory state" logger="Memory Manager state checkpoint" Jan 14 13:51:41.582818 kubelet[2951]: I0114 13:51:41.581888 2951 policy_none.go:47] "Start" Jan 14 13:51:41.698375 kubelet[2951]: E0114 13:51:41.697485 2951 kubelet.go:2451] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jan 14 13:51:41.708316 kubelet[2951]: E0114 13:51:41.706813 2951 manager.go:513] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Jan 14 13:51:41.721346 kubelet[2951]: I0114 13:51:41.719094 2951 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 14 13:51:41.723125 kubelet[2951]: I0114 13:51:41.722135 2951 container_log_manager.go:146] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 14 13:51:41.742440 kubelet[2951]: I0114 13:51:41.741479 2951 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 14 13:51:41.789396 kubelet[2951]: I0114 13:51:41.788304 2951 kuberuntime_manager.go:1828] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jan 14 13:51:41.805455 containerd[1635]: time="2026-01-14T13:51:41.799290401Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jan 14 13:51:41.871428 kubelet[2951]: I0114 13:51:41.862355 2951 kubelet_network.go:47] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jan 14 13:51:41.901126 kubelet[2951]: E0114 13:51:41.897342 2951 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jan 14 13:51:42.331228 kubelet[2951]: I0114 13:51:42.326069 2951 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 14 13:51:42.621181 kubelet[2951]: I0114 13:51:42.619208 2951 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Jan 14 13:51:42.701084 kubelet[2951]: I0114 13:51:42.623452 2951 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Jan 14 13:51:44.936477 kubelet[2951]: I0114 13:51:44.935240 2951 kubelet.go:3219] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Jan 14 13:51:44.974874 kubelet[2951]: I0114 13:51:44.974394 2951 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 14 13:51:45.044800 kubelet[2951]: E0114 13:51:45.039280 2951 kubelet.go:3221] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Jan 14 13:51:45.139730 kubelet[2951]: I0114 13:51:45.139144 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 14 13:51:45.139730 kubelet[2951]: I0114 13:51:45.139195 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 14 13:51:45.139730 kubelet[2951]: I0114 13:51:45.139392 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4pgx\" (UniqueName: \"kubernetes.io/projected/dda2e25d-1c0b-4953-bf40-c3e0ec54eebc-kube-api-access-p4pgx\") pod \"kube-proxy-9csc2\" (UID: \"dda2e25d-1c0b-4953-bf40-c3e0ec54eebc\") " pod="kube-system/kube-proxy-9csc2" Jan 14 13:51:45.139730 kubelet[2951]: I0114 13:51:45.139411 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0ae88e709ded7aa62d05a5e6bfe67ee8-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"0ae88e709ded7aa62d05a5e6bfe67ee8\") " pod="kube-system/kube-apiserver-localhost" Jan 14 13:51:45.139730 kubelet[2951]: I0114 13:51:45.139427 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0ae88e709ded7aa62d05a5e6bfe67ee8-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"0ae88e709ded7aa62d05a5e6bfe67ee8\") " pod="kube-system/kube-apiserver-localhost" Jan 14 13:51:45.140386 kubelet[2951]: I0114 13:51:45.139441 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 14 13:51:45.140386 kubelet[2951]: I0114 13:51:45.139455 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/07ca0cbf79ad6ba9473d8e9f7715e571-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"07ca0cbf79ad6ba9473d8e9f7715e571\") " pod="kube-system/kube-scheduler-localhost" Jan 14 13:51:45.140386 kubelet[2951]: I0114 13:51:45.139468 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/dda2e25d-1c0b-4953-bf40-c3e0ec54eebc-kube-proxy\") pod \"kube-proxy-9csc2\" (UID: \"dda2e25d-1c0b-4953-bf40-c3e0ec54eebc\") " pod="kube-system/kube-proxy-9csc2" Jan 14 13:51:45.140386 kubelet[2951]: I0114 13:51:45.139482 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/dda2e25d-1c0b-4953-bf40-c3e0ec54eebc-xtables-lock\") pod \"kube-proxy-9csc2\" (UID: \"dda2e25d-1c0b-4953-bf40-c3e0ec54eebc\") " pod="kube-system/kube-proxy-9csc2" Jan 14 13:51:45.140386 kubelet[2951]: I0114 13:51:45.139494 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/dda2e25d-1c0b-4953-bf40-c3e0ec54eebc-lib-modules\") pod \"kube-proxy-9csc2\" (UID: \"dda2e25d-1c0b-4953-bf40-c3e0ec54eebc\") " pod="kube-system/kube-proxy-9csc2" Jan 14 13:51:45.140843 kubelet[2951]: I0114 13:51:45.139505 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0ae88e709ded7aa62d05a5e6bfe67ee8-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"0ae88e709ded7aa62d05a5e6bfe67ee8\") " pod="kube-system/kube-apiserver-localhost" Jan 14 13:51:45.140843 kubelet[2951]: I0114 13:51:45.140199 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 14 13:51:45.140843 kubelet[2951]: I0114 13:51:45.140215 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5bbfee13ce9e07281eca876a0b8067f2-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5bbfee13ce9e07281eca876a0b8067f2\") " pod="kube-system/kube-controller-manager-localhost" Jan 14 13:51:45.152910 systemd[1]: Created slice kubepods-besteffort-poddda2e25d_1c0b_4953_bf40_c3e0ec54eebc.slice - libcontainer container kubepods-besteffort-poddda2e25d_1c0b_4953_bf40_c3e0ec54eebc.slice. Jan 14 13:51:45.366466 kubelet[2951]: E0114 13:51:45.366232 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:45.453199 kubelet[2951]: E0114 13:51:45.451843 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:45.532227 kubelet[2951]: E0114 13:51:45.531228 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:45.555121 kubelet[2951]: E0114 13:51:45.553842 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:45.864904 kubelet[2951]: E0114 13:51:45.855476 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:45.872434 containerd[1635]: time="2026-01-14T13:51:45.870282596Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-9csc2,Uid:dda2e25d-1c0b-4953-bf40-c3e0ec54eebc,Namespace:kube-system,Attempt:0,}" Jan 14 13:51:46.225267 systemd[1]: Created slice kubepods-besteffort-pod5837ffc0_f616_411c_a861_cccb9374c9c4.slice - libcontainer container kubepods-besteffort-pod5837ffc0_f616_411c_a861_cccb9374c9c4.slice. Jan 14 13:51:46.314220 kubelet[2951]: I0114 13:51:46.310853 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/5837ffc0-f616-411c-a861-cccb9374c9c4-var-lib-calico\") pod \"tigera-operator-65cdcdfd6d-86fkx\" (UID: \"5837ffc0-f616-411c-a861-cccb9374c9c4\") " pod="tigera-operator/tigera-operator-65cdcdfd6d-86fkx" Jan 14 13:51:46.314220 kubelet[2951]: I0114 13:51:46.311075 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j24vk\" (UniqueName: \"kubernetes.io/projected/5837ffc0-f616-411c-a861-cccb9374c9c4-kube-api-access-j24vk\") pod \"tigera-operator-65cdcdfd6d-86fkx\" (UID: \"5837ffc0-f616-411c-a861-cccb9374c9c4\") " pod="tigera-operator/tigera-operator-65cdcdfd6d-86fkx" Jan 14 13:51:46.330178 containerd[1635]: time="2026-01-14T13:51:46.319412078Z" level=info msg="connecting to shim c376e84e49caad36aa45b22144deacfd7c14e6335dfbfad4f19f278e4375b06e" address="unix:///run/containerd/s/96f5acf8a70e3e6df5e9f0f4ba7bd7bc4f82ea5052671069283ce894d830bba7" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:51:46.508235 kubelet[2951]: E0114 13:51:46.505846 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:46.518297 kubelet[2951]: E0114 13:51:46.512495 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:46.760408 systemd[1]: Started cri-containerd-c376e84e49caad36aa45b22144deacfd7c14e6335dfbfad4f19f278e4375b06e.scope - libcontainer container c376e84e49caad36aa45b22144deacfd7c14e6335dfbfad4f19f278e4375b06e. Jan 14 13:51:46.934233 kernel: kauditd_printk_skb: 32 callbacks suppressed Jan 14 13:51:46.968873 kernel: audit: type=1334 audit(1768398706.906:434): prog-id=131 op=LOAD Jan 14 13:51:46.984049 kernel: audit: type=1334 audit(1768398706.947:435): prog-id=132 op=LOAD Jan 14 13:51:46.984165 kernel: audit: type=1300 audit(1768398706.947:435): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00018c238 a2=98 a3=0 items=0 ppid=3025 pid=3036 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:46.906000 audit: BPF prog-id=131 op=LOAD Jan 14 13:51:46.947000 audit: BPF prog-id=132 op=LOAD Jan 14 13:51:46.947000 audit[3036]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00018c238 a2=98 a3=0 items=0 ppid=3025 pid=3036 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:46.987859 containerd[1635]: time="2026-01-14T13:51:46.925249248Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-65cdcdfd6d-86fkx,Uid:5837ffc0-f616-411c-a861-cccb9374c9c4,Namespace:tigera-operator,Attempt:0,}" Jan 14 13:51:46.947000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333373665383465343963616164333661613435623232313434646561 Jan 14 13:51:47.123879 kernel: audit: type=1327 audit(1768398706.947:435): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333373665383465343963616164333661613435623232313434646561 Jan 14 13:51:47.127302 kernel: audit: type=1334 audit(1768398706.947:436): prog-id=132 op=UNLOAD Jan 14 13:51:46.947000 audit: BPF prog-id=132 op=UNLOAD Jan 14 13:51:46.947000 audit[3036]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3025 pid=3036 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:46.947000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333373665383465343963616164333661613435623232313434646561 Jan 14 13:51:47.244439 containerd[1635]: time="2026-01-14T13:51:47.216379503Z" level=info msg="connecting to shim 76ca8881a31453feb9f0da5ee45296536a0d894afcc0c120f225f6849c64e3a7" address="unix:///run/containerd/s/476691f8fcf272a288e87caea25effbde5e0949e0c1a002074fb6497c7b85eda" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:51:47.262295 kernel: audit: type=1300 audit(1768398706.947:436): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3025 pid=3036 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:47.263850 kernel: audit: type=1327 audit(1768398706.947:436): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333373665383465343963616164333661613435623232313434646561 Jan 14 13:51:46.949000 audit: BPF prog-id=133 op=LOAD Jan 14 13:51:46.949000 audit[3036]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00018c488 a2=98 a3=0 items=0 ppid=3025 pid=3036 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:47.342893 kernel: audit: type=1334 audit(1768398706.949:437): prog-id=133 op=LOAD Jan 14 13:51:47.343259 kernel: audit: type=1300 audit(1768398706.949:437): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00018c488 a2=98 a3=0 items=0 ppid=3025 pid=3036 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:47.343792 kernel: audit: type=1327 audit(1768398706.949:437): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333373665383465343963616164333661613435623232313434646561 Jan 14 13:51:46.949000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333373665383465343963616164333661613435623232313434646561 Jan 14 13:51:46.949000 audit: BPF prog-id=134 op=LOAD Jan 14 13:51:46.949000 audit[3036]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00018c218 a2=98 a3=0 items=0 ppid=3025 pid=3036 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:46.949000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333373665383465343963616164333661613435623232313434646561 Jan 14 13:51:46.949000 audit: BPF prog-id=134 op=UNLOAD Jan 14 13:51:46.949000 audit[3036]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3025 pid=3036 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:46.949000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333373665383465343963616164333661613435623232313434646561 Jan 14 13:51:46.949000 audit: BPF prog-id=133 op=UNLOAD Jan 14 13:51:46.949000 audit[3036]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3025 pid=3036 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:46.949000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333373665383465343963616164333661613435623232313434646561 Jan 14 13:51:46.949000 audit: BPF prog-id=135 op=LOAD Jan 14 13:51:46.949000 audit[3036]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00018c6e8 a2=98 a3=0 items=0 ppid=3025 pid=3036 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:46.949000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6333373665383465343963616164333661613435623232313434646561 Jan 14 13:51:47.433311 containerd[1635]: time="2026-01-14T13:51:47.433169659Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-9csc2,Uid:dda2e25d-1c0b-4953-bf40-c3e0ec54eebc,Namespace:kube-system,Attempt:0,} returns sandbox id \"c376e84e49caad36aa45b22144deacfd7c14e6335dfbfad4f19f278e4375b06e\"" Jan 14 13:51:47.462293 kubelet[2951]: E0114 13:51:47.460147 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:47.460329 systemd[1]: Started cri-containerd-76ca8881a31453feb9f0da5ee45296536a0d894afcc0c120f225f6849c64e3a7.scope - libcontainer container 76ca8881a31453feb9f0da5ee45296536a0d894afcc0c120f225f6849c64e3a7. Jan 14 13:51:47.515478 containerd[1635]: time="2026-01-14T13:51:47.515056661Z" level=info msg="CreateContainer within sandbox \"c376e84e49caad36aa45b22144deacfd7c14e6335dfbfad4f19f278e4375b06e\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jan 14 13:51:47.537883 kubelet[2951]: E0114 13:51:47.537331 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:47.610398 containerd[1635]: time="2026-01-14T13:51:47.607148285Z" level=info msg="Container 6c8648321e949572420f9aace55d524149eb80b6e1ea5aa5b9b770cd35b9d764: CDI devices from CRI Config.CDIDevices: []" Jan 14 13:51:47.609262 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4218870665.mount: Deactivated successfully. Jan 14 13:51:47.661000 audit: BPF prog-id=136 op=LOAD Jan 14 13:51:47.662000 audit: BPF prog-id=137 op=LOAD Jan 14 13:51:47.662000 audit[3081]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=3064 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:47.662000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736636138383831613331343533666562396630646135656534353239 Jan 14 13:51:47.662000 audit: BPF prog-id=137 op=UNLOAD Jan 14 13:51:47.662000 audit[3081]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3064 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:47.662000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736636138383831613331343533666562396630646135656534353239 Jan 14 13:51:47.663000 audit: BPF prog-id=138 op=LOAD Jan 14 13:51:47.663000 audit[3081]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3064 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:47.663000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736636138383831613331343533666562396630646135656534353239 Jan 14 13:51:47.663000 audit: BPF prog-id=139 op=LOAD Jan 14 13:51:47.663000 audit[3081]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3064 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:47.663000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736636138383831613331343533666562396630646135656534353239 Jan 14 13:51:47.663000 audit: BPF prog-id=139 op=UNLOAD Jan 14 13:51:47.663000 audit[3081]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3064 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:47.663000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736636138383831613331343533666562396630646135656534353239 Jan 14 13:51:47.663000 audit: BPF prog-id=138 op=UNLOAD Jan 14 13:51:47.663000 audit[3081]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3064 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:47.663000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736636138383831613331343533666562396630646135656534353239 Jan 14 13:51:47.663000 audit: BPF prog-id=140 op=LOAD Jan 14 13:51:47.663000 audit[3081]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=3064 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:47.663000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736636138383831613331343533666562396630646135656534353239 Jan 14 13:51:47.673320 containerd[1635]: time="2026-01-14T13:51:47.671235208Z" level=info msg="CreateContainer within sandbox \"c376e84e49caad36aa45b22144deacfd7c14e6335dfbfad4f19f278e4375b06e\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"6c8648321e949572420f9aace55d524149eb80b6e1ea5aa5b9b770cd35b9d764\"" Jan 14 13:51:47.675389 containerd[1635]: time="2026-01-14T13:51:47.674129812Z" level=info msg="StartContainer for \"6c8648321e949572420f9aace55d524149eb80b6e1ea5aa5b9b770cd35b9d764\"" Jan 14 13:51:47.680371 containerd[1635]: time="2026-01-14T13:51:47.680329889Z" level=info msg="connecting to shim 6c8648321e949572420f9aace55d524149eb80b6e1ea5aa5b9b770cd35b9d764" address="unix:///run/containerd/s/96f5acf8a70e3e6df5e9f0f4ba7bd7bc4f82ea5052671069283ce894d830bba7" protocol=ttrpc version=3 Jan 14 13:51:48.011450 systemd[1]: Started cri-containerd-6c8648321e949572420f9aace55d524149eb80b6e1ea5aa5b9b770cd35b9d764.scope - libcontainer container 6c8648321e949572420f9aace55d524149eb80b6e1ea5aa5b9b770cd35b9d764. Jan 14 13:51:48.908306 kubelet[2951]: E0114 13:51:48.906346 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:49.919099 kubelet[2951]: E0114 13:51:49.917451 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:49.981475 containerd[1635]: time="2026-01-14T13:51:49.980419953Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-65cdcdfd6d-86fkx,Uid:5837ffc0-f616-411c-a861-cccb9374c9c4,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"76ca8881a31453feb9f0da5ee45296536a0d894afcc0c120f225f6849c64e3a7\"" Jan 14 13:51:49.993361 containerd[1635]: time="2026-01-14T13:51:49.993187347Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\"" Jan 14 13:51:50.042000 audit: BPF prog-id=141 op=LOAD Jan 14 13:51:50.042000 audit[3101]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001b0488 a2=98 a3=0 items=0 ppid=3025 pid=3101 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:50.042000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3663383634383332316539343935373234323066396161636535356435 Jan 14 13:51:50.044000 audit: BPF prog-id=142 op=LOAD Jan 14 13:51:50.044000 audit[3101]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001b0218 a2=98 a3=0 items=0 ppid=3025 pid=3101 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:50.044000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3663383634383332316539343935373234323066396161636535356435 Jan 14 13:51:50.044000 audit: BPF prog-id=142 op=UNLOAD Jan 14 13:51:50.044000 audit[3101]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3025 pid=3101 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:50.044000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3663383634383332316539343935373234323066396161636535356435 Jan 14 13:51:50.044000 audit: BPF prog-id=141 op=UNLOAD Jan 14 13:51:50.044000 audit[3101]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3025 pid=3101 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:50.044000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3663383634383332316539343935373234323066396161636535356435 Jan 14 13:51:50.044000 audit: BPF prog-id=143 op=LOAD Jan 14 13:51:50.044000 audit[3101]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001b06e8 a2=98 a3=0 items=0 ppid=3025 pid=3101 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:50.044000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3663383634383332316539343935373234323066396161636535356435 Jan 14 13:51:50.211196 containerd[1635]: time="2026-01-14T13:51:50.209864854Z" level=info msg="StartContainer for \"6c8648321e949572420f9aace55d524149eb80b6e1ea5aa5b9b770cd35b9d764\" returns successfully" Jan 14 13:51:50.921290 kubelet[2951]: E0114 13:51:50.920501 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:51.246000 audit[3177]: NETFILTER_CFG table=mangle:54 family=2 entries=1 op=nft_register_chain pid=3177 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.246000 audit[3177]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffd4f24b290 a2=0 a3=7ffd4f24b27c items=0 ppid=3117 pid=3177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.246000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Jan 14 13:51:51.275000 audit[3176]: NETFILTER_CFG table=mangle:55 family=10 entries=1 op=nft_register_chain pid=3176 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:51.275000 audit[3176]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fff2b45e620 a2=0 a3=7fff2b45e60c items=0 ppid=3117 pid=3176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.275000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Jan 14 13:51:51.304000 audit[3183]: NETFILTER_CFG table=nat:56 family=10 entries=1 op=nft_register_chain pid=3183 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:51.304000 audit[3183]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffcae1bb910 a2=0 a3=7ffcae1bb8fc items=0 ppid=3117 pid=3183 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.304000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Jan 14 13:51:51.307000 audit[3180]: NETFILTER_CFG table=nat:57 family=2 entries=1 op=nft_register_chain pid=3180 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.307000 audit[3180]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe73582fe0 a2=0 a3=7ffe73582fcc items=0 ppid=3117 pid=3180 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.307000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Jan 14 13:51:51.318000 audit[3184]: NETFILTER_CFG table=filter:58 family=10 entries=1 op=nft_register_chain pid=3184 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:51.318000 audit[3184]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffcac84aa10 a2=0 a3=7ffcac84a9fc items=0 ppid=3117 pid=3184 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.318000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Jan 14 13:51:51.323000 audit[3185]: NETFILTER_CFG table=filter:59 family=2 entries=1 op=nft_register_chain pid=3185 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.323000 audit[3185]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffc15e9ba40 a2=0 a3=7ffc15e9ba2c items=0 ppid=3117 pid=3185 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.323000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Jan 14 13:51:51.370000 audit[3186]: NETFILTER_CFG table=filter:60 family=2 entries=1 op=nft_register_chain pid=3186 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.370000 audit[3186]: SYSCALL arch=c000003e syscall=46 success=yes exit=108 a0=3 a1=7ffe6e12f860 a2=0 a3=7ffe6e12f84c items=0 ppid=3117 pid=3186 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.370000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 Jan 14 13:51:51.403000 audit[3188]: NETFILTER_CFG table=filter:61 family=2 entries=1 op=nft_register_rule pid=3188 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.403000 audit[3188]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7fff57088150 a2=0 a3=7fff5708813c items=0 ppid=3117 pid=3188 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.403000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669636520706F7274616C73002D Jan 14 13:51:51.451000 audit[3191]: NETFILTER_CFG table=filter:62 family=2 entries=1 op=nft_register_rule pid=3191 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.451000 audit[3191]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffc1efd7b10 a2=0 a3=7ffc1efd7afc items=0 ppid=3117 pid=3191 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.451000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669636520706F7274616C73 Jan 14 13:51:51.465000 audit[3192]: NETFILTER_CFG table=filter:63 family=2 entries=1 op=nft_register_chain pid=3192 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.465000 audit[3192]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffcb901baa0 a2=0 a3=7ffcb901ba8c items=0 ppid=3117 pid=3192 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.465000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D4E4F4445504F525453002D740066696C746572 Jan 14 13:51:51.500000 audit[3194]: NETFILTER_CFG table=filter:64 family=2 entries=1 op=nft_register_rule pid=3194 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.500000 audit[3194]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7fff534e3610 a2=0 a3=7fff534e35fc items=0 ppid=3117 pid=3194 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.500000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 Jan 14 13:51:51.521000 audit[3195]: NETFILTER_CFG table=filter:65 family=2 entries=1 op=nft_register_chain pid=3195 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.521000 audit[3195]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff17f8c320 a2=0 a3=7fff17f8c30c items=0 ppid=3117 pid=3195 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.521000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D5345525649434553002D740066696C746572 Jan 14 13:51:51.561000 audit[3197]: NETFILTER_CFG table=filter:66 family=2 entries=1 op=nft_register_rule pid=3197 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.561000 audit[3197]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffcd81e7640 a2=0 a3=7ffcd81e762c items=0 ppid=3117 pid=3197 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.561000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 14 13:51:51.603000 audit[3200]: NETFILTER_CFG table=filter:67 family=2 entries=1 op=nft_register_rule pid=3200 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.603000 audit[3200]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffc65e1ecc0 a2=0 a3=7ffc65e1ecac items=0 ppid=3117 pid=3200 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.603000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 14 13:51:51.629000 audit[3201]: NETFILTER_CFG table=filter:68 family=2 entries=1 op=nft_register_chain pid=3201 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.629000 audit[3201]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffdd6df5e90 a2=0 a3=7ffdd6df5e7c items=0 ppid=3117 pid=3201 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.629000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D464F5257415244002D740066696C746572 Jan 14 13:51:51.676000 audit[3203]: NETFILTER_CFG table=filter:69 family=2 entries=1 op=nft_register_rule pid=3203 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.676000 audit[3203]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffef01b42c0 a2=0 a3=7ffef01b42ac items=0 ppid=3117 pid=3203 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.676000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 Jan 14 13:51:51.690000 audit[3204]: NETFILTER_CFG table=filter:70 family=2 entries=1 op=nft_register_chain pid=3204 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.690000 audit[3204]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fff6c4f0160 a2=0 a3=7fff6c4f014c items=0 ppid=3117 pid=3204 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.690000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 Jan 14 13:51:51.738000 audit[3206]: NETFILTER_CFG table=filter:71 family=2 entries=1 op=nft_register_rule pid=3206 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.738000 audit[3206]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7fffeba687f0 a2=0 a3=7fffeba687dc items=0 ppid=3117 pid=3206 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.738000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A004B5542452D50524F5859 Jan 14 13:51:51.797000 audit[3209]: NETFILTER_CFG table=filter:72 family=2 entries=1 op=nft_register_rule pid=3209 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.797000 audit[3209]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffd3c7ee330 a2=0 a3=7ffd3c7ee31c items=0 ppid=3117 pid=3209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.797000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A004B5542452D50524F58 Jan 14 13:51:51.872000 audit[3212]: NETFILTER_CFG table=filter:73 family=2 entries=1 op=nft_register_rule pid=3212 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.872000 audit[3212]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffca6d54880 a2=0 a3=7ffca6d5486c items=0 ppid=3117 pid=3212 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.872000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A004B5542452D50524F Jan 14 13:51:51.897000 audit[3213]: NETFILTER_CFG table=nat:74 family=2 entries=1 op=nft_register_chain pid=3213 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.897000 audit[3213]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7fffa37522b0 a2=0 a3=7fffa375229c items=0 ppid=3117 pid=3213 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:51.897000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D5345525649434553002D74006E6174 Jan 14 13:51:52.004842 kernel: kauditd_printk_skb: 112 callbacks suppressed Jan 14 13:51:52.005170 kernel: audit: type=1325 audit(1768398711.955:476): table=nat:75 family=2 entries=1 op=nft_register_rule pid=3215 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:51.955000 audit[3215]: NETFILTER_CFG table=nat:75 family=2 entries=1 op=nft_register_rule pid=3215 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:52.005451 kubelet[2951]: E0114 13:51:51.967292 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:51.955000 audit[3215]: SYSCALL arch=c000003e syscall=46 success=yes exit=524 a0=3 a1=7fffa009e3d0 a2=0 a3=7fffa009e3bc items=0 ppid=3117 pid=3215 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.142502 kernel: audit: type=1300 audit(1768398711.955:476): arch=c000003e syscall=46 success=yes exit=524 a0=3 a1=7fffa009e3d0 a2=0 a3=7fffa009e3bc items=0 ppid=3117 pid=3215 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.142900 kernel: audit: type=1327 audit(1768398711.955:476): proctitle=69707461626C6573002D770035002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 14 13:51:51.955000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 14 13:51:52.246090 kernel: audit: type=1325 audit(1768398712.033:477): table=nat:76 family=2 entries=1 op=nft_register_rule pid=3218 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:52.033000 audit[3218]: NETFILTER_CFG table=nat:76 family=2 entries=1 op=nft_register_rule pid=3218 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:52.033000 audit[3218]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffc50f9d310 a2=0 a3=7ffc50f9d2fc items=0 ppid=3117 pid=3218 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.333248 kernel: audit: type=1300 audit(1768398712.033:477): arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffc50f9d310 a2=0 a3=7ffc50f9d2fc items=0 ppid=3117 pid=3218 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.033000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 14 13:51:52.388819 kernel: audit: type=1327 audit(1768398712.033:477): proctitle=69707461626C6573002D770035002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 14 13:51:52.082000 audit[3219]: NETFILTER_CFG table=nat:77 family=2 entries=1 op=nft_register_chain pid=3219 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:52.425361 kernel: audit: type=1325 audit(1768398712.082:478): table=nat:77 family=2 entries=1 op=nft_register_chain pid=3219 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:52.082000 audit[3219]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff0058ec10 a2=0 a3=7fff0058ebfc items=0 ppid=3117 pid=3219 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.521182 kernel: audit: type=1300 audit(1768398712.082:478): arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff0058ec10 a2=0 a3=7fff0058ebfc items=0 ppid=3117 pid=3219 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.529506 kernel: audit: type=1327 audit(1768398712.082:478): proctitle=69707461626C6573002D770035002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Jan 14 13:51:52.082000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Jan 14 13:51:52.155000 audit[3221]: NETFILTER_CFG table=nat:78 family=2 entries=1 op=nft_register_rule pid=3221 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:52.569364 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3199212634.mount: Deactivated successfully. Jan 14 13:51:52.580340 kernel: audit: type=1325 audit(1768398712.155:479): table=nat:78 family=2 entries=1 op=nft_register_rule pid=3221 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 13:51:52.155000 audit[3221]: SYSCALL arch=c000003e syscall=46 success=yes exit=532 a0=3 a1=7ffcfb312120 a2=0 a3=7ffcfb31210c items=0 ppid=3117 pid=3221 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.155000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 Jan 14 13:51:52.452000 audit[3231]: NETFILTER_CFG table=filter:79 family=2 entries=8 op=nft_register_rule pid=3231 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:51:52.452000 audit[3231]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffe16195910 a2=0 a3=7ffe161958fc items=0 ppid=3117 pid=3231 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.452000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:51:52.600000 audit[3231]: NETFILTER_CFG table=nat:80 family=2 entries=14 op=nft_register_chain pid=3231 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:51:52.600000 audit[3231]: SYSCALL arch=c000003e syscall=46 success=yes exit=5508 a0=3 a1=7ffe16195910 a2=0 a3=7ffe161958fc items=0 ppid=3117 pid=3231 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.600000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:51:52.613000 audit[3236]: NETFILTER_CFG table=filter:81 family=10 entries=1 op=nft_register_chain pid=3236 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:52.613000 audit[3236]: SYSCALL arch=c000003e syscall=46 success=yes exit=108 a0=3 a1=7fff49400300 a2=0 a3=7fff494002ec items=0 ppid=3117 pid=3236 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.613000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 Jan 14 13:51:52.661000 audit[3238]: NETFILTER_CFG table=filter:82 family=10 entries=2 op=nft_register_chain pid=3238 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:52.661000 audit[3238]: SYSCALL arch=c000003e syscall=46 success=yes exit=836 a0=3 a1=7fffaff95b60 a2=0 a3=7fffaff95b4c items=0 ppid=3117 pid=3238 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.661000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669636520706F7274616C73 Jan 14 13:51:52.755000 audit[3241]: NETFILTER_CFG table=filter:83 family=10 entries=1 op=nft_register_rule pid=3241 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:52.755000 audit[3241]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffd457997c0 a2=0 a3=7ffd457997ac items=0 ppid=3117 pid=3241 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.755000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669636520706F7274616C Jan 14 13:51:52.771000 audit[3242]: NETFILTER_CFG table=filter:84 family=10 entries=1 op=nft_register_chain pid=3242 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:52.771000 audit[3242]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd43e544c0 a2=0 a3=7ffd43e544ac items=0 ppid=3117 pid=3242 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.771000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D4E4F4445504F525453002D740066696C746572 Jan 14 13:51:52.808000 audit[3244]: NETFILTER_CFG table=filter:85 family=10 entries=1 op=nft_register_rule pid=3244 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:52.808000 audit[3244]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffe6c3c7870 a2=0 a3=7ffe6c3c785c items=0 ppid=3117 pid=3244 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.808000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 Jan 14 13:51:52.831000 audit[3245]: NETFILTER_CFG table=filter:86 family=10 entries=1 op=nft_register_chain pid=3245 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:52.831000 audit[3245]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe85997700 a2=0 a3=7ffe859976ec items=0 ppid=3117 pid=3245 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.831000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D5345525649434553002D740066696C746572 Jan 14 13:51:52.865000 audit[3247]: NETFILTER_CFG table=filter:87 family=10 entries=1 op=nft_register_rule pid=3247 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:52.865000 audit[3247]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7fff8a096460 a2=0 a3=7fff8a09644c items=0 ppid=3117 pid=3247 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.865000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 14 13:51:52.932000 audit[3250]: NETFILTER_CFG table=filter:88 family=10 entries=2 op=nft_register_chain pid=3250 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:52.932000 audit[3250]: SYSCALL arch=c000003e syscall=46 success=yes exit=828 a0=3 a1=7ffeec20df20 a2=0 a3=7ffeec20df0c items=0 ppid=3117 pid=3250 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.932000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 14 13:51:52.945000 audit[3251]: NETFILTER_CFG table=filter:89 family=10 entries=1 op=nft_register_chain pid=3251 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:52.945000 audit[3251]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc413d7630 a2=0 a3=7ffc413d761c items=0 ppid=3117 pid=3251 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.945000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D464F5257415244002D740066696C746572 Jan 14 13:51:52.981000 audit[3253]: NETFILTER_CFG table=filter:90 family=10 entries=1 op=nft_register_rule pid=3253 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:52.981000 audit[3253]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffde35f9cc0 a2=0 a3=7ffde35f9cac items=0 ppid=3117 pid=3253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.981000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 Jan 14 13:51:52.994000 audit[3254]: NETFILTER_CFG table=filter:91 family=10 entries=1 op=nft_register_chain pid=3254 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:52.994000 audit[3254]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fff6f91f060 a2=0 a3=7fff6f91f04c items=0 ppid=3117 pid=3254 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:52.994000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 Jan 14 13:51:53.028000 audit[3260]: NETFILTER_CFG table=filter:92 family=10 entries=1 op=nft_register_rule pid=3260 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:53.028000 audit[3260]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffc11a3ce70 a2=0 a3=7ffc11a3ce5c items=0 ppid=3117 pid=3260 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:53.028000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A004B5542452D50524F58 Jan 14 13:51:53.089000 audit[3263]: NETFILTER_CFG table=filter:93 family=10 entries=1 op=nft_register_rule pid=3263 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:53.089000 audit[3263]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffe0a4b8d40 a2=0 a3=7ffe0a4b8d2c items=0 ppid=3117 pid=3263 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:53.089000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A004B5542452D50524F Jan 14 13:51:53.135000 audit[3266]: NETFILTER_CFG table=filter:94 family=10 entries=1 op=nft_register_rule pid=3266 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:53.135000 audit[3266]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffc797b3050 a2=0 a3=7ffc797b303c items=0 ppid=3117 pid=3266 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:53.135000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A004B5542452D5052 Jan 14 13:51:53.152000 audit[3267]: NETFILTER_CFG table=nat:95 family=10 entries=1 op=nft_register_chain pid=3267 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:53.152000 audit[3267]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffc297215f0 a2=0 a3=7ffc297215dc items=0 ppid=3117 pid=3267 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:53.152000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D5345525649434553002D74006E6174 Jan 14 13:51:53.182000 audit[3269]: NETFILTER_CFG table=nat:96 family=10 entries=1 op=nft_register_rule pid=3269 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:53.182000 audit[3269]: SYSCALL arch=c000003e syscall=46 success=yes exit=524 a0=3 a1=7ffd72da9e40 a2=0 a3=7ffd72da9e2c items=0 ppid=3117 pid=3269 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:53.182000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 14 13:51:53.223000 audit[3272]: NETFILTER_CFG table=nat:97 family=10 entries=1 op=nft_register_rule pid=3272 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:53.223000 audit[3272]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffc6abc6d70 a2=0 a3=7ffc6abc6d5c items=0 ppid=3117 pid=3272 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:53.223000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 14 13:51:53.246000 audit[3273]: NETFILTER_CFG table=nat:98 family=10 entries=1 op=nft_register_chain pid=3273 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:53.246000 audit[3273]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff7dc25f80 a2=0 a3=7fff7dc25f6c items=0 ppid=3117 pid=3273 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:53.246000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Jan 14 13:51:53.295000 audit[3275]: NETFILTER_CFG table=nat:99 family=10 entries=2 op=nft_register_chain pid=3275 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:53.295000 audit[3275]: SYSCALL arch=c000003e syscall=46 success=yes exit=612 a0=3 a1=7ffc9cfe5630 a2=0 a3=7ffc9cfe561c items=0 ppid=3117 pid=3275 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:53.295000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 Jan 14 13:51:53.310000 audit[3276]: NETFILTER_CFG table=filter:100 family=10 entries=1 op=nft_register_chain pid=3276 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:53.310000 audit[3276]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffdebeedd30 a2=0 a3=7ffdebeedd1c items=0 ppid=3117 pid=3276 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:53.310000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4E004B5542452D4649524557414C4C002D740066696C746572 Jan 14 13:51:53.344000 audit[3278]: NETFILTER_CFG table=filter:101 family=10 entries=1 op=nft_register_rule pid=3278 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:53.344000 audit[3278]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7ffcca810690 a2=0 a3=7ffcca81067c items=0 ppid=3117 pid=3278 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:53.344000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 14 13:51:53.393000 audit[3281]: NETFILTER_CFG table=filter:102 family=10 entries=1 op=nft_register_rule pid=3281 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 13:51:53.393000 audit[3281]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7fff9a3c8530 a2=0 a3=7fff9a3c851c items=0 ppid=3117 pid=3281 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:53.393000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 14 13:51:53.446000 audit[3283]: NETFILTER_CFG table=filter:103 family=10 entries=3 op=nft_register_rule pid=3283 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" Jan 14 13:51:53.446000 audit[3283]: SYSCALL arch=c000003e syscall=46 success=yes exit=2088 a0=3 a1=7ffda5a1f020 a2=0 a3=7ffda5a1f00c items=0 ppid=3117 pid=3283 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:53.446000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:51:53.448000 audit[3283]: NETFILTER_CFG table=nat:104 family=10 entries=7 op=nft_register_chain pid=3283 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" Jan 14 13:51:53.448000 audit[3283]: SYSCALL arch=c000003e syscall=46 success=yes exit=2056 a0=3 a1=7ffda5a1f020 a2=0 a3=7ffda5a1f00c items=0 ppid=3117 pid=3283 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:53.448000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:51:53.983346 kubelet[2951]: E0114 13:51:53.982182 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:54.034475 kubelet[2951]: I0114 13:51:54.031890 2951 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-9csc2" podStartSLOduration=19.031867281 podStartE2EDuration="19.031867281s" podCreationTimestamp="2026-01-14 13:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-14 13:51:50.976720422 +0000 UTC m=+18.224945690" watchObservedRunningTime="2026-01-14 13:51:54.031867281 +0000 UTC m=+21.280092550" Jan 14 13:51:55.744392 kubelet[2951]: E0114 13:51:55.743202 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:51:57.526784 containerd[1635]: time="2026-01-14T13:51:57.526291519Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:51:57.529468 containerd[1635]: time="2026-01-14T13:51:57.529443660Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.7: active requests=0, bytes read=23558205" Jan 14 13:51:57.538084 containerd[1635]: time="2026-01-14T13:51:57.536286990Z" level=info msg="ImageCreate event name:\"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:51:57.544873 containerd[1635]: time="2026-01-14T13:51:57.544835307Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:51:57.547252 containerd[1635]: time="2026-01-14T13:51:57.547220538Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.7\" with image id \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\", repo tag \"quay.io/tigera/operator:v1.38.7\", repo digest \"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\", size \"25057686\" in 7.553858336s" Jan 14 13:51:57.550356 containerd[1635]: time="2026-01-14T13:51:57.547335932Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\" returns image reference \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\"" Jan 14 13:51:57.562816 containerd[1635]: time="2026-01-14T13:51:57.561884125Z" level=info msg="CreateContainer within sandbox \"76ca8881a31453feb9f0da5ee45296536a0d894afcc0c120f225f6849c64e3a7\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Jan 14 13:51:57.607337 containerd[1635]: time="2026-01-14T13:51:57.607143242Z" level=info msg="Container 87587a98dc9d999dbf4d9e84299fdd7de12c3e5605fc5d85d275e64bf733c6f0: CDI devices from CRI Config.CDIDevices: []" Jan 14 13:51:57.634764 containerd[1635]: time="2026-01-14T13:51:57.634457003Z" level=info msg="CreateContainer within sandbox \"76ca8881a31453feb9f0da5ee45296536a0d894afcc0c120f225f6849c64e3a7\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"87587a98dc9d999dbf4d9e84299fdd7de12c3e5605fc5d85d275e64bf733c6f0\"" Jan 14 13:51:57.637794 containerd[1635]: time="2026-01-14T13:51:57.637217474Z" level=info msg="StartContainer for \"87587a98dc9d999dbf4d9e84299fdd7de12c3e5605fc5d85d275e64bf733c6f0\"" Jan 14 13:51:57.640129 containerd[1635]: time="2026-01-14T13:51:57.640092991Z" level=info msg="connecting to shim 87587a98dc9d999dbf4d9e84299fdd7de12c3e5605fc5d85d275e64bf733c6f0" address="unix:///run/containerd/s/476691f8fcf272a288e87caea25effbde5e0949e0c1a002074fb6497c7b85eda" protocol=ttrpc version=3 Jan 14 13:51:57.750143 systemd[1]: Started cri-containerd-87587a98dc9d999dbf4d9e84299fdd7de12c3e5605fc5d85d275e64bf733c6f0.scope - libcontainer container 87587a98dc9d999dbf4d9e84299fdd7de12c3e5605fc5d85d275e64bf733c6f0. Jan 14 13:51:57.823000 audit: BPF prog-id=144 op=LOAD Jan 14 13:51:57.838053 kernel: kauditd_printk_skb: 80 callbacks suppressed Jan 14 13:51:57.838320 kernel: audit: type=1334 audit(1768398717.823:506): prog-id=144 op=LOAD Jan 14 13:51:57.827000 audit: BPF prog-id=145 op=LOAD Jan 14 13:51:57.870825 kernel: audit: type=1334 audit(1768398717.827:507): prog-id=145 op=LOAD Jan 14 13:51:57.871093 kernel: audit: type=1300 audit(1768398717.827:507): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=3064 pid=3284 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:57.827000 audit[3284]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=3064 pid=3284 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:57.827000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837353837613938646339643939396462663464396538343239396664 Jan 14 13:51:57.998825 kernel: audit: type=1327 audit(1768398717.827:507): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837353837613938646339643939396462663464396538343239396664 Jan 14 13:51:57.827000 audit: BPF prog-id=145 op=UNLOAD Jan 14 13:51:58.014782 kernel: audit: type=1334 audit(1768398717.827:508): prog-id=145 op=UNLOAD Jan 14 13:51:57.827000 audit[3284]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3064 pid=3284 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:58.070105 kernel: audit: type=1300 audit(1768398717.827:508): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3064 pid=3284 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:58.070407 kernel: audit: type=1327 audit(1768398717.827:508): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837353837613938646339643939396462663464396538343239396664 Jan 14 13:51:57.827000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837353837613938646339643939396462663464396538343239396664 Jan 14 13:51:57.828000 audit: BPF prog-id=146 op=LOAD Jan 14 13:51:58.141466 kernel: audit: type=1334 audit(1768398717.828:509): prog-id=146 op=LOAD Jan 14 13:51:58.142110 kernel: audit: type=1300 audit(1768398717.828:509): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=3064 pid=3284 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:57.828000 audit[3284]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=3064 pid=3284 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:58.154251 containerd[1635]: time="2026-01-14T13:51:58.154183458Z" level=info msg="StartContainer for \"87587a98dc9d999dbf4d9e84299fdd7de12c3e5605fc5d85d275e64bf733c6f0\" returns successfully" Jan 14 13:51:57.828000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837353837613938646339643939396462663464396538343239396664 Jan 14 13:51:58.256323 kernel: audit: type=1327 audit(1768398717.828:509): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837353837613938646339643939396462663464396538343239396664 Jan 14 13:51:57.828000 audit: BPF prog-id=147 op=LOAD Jan 14 13:51:57.828000 audit[3284]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=3064 pid=3284 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:57.828000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837353837613938646339643939396462663464396538343239396664 Jan 14 13:51:57.828000 audit: BPF prog-id=147 op=UNLOAD Jan 14 13:51:57.828000 audit[3284]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3064 pid=3284 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:57.828000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837353837613938646339643939396462663464396538343239396664 Jan 14 13:51:57.828000 audit: BPF prog-id=146 op=UNLOAD Jan 14 13:51:57.828000 audit[3284]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3064 pid=3284 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:57.828000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837353837613938646339643939396462663464396538343239396664 Jan 14 13:51:57.828000 audit: BPF prog-id=148 op=LOAD Jan 14 13:51:57.828000 audit[3284]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=3064 pid=3284 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:51:57.828000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3837353837613938646339643939396462663464396538343239396664 Jan 14 13:52:04.010154 kubelet[2951]: E0114 13:52:04.009419 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:04.266395 kubelet[2951]: I0114 13:52:04.264494 2951 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-65cdcdfd6d-86fkx" podStartSLOduration=10.707866646 podStartE2EDuration="18.26447681s" podCreationTimestamp="2026-01-14 13:51:46 +0000 UTC" firstStartedPulling="2026-01-14 13:51:49.992493047 +0000 UTC m=+17.240718315" lastFinishedPulling="2026-01-14 13:51:57.549103211 +0000 UTC m=+24.797328479" observedRunningTime="2026-01-14 13:51:59.093234242 +0000 UTC m=+26.341459530" watchObservedRunningTime="2026-01-14 13:52:04.26447681 +0000 UTC m=+31.512702078" Jan 14 13:52:13.078769 sudo[1876]: pam_unix(sudo:session): session closed for user root Jan 14 13:52:13.150797 kernel: kauditd_printk_skb: 12 callbacks suppressed Jan 14 13:52:13.150924 kernel: audit: type=1106 audit(1768398733.083:514): pid=1876 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 13:52:13.083000 audit[1876]: USER_END pid=1876 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 13:52:13.151233 sshd[1875]: Connection closed by 10.0.0.1 port 49234 Jan 14 13:52:13.092000 audit[1876]: CRED_DISP pid=1876 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 13:52:13.126318 sshd-session[1871]: pam_unix(sshd:session): session closed for user core Jan 14 13:52:13.177134 systemd-logind[1617]: Session 8 logged out. Waiting for processes to exit. Jan 14 13:52:13.180312 systemd[1]: sshd@6-10.0.0.127:22-10.0.0.1:49234.service: Deactivated successfully. Jan 14 13:52:13.198865 systemd[1]: session-8.scope: Deactivated successfully. Jan 14 13:52:13.206280 kernel: audit: type=1104 audit(1768398733.092:515): pid=1876 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 13:52:13.206342 kernel: audit: type=1106 audit(1768398733.166:516): pid=1871 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:52:13.166000 audit[1871]: USER_END pid=1871 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:52:13.199879 systemd[1]: session-8.scope: Consumed 27.041s CPU time, 225.9M memory peak. Jan 14 13:52:13.212124 systemd-logind[1617]: Removed session 8. Jan 14 13:52:13.281313 kernel: audit: type=1104 audit(1768398733.166:517): pid=1871 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:52:13.166000 audit[1871]: CRED_DISP pid=1871 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:52:13.181000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-10.0.0.127:22-10.0.0.1:49234 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:52:13.377100 kernel: audit: type=1131 audit(1768398733.181:518): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-10.0.0.127:22-10.0.0.1:49234 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:52:17.478000 audit[3378]: NETFILTER_CFG table=filter:105 family=2 entries=15 op=nft_register_rule pid=3378 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:17.518120 kernel: audit: type=1325 audit(1768398737.478:519): table=filter:105 family=2 entries=15 op=nft_register_rule pid=3378 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:17.518279 kernel: audit: type=1300 audit(1768398737.478:519): arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7ffd5e019a10 a2=0 a3=7ffd5e0199fc items=0 ppid=3117 pid=3378 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:17.478000 audit[3378]: SYSCALL arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7ffd5e019a10 a2=0 a3=7ffd5e0199fc items=0 ppid=3117 pid=3378 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:17.619936 kernel: audit: type=1327 audit(1768398737.478:519): proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:17.478000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:17.627000 audit[3378]: NETFILTER_CFG table=nat:106 family=2 entries=12 op=nft_register_rule pid=3378 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:17.627000 audit[3378]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd5e019a10 a2=0 a3=0 items=0 ppid=3117 pid=3378 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:17.740121 kernel: audit: type=1325 audit(1768398737.627:520): table=nat:106 family=2 entries=12 op=nft_register_rule pid=3378 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:17.740232 kernel: audit: type=1300 audit(1768398737.627:520): arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd5e019a10 a2=0 a3=0 items=0 ppid=3117 pid=3378 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:17.627000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:19.223000 audit[3381]: NETFILTER_CFG table=filter:107 family=2 entries=18 op=nft_register_rule pid=3381 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:19.243146 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 14 13:52:19.243300 kernel: audit: type=1325 audit(1768398739.223:521): table=filter:107 family=2 entries=18 op=nft_register_rule pid=3381 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:19.223000 audit[3381]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffce0f693a0 a2=0 a3=7ffce0f6938c items=0 ppid=3117 pid=3381 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:19.351922 kernel: audit: type=1300 audit(1768398739.223:521): arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffce0f693a0 a2=0 a3=7ffce0f6938c items=0 ppid=3117 pid=3381 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:19.223000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:19.384858 kernel: audit: type=1327 audit(1768398739.223:521): proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:19.351000 audit[3381]: NETFILTER_CFG table=nat:108 family=2 entries=12 op=nft_register_rule pid=3381 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:19.422861 kernel: audit: type=1325 audit(1768398739.351:522): table=nat:108 family=2 entries=12 op=nft_register_rule pid=3381 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:19.351000 audit[3381]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffce0f693a0 a2=0 a3=0 items=0 ppid=3117 pid=3381 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:19.497287 kernel: audit: type=1300 audit(1768398739.351:522): arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffce0f693a0 a2=0 a3=0 items=0 ppid=3117 pid=3381 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:19.351000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:19.529429 kernel: audit: type=1327 audit(1768398739.351:522): proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:26.266000 audit[3385]: NETFILTER_CFG table=filter:109 family=2 entries=21 op=nft_register_rule pid=3385 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:26.309425 kernel: audit: type=1325 audit(1768398746.266:523): table=filter:109 family=2 entries=21 op=nft_register_rule pid=3385 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:26.395226 kernel: audit: type=1300 audit(1768398746.266:523): arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7fff1457aa30 a2=0 a3=7fff1457aa1c items=0 ppid=3117 pid=3385 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:26.266000 audit[3385]: SYSCALL arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7fff1457aa30 a2=0 a3=7fff1457aa1c items=0 ppid=3117 pid=3385 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:26.396868 kubelet[2951]: I0114 13:52:26.361154 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c72902a2-6254-4e22-9e4c-44098b40c0f6-tigera-ca-bundle\") pod \"calico-typha-5677687b8b-zn66h\" (UID: \"c72902a2-6254-4e22-9e4c-44098b40c0f6\") " pod="calico-system/calico-typha-5677687b8b-zn66h" Jan 14 13:52:26.396868 kubelet[2951]: I0114 13:52:26.361194 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/c72902a2-6254-4e22-9e4c-44098b40c0f6-typha-certs\") pod \"calico-typha-5677687b8b-zn66h\" (UID: \"c72902a2-6254-4e22-9e4c-44098b40c0f6\") " pod="calico-system/calico-typha-5677687b8b-zn66h" Jan 14 13:52:26.396868 kubelet[2951]: I0114 13:52:26.361212 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbq8l\" (UniqueName: \"kubernetes.io/projected/c72902a2-6254-4e22-9e4c-44098b40c0f6-kube-api-access-lbq8l\") pod \"calico-typha-5677687b8b-zn66h\" (UID: \"c72902a2-6254-4e22-9e4c-44098b40c0f6\") " pod="calico-system/calico-typha-5677687b8b-zn66h" Jan 14 13:52:26.332351 systemd[1]: Created slice kubepods-besteffort-podc72902a2_6254_4e22_9e4c_44098b40c0f6.slice - libcontainer container kubepods-besteffort-podc72902a2_6254_4e22_9e4c_44098b40c0f6.slice. Jan 14 13:52:26.266000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:26.405000 audit[3385]: NETFILTER_CFG table=nat:110 family=2 entries=12 op=nft_register_rule pid=3385 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:26.483457 kernel: audit: type=1327 audit(1768398746.266:523): proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:26.483942 kernel: audit: type=1325 audit(1768398746.405:524): table=nat:110 family=2 entries=12 op=nft_register_rule pid=3385 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:26.405000 audit[3385]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7fff1457aa30 a2=0 a3=0 items=0 ppid=3117 pid=3385 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:26.564232 kernel: audit: type=1300 audit(1768398746.405:524): arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7fff1457aa30 a2=0 a3=0 items=0 ppid=3117 pid=3385 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:26.405000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:26.626236 kernel: audit: type=1327 audit(1768398746.405:524): proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:26.678853 kubelet[2951]: E0114 13:52:26.678161 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:26.689866 containerd[1635]: time="2026-01-14T13:52:26.688241783Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5677687b8b-zn66h,Uid:c72902a2-6254-4e22-9e4c-44098b40c0f6,Namespace:calico-system,Attempt:0,}" Jan 14 13:52:26.796447 systemd[1]: Created slice kubepods-besteffort-pode067985e_0631_4d20_b684_e59dafe9a54d.slice - libcontainer container kubepods-besteffort-pode067985e_0631_4d20_b684_e59dafe9a54d.slice. Jan 14 13:52:26.872868 containerd[1635]: time="2026-01-14T13:52:26.872427075Z" level=info msg="connecting to shim 33111a5b67beb1963ac52a343dac01563ee0c2e3fdcc951af612a9f4c0163ece" address="unix:///run/containerd/s/5d72708a39df1695ee7f76e1e0bd0d10bf4413a2936634aa81de4efbd141ed2e" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:52:26.874450 kubelet[2951]: I0114 13:52:26.873910 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/e067985e-0631-4d20-b684-e59dafe9a54d-policysync\") pod \"calico-node-mt5qd\" (UID: \"e067985e-0631-4d20-b684-e59dafe9a54d\") " pod="calico-system/calico-node-mt5qd" Jan 14 13:52:26.874450 kubelet[2951]: I0114 13:52:26.873942 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e067985e-0631-4d20-b684-e59dafe9a54d-xtables-lock\") pod \"calico-node-mt5qd\" (UID: \"e067985e-0631-4d20-b684-e59dafe9a54d\") " pod="calico-system/calico-node-mt5qd" Jan 14 13:52:26.874450 kubelet[2951]: I0114 13:52:26.873958 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/e067985e-0631-4d20-b684-e59dafe9a54d-cni-log-dir\") pod \"calico-node-mt5qd\" (UID: \"e067985e-0631-4d20-b684-e59dafe9a54d\") " pod="calico-system/calico-node-mt5qd" Jan 14 13:52:26.874450 kubelet[2951]: I0114 13:52:26.873972 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/e067985e-0631-4d20-b684-e59dafe9a54d-flexvol-driver-host\") pod \"calico-node-mt5qd\" (UID: \"e067985e-0631-4d20-b684-e59dafe9a54d\") " pod="calico-system/calico-node-mt5qd" Jan 14 13:52:26.874450 kubelet[2951]: I0114 13:52:26.874165 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/e067985e-0631-4d20-b684-e59dafe9a54d-cni-bin-dir\") pod \"calico-node-mt5qd\" (UID: \"e067985e-0631-4d20-b684-e59dafe9a54d\") " pod="calico-system/calico-node-mt5qd" Jan 14 13:52:26.874970 kubelet[2951]: I0114 13:52:26.874190 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e067985e-0631-4d20-b684-e59dafe9a54d-lib-modules\") pod \"calico-node-mt5qd\" (UID: \"e067985e-0631-4d20-b684-e59dafe9a54d\") " pod="calico-system/calico-node-mt5qd" Jan 14 13:52:26.874970 kubelet[2951]: I0114 13:52:26.874204 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e067985e-0631-4d20-b684-e59dafe9a54d-tigera-ca-bundle\") pod \"calico-node-mt5qd\" (UID: \"e067985e-0631-4d20-b684-e59dafe9a54d\") " pod="calico-system/calico-node-mt5qd" Jan 14 13:52:26.874970 kubelet[2951]: I0114 13:52:26.874223 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/e067985e-0631-4d20-b684-e59dafe9a54d-var-lib-calico\") pod \"calico-node-mt5qd\" (UID: \"e067985e-0631-4d20-b684-e59dafe9a54d\") " pod="calico-system/calico-node-mt5qd" Jan 14 13:52:26.874970 kubelet[2951]: I0114 13:52:26.874236 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktsw5\" (UniqueName: \"kubernetes.io/projected/e067985e-0631-4d20-b684-e59dafe9a54d-kube-api-access-ktsw5\") pod \"calico-node-mt5qd\" (UID: \"e067985e-0631-4d20-b684-e59dafe9a54d\") " pod="calico-system/calico-node-mt5qd" Jan 14 13:52:26.874970 kubelet[2951]: I0114 13:52:26.874251 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/e067985e-0631-4d20-b684-e59dafe9a54d-cni-net-dir\") pod \"calico-node-mt5qd\" (UID: \"e067985e-0631-4d20-b684-e59dafe9a54d\") " pod="calico-system/calico-node-mt5qd" Jan 14 13:52:26.875262 kubelet[2951]: I0114 13:52:26.874267 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/e067985e-0631-4d20-b684-e59dafe9a54d-node-certs\") pod \"calico-node-mt5qd\" (UID: \"e067985e-0631-4d20-b684-e59dafe9a54d\") " pod="calico-system/calico-node-mt5qd" Jan 14 13:52:26.875262 kubelet[2951]: I0114 13:52:26.874279 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/e067985e-0631-4d20-b684-e59dafe9a54d-var-run-calico\") pod \"calico-node-mt5qd\" (UID: \"e067985e-0631-4d20-b684-e59dafe9a54d\") " pod="calico-system/calico-node-mt5qd" Jan 14 13:52:26.982832 kubelet[2951]: E0114 13:52:26.981477 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.982832 kubelet[2951]: W0114 13:52:26.982494 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.983928 kubelet[2951]: E0114 13:52:26.983380 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.984346 kubelet[2951]: E0114 13:52:26.984249 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.984346 kubelet[2951]: W0114 13:52:26.984261 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.984346 kubelet[2951]: E0114 13:52:26.984275 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.985178 kubelet[2951]: E0114 13:52:26.984467 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.985178 kubelet[2951]: W0114 13:52:26.984484 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.985178 kubelet[2951]: E0114 13:52:26.984493 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.987239 kubelet[2951]: E0114 13:52:26.986236 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.987239 kubelet[2951]: W0114 13:52:26.986249 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.987239 kubelet[2951]: E0114 13:52:26.986259 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.987239 kubelet[2951]: E0114 13:52:26.986430 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.987239 kubelet[2951]: W0114 13:52:26.986441 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.988861 kubelet[2951]: E0114 13:52:26.987441 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.988861 kubelet[2951]: E0114 13:52:26.987929 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.988861 kubelet[2951]: W0114 13:52:26.987938 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.989313 kubelet[2951]: E0114 13:52:26.988961 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.989313 kubelet[2951]: E0114 13:52:26.989295 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.989313 kubelet[2951]: W0114 13:52:26.989303 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.989313 kubelet[2951]: E0114 13:52:26.989312 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.991355 kubelet[2951]: E0114 13:52:26.989475 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.991355 kubelet[2951]: W0114 13:52:26.990406 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.991355 kubelet[2951]: E0114 13:52:26.990417 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.991475 kubelet[2951]: E0114 13:52:26.991428 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.991475 kubelet[2951]: W0114 13:52:26.991436 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.991475 kubelet[2951]: E0114 13:52:26.991445 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.991947 kubelet[2951]: E0114 13:52:26.991913 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.991947 kubelet[2951]: W0114 13:52:26.991922 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.991947 kubelet[2951]: E0114 13:52:26.991930 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.992289 kubelet[2951]: E0114 13:52:26.992253 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.992289 kubelet[2951]: W0114 13:52:26.992261 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.992289 kubelet[2951]: E0114 13:52:26.992270 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.993868 kubelet[2951]: E0114 13:52:26.992424 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.993868 kubelet[2951]: W0114 13:52:26.992435 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.993868 kubelet[2951]: E0114 13:52:26.992444 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.993868 kubelet[2951]: E0114 13:52:26.992899 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.993868 kubelet[2951]: W0114 13:52:26.992907 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.993868 kubelet[2951]: E0114 13:52:26.992915 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.993868 kubelet[2951]: E0114 13:52:26.993259 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.993868 kubelet[2951]: W0114 13:52:26.993271 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.993868 kubelet[2951]: E0114 13:52:26.993284 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.993868 kubelet[2951]: E0114 13:52:26.993877 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.994363 kubelet[2951]: W0114 13:52:26.993890 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.994363 kubelet[2951]: E0114 13:52:26.993906 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.994363 kubelet[2951]: E0114 13:52:26.994307 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.994363 kubelet[2951]: W0114 13:52:26.994320 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.994363 kubelet[2951]: E0114 13:52:26.994331 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.995361 kubelet[2951]: E0114 13:52:26.995168 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:26.995361 kubelet[2951]: W0114 13:52:26.995347 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:26.995361 kubelet[2951]: E0114 13:52:26.995363 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:26.999951 kubelet[2951]: E0114 13:52:26.999925 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.000219 kubelet[2951]: W0114 13:52:27.000202 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.000282 kubelet[2951]: E0114 13:52:27.000270 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.010941 kubelet[2951]: E0114 13:52:27.010917 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.011242 kubelet[2951]: W0114 13:52:27.011223 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.011322 kubelet[2951]: E0114 13:52:27.011309 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.026173 kubelet[2951]: E0114 13:52:27.025964 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.026460 kubelet[2951]: W0114 13:52:27.026296 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.026460 kubelet[2951]: E0114 13:52:27.026326 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.032255 kubelet[2951]: E0114 13:52:27.031901 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.032255 kubelet[2951]: W0114 13:52:27.031919 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.032255 kubelet[2951]: E0114 13:52:27.031938 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.034850 kubelet[2951]: E0114 13:52:27.033436 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:27.037856 kubelet[2951]: E0114 13:52:27.036916 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.037856 kubelet[2951]: W0114 13:52:27.036933 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.037856 kubelet[2951]: E0114 13:52:27.036950 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.038390 kubelet[2951]: E0114 13:52:27.038203 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.038390 kubelet[2951]: W0114 13:52:27.038366 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.038479 kubelet[2951]: E0114 13:52:27.038396 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.052947 kubelet[2951]: E0114 13:52:27.050819 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.052947 kubelet[2951]: W0114 13:52:27.051131 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.052947 kubelet[2951]: E0114 13:52:27.051150 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.070726 kubelet[2951]: E0114 13:52:27.069906 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.070726 kubelet[2951]: W0114 13:52:27.070217 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.070726 kubelet[2951]: E0114 13:52:27.070235 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.079415 kubelet[2951]: E0114 13:52:27.078934 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.079415 kubelet[2951]: W0114 13:52:27.078953 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.079415 kubelet[2951]: E0114 13:52:27.078968 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.085342 kubelet[2951]: E0114 13:52:27.085159 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.085342 kubelet[2951]: W0114 13:52:27.085330 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.087284 kubelet[2951]: E0114 13:52:27.085349 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.087284 kubelet[2951]: E0114 13:52:27.086322 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.087284 kubelet[2951]: W0114 13:52:27.086331 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.087284 kubelet[2951]: E0114 13:52:27.086341 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.088332 kubelet[2951]: E0114 13:52:27.088172 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.088332 kubelet[2951]: W0114 13:52:27.088185 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.088332 kubelet[2951]: E0114 13:52:27.088195 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.090411 kubelet[2951]: E0114 13:52:27.089818 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.090411 kubelet[2951]: W0114 13:52:27.089831 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.090411 kubelet[2951]: E0114 13:52:27.089841 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.094977 kubelet[2951]: E0114 13:52:27.092346 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.094977 kubelet[2951]: W0114 13:52:27.092360 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.094977 kubelet[2951]: E0114 13:52:27.092369 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.097459 kubelet[2951]: E0114 13:52:27.095337 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.097459 kubelet[2951]: W0114 13:52:27.095366 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.097459 kubelet[2951]: E0114 13:52:27.095400 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.101222 kubelet[2951]: E0114 13:52:27.101198 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.101319 kubelet[2951]: W0114 13:52:27.101299 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.101400 kubelet[2951]: E0114 13:52:27.101381 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.104341 kubelet[2951]: E0114 13:52:27.104323 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.104422 kubelet[2951]: W0114 13:52:27.104406 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.104902 kubelet[2951]: E0114 13:52:27.104510 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.126288 kubelet[2951]: E0114 13:52:27.124240 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.128901 kubelet[2951]: W0114 13:52:27.128506 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.130227 kubelet[2951]: E0114 13:52:27.130202 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.148372 kubelet[2951]: E0114 13:52:27.141898 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.148372 kubelet[2951]: W0114 13:52:27.141920 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.148372 kubelet[2951]: E0114 13:52:27.141943 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.162380 kubelet[2951]: E0114 13:52:27.160879 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.162380 kubelet[2951]: W0114 13:52:27.161167 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.162380 kubelet[2951]: E0114 13:52:27.161190 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.173448 kubelet[2951]: E0114 13:52:27.171501 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.172138 systemd[1]: Started cri-containerd-33111a5b67beb1963ac52a343dac01563ee0c2e3fdcc951af612a9f4c0163ece.scope - libcontainer container 33111a5b67beb1963ac52a343dac01563ee0c2e3fdcc951af612a9f4c0163ece. Jan 14 13:52:27.182177 kubelet[2951]: W0114 13:52:27.182157 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.183841 kubelet[2951]: E0114 13:52:27.182242 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.190805 kubelet[2951]: E0114 13:52:27.189217 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.190918 kubelet[2951]: W0114 13:52:27.190900 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.191147 kubelet[2951]: E0114 13:52:27.190971 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.198249 kubelet[2951]: E0114 13:52:27.193420 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.198249 kubelet[2951]: W0114 13:52:27.193435 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.198249 kubelet[2951]: E0114 13:52:27.193446 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.202814 kubelet[2951]: E0114 13:52:27.202794 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.202937 kubelet[2951]: W0114 13:52:27.202896 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.202937 kubelet[2951]: E0114 13:52:27.202916 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.228184 kubelet[2951]: E0114 13:52:27.228149 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.228383 kubelet[2951]: W0114 13:52:27.228330 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.228383 kubelet[2951]: E0114 13:52:27.228361 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.232151 kubelet[2951]: I0114 13:52:27.231966 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/967b7513-2769-43e5-a113-9344f17909c3-varrun\") pod \"csi-node-driver-pvxlq\" (UID: \"967b7513-2769-43e5-a113-9344f17909c3\") " pod="calico-system/csi-node-driver-pvxlq" Jan 14 13:52:27.234294 kubelet[2951]: E0114 13:52:27.234192 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.234294 kubelet[2951]: W0114 13:52:27.234205 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.234294 kubelet[2951]: E0114 13:52:27.234218 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.237857 kubelet[2951]: E0114 13:52:27.237840 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.239380 kubelet[2951]: W0114 13:52:27.237920 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.239380 kubelet[2951]: E0114 13:52:27.237940 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.240833 kubelet[2951]: E0114 13:52:27.240819 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.240896 kubelet[2951]: W0114 13:52:27.240885 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.240956 kubelet[2951]: E0114 13:52:27.240945 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.241867 kubelet[2951]: I0114 13:52:27.241844 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/967b7513-2769-43e5-a113-9344f17909c3-kubelet-dir\") pod \"csi-node-driver-pvxlq\" (UID: \"967b7513-2769-43e5-a113-9344f17909c3\") " pod="calico-system/csi-node-driver-pvxlq" Jan 14 13:52:27.244836 kubelet[2951]: E0114 13:52:27.243929 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.244836 kubelet[2951]: W0114 13:52:27.243953 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.244836 kubelet[2951]: E0114 13:52:27.243975 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.251438 kubelet[2951]: E0114 13:52:27.250432 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.253331 kubelet[2951]: W0114 13:52:27.252504 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.253331 kubelet[2951]: E0114 13:52:27.252936 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.259365 kubelet[2951]: E0114 13:52:27.258886 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.259365 kubelet[2951]: W0114 13:52:27.259197 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.259365 kubelet[2951]: E0114 13:52:27.259238 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.270982 kubelet[2951]: E0114 13:52:27.269504 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.270982 kubelet[2951]: W0114 13:52:27.269823 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.270982 kubelet[2951]: E0114 13:52:27.269839 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.272356 kubelet[2951]: I0114 13:52:27.272185 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/967b7513-2769-43e5-a113-9344f17909c3-registration-dir\") pod \"csi-node-driver-pvxlq\" (UID: \"967b7513-2769-43e5-a113-9344f17909c3\") " pod="calico-system/csi-node-driver-pvxlq" Jan 14 13:52:27.275812 kubelet[2951]: E0114 13:52:27.275480 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.275812 kubelet[2951]: W0114 13:52:27.275498 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.276491 kubelet[2951]: E0114 13:52:27.275511 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.291803 kubelet[2951]: E0114 13:52:27.288796 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.291803 kubelet[2951]: W0114 13:52:27.290788 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.291803 kubelet[2951]: E0114 13:52:27.290825 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.295925 kubelet[2951]: E0114 13:52:27.295907 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.296187 kubelet[2951]: W0114 13:52:27.296170 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.297969 kubelet[2951]: E0114 13:52:27.297953 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.299891 kubelet[2951]: E0114 13:52:27.299876 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.299966 kubelet[2951]: W0114 13:52:27.299951 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.301728 kubelet[2951]: E0114 13:52:27.300175 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.308881 kubelet[2951]: E0114 13:52:27.308859 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.308980 kubelet[2951]: W0114 13:52:27.308965 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.310826 kubelet[2951]: E0114 13:52:27.310806 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.321350 kubelet[2951]: E0114 13:52:27.320900 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.321421 kubelet[2951]: W0114 13:52:27.321371 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.321421 kubelet[2951]: E0114 13:52:27.321404 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.324796 kubelet[2951]: E0114 13:52:27.324431 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.324796 kubelet[2951]: W0114 13:52:27.324448 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.324796 kubelet[2951]: E0114 13:52:27.324465 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.330251 kubelet[2951]: I0114 13:52:27.324499 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/967b7513-2769-43e5-a113-9344f17909c3-socket-dir\") pod \"csi-node-driver-pvxlq\" (UID: \"967b7513-2769-43e5-a113-9344f17909c3\") " pod="calico-system/csi-node-driver-pvxlq" Jan 14 13:52:27.330251 kubelet[2951]: E0114 13:52:27.330196 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.330251 kubelet[2951]: W0114 13:52:27.330207 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.330251 kubelet[2951]: E0114 13:52:27.330220 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.330842 kubelet[2951]: E0114 13:52:27.330505 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.332152 kubelet[2951]: W0114 13:52:27.331819 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.332152 kubelet[2951]: E0114 13:52:27.331843 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.334800 kubelet[2951]: E0114 13:52:27.333897 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.334800 kubelet[2951]: W0114 13:52:27.333913 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.334800 kubelet[2951]: E0114 13:52:27.333924 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.372000 audit: BPF prog-id=149 op=LOAD Jan 14 13:52:27.391942 kernel: audit: type=1334 audit(1768398747.372:525): prog-id=149 op=LOAD Jan 14 13:52:27.376000 audit: BPF prog-id=150 op=LOAD Jan 14 13:52:27.418941 kernel: audit: type=1334 audit(1768398747.376:526): prog-id=150 op=LOAD Jan 14 13:52:27.376000 audit[3408]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000174238 a2=98 a3=0 items=0 ppid=3395 pid=3408 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:27.481259 kernel: audit: type=1300 audit(1768398747.376:526): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000174238 a2=98 a3=0 items=0 ppid=3395 pid=3408 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:27.481297 kubelet[2951]: E0114 13:52:27.436831 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.481297 kubelet[2951]: W0114 13:52:27.436847 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.481297 kubelet[2951]: E0114 13:52:27.436864 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.481297 kubelet[2951]: E0114 13:52:27.448822 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.481297 kubelet[2951]: W0114 13:52:27.448838 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.481297 kubelet[2951]: E0114 13:52:27.448856 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.481297 kubelet[2951]: E0114 13:52:27.450150 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.481297 kubelet[2951]: W0114 13:52:27.450165 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.481297 kubelet[2951]: E0114 13:52:27.450178 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.481297 kubelet[2951]: E0114 13:52:27.450372 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.482405 kubelet[2951]: W0114 13:52:27.450379 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.482405 kubelet[2951]: E0114 13:52:27.450389 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.482405 kubelet[2951]: E0114 13:52:27.451448 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.482405 kubelet[2951]: W0114 13:52:27.451457 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.482405 kubelet[2951]: E0114 13:52:27.451467 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.482405 kubelet[2951]: E0114 13:52:27.452790 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.482405 kubelet[2951]: W0114 13:52:27.452798 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.482405 kubelet[2951]: E0114 13:52:27.452808 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.482405 kubelet[2951]: E0114 13:52:27.454836 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.482405 kubelet[2951]: W0114 13:52:27.454845 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.482902 kubelet[2951]: E0114 13:52:27.454855 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.482902 kubelet[2951]: E0114 13:52:27.459753 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.482902 kubelet[2951]: W0114 13:52:27.459762 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.482902 kubelet[2951]: E0114 13:52:27.459772 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.482902 kubelet[2951]: E0114 13:52:27.461164 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.482902 kubelet[2951]: W0114 13:52:27.461188 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.482902 kubelet[2951]: E0114 13:52:27.461213 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.482902 kubelet[2951]: E0114 13:52:27.463850 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.482902 kubelet[2951]: W0114 13:52:27.463870 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.482902 kubelet[2951]: E0114 13:52:27.463892 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.483250 kubelet[2951]: I0114 13:52:27.464161 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq7hq\" (UniqueName: \"kubernetes.io/projected/967b7513-2769-43e5-a113-9344f17909c3-kube-api-access-kq7hq\") pod \"csi-node-driver-pvxlq\" (UID: \"967b7513-2769-43e5-a113-9344f17909c3\") " pod="calico-system/csi-node-driver-pvxlq" Jan 14 13:52:27.483250 kubelet[2951]: E0114 13:52:27.468139 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.483250 kubelet[2951]: W0114 13:52:27.468152 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.483250 kubelet[2951]: E0114 13:52:27.468163 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.483250 kubelet[2951]: E0114 13:52:27.468406 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.483250 kubelet[2951]: W0114 13:52:27.468414 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.483250 kubelet[2951]: E0114 13:52:27.468422 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.483250 kubelet[2951]: E0114 13:52:27.468912 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.483250 kubelet[2951]: W0114 13:52:27.468921 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.483421 kubelet[2951]: E0114 13:52:27.468930 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.483421 kubelet[2951]: E0114 13:52:27.470828 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.483421 kubelet[2951]: W0114 13:52:27.470837 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.483421 kubelet[2951]: E0114 13:52:27.470845 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.483421 kubelet[2951]: E0114 13:52:27.473916 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.483421 kubelet[2951]: W0114 13:52:27.473927 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.483421 kubelet[2951]: E0114 13:52:27.473938 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.483421 kubelet[2951]: E0114 13:52:27.478869 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.483421 kubelet[2951]: W0114 13:52:27.478880 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.483421 kubelet[2951]: E0114 13:52:27.478890 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.376000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333313131613562363762656231393633616335326133343364616330 Jan 14 13:52:27.550337 containerd[1635]: time="2026-01-14T13:52:27.502956146Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-mt5qd,Uid:e067985e-0631-4d20-b684-e59dafe9a54d,Namespace:calico-system,Attempt:0,}" Jan 14 13:52:27.550437 kubelet[2951]: E0114 13:52:27.484404 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.550437 kubelet[2951]: W0114 13:52:27.484416 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.550437 kubelet[2951]: E0114 13:52:27.484427 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.550437 kubelet[2951]: E0114 13:52:27.485906 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.550437 kubelet[2951]: W0114 13:52:27.485916 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.550437 kubelet[2951]: E0114 13:52:27.485926 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.550437 kubelet[2951]: E0114 13:52:27.486314 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.550437 kubelet[2951]: W0114 13:52:27.486323 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.550437 kubelet[2951]: E0114 13:52:27.486331 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.550437 kubelet[2951]: E0114 13:52:27.490480 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.551471 kernel: audit: type=1327 audit(1768398747.376:526): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333313131613562363762656231393633616335326133343364616330 Jan 14 13:52:27.551840 kubelet[2951]: W0114 13:52:27.490492 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.551840 kubelet[2951]: E0114 13:52:27.490505 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.551840 kubelet[2951]: E0114 13:52:27.491873 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:27.551840 kubelet[2951]: E0114 13:52:27.498231 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.551840 kubelet[2951]: W0114 13:52:27.498240 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.551840 kubelet[2951]: E0114 13:52:27.498249 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.551840 kubelet[2951]: E0114 13:52:27.498463 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.551840 kubelet[2951]: W0114 13:52:27.498472 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.551840 kubelet[2951]: E0114 13:52:27.498481 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.551840 kubelet[2951]: E0114 13:52:27.502325 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.553190 kubelet[2951]: W0114 13:52:27.502335 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.553190 kubelet[2951]: E0114 13:52:27.502345 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.376000 audit: BPF prog-id=150 op=UNLOAD Jan 14 13:52:27.376000 audit[3408]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3395 pid=3408 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:27.376000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333313131613562363762656231393633616335326133343364616330 Jan 14 13:52:27.376000 audit: BPF prog-id=151 op=LOAD Jan 14 13:52:27.376000 audit[3408]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000174488 a2=98 a3=0 items=0 ppid=3395 pid=3408 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:27.376000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333313131613562363762656231393633616335326133343364616330 Jan 14 13:52:27.376000 audit: BPF prog-id=152 op=LOAD Jan 14 13:52:27.376000 audit[3408]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000174218 a2=98 a3=0 items=0 ppid=3395 pid=3408 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:27.376000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333313131613562363762656231393633616335326133343364616330 Jan 14 13:52:27.376000 audit: BPF prog-id=152 op=UNLOAD Jan 14 13:52:27.376000 audit[3408]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3395 pid=3408 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:27.376000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333313131613562363762656231393633616335326133343364616330 Jan 14 13:52:27.376000 audit: BPF prog-id=151 op=UNLOAD Jan 14 13:52:27.376000 audit[3408]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3395 pid=3408 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:27.376000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333313131613562363762656231393633616335326133343364616330 Jan 14 13:52:27.377000 audit: BPF prog-id=153 op=LOAD Jan 14 13:52:27.377000 audit[3408]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001746e8 a2=98 a3=0 items=0 ppid=3395 pid=3408 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:27.377000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333313131613562363762656231393633616335326133343364616330 Jan 14 13:52:27.568315 kubelet[2951]: E0114 13:52:27.567869 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.568315 kubelet[2951]: W0114 13:52:27.567901 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.568315 kubelet[2951]: E0114 13:52:27.567928 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.571296 kubelet[2951]: E0114 13:52:27.571278 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.574295 kubelet[2951]: W0114 13:52:27.574275 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.577350 kubelet[2951]: E0114 13:52:27.577328 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.578298 kubelet[2951]: E0114 13:52:27.578283 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.580448 kubelet[2951]: W0114 13:52:27.580176 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.580448 kubelet[2951]: E0114 13:52:27.580197 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.581250 kubelet[2951]: E0114 13:52:27.580943 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.581250 kubelet[2951]: W0114 13:52:27.580957 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.581250 kubelet[2951]: E0114 13:52:27.580970 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.584206 kubelet[2951]: E0114 13:52:27.582294 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.584304 kubelet[2951]: W0114 13:52:27.584286 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.584376 kubelet[2951]: E0114 13:52:27.584359 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.720509 kubelet[2951]: E0114 13:52:27.715429 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:27.721000 audit[3527]: NETFILTER_CFG table=filter:111 family=2 entries=22 op=nft_register_rule pid=3527 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:27.725228 kubelet[2951]: W0114 13:52:27.723501 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:27.725228 kubelet[2951]: E0114 13:52:27.723849 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:27.721000 audit[3527]: SYSCALL arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7ffd2aee68e0 a2=0 a3=7ffd2aee68cc items=0 ppid=3117 pid=3527 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:27.721000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:27.748000 audit[3527]: NETFILTER_CFG table=nat:112 family=2 entries=12 op=nft_register_rule pid=3527 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:27.748000 audit[3527]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd2aee68e0 a2=0 a3=0 items=0 ppid=3117 pid=3527 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:27.748000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:27.807275 containerd[1635]: time="2026-01-14T13:52:27.805853444Z" level=info msg="connecting to shim 14ebc5df29f97bd4359b3e55957c05fc18c0d3a2b470daa31b428fd78cb16055" address="unix:///run/containerd/s/aa914e9f8d13f41cdee1957eb5a9836fb7fc6d9f54222f9c02ed706a6d9468f8" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:52:27.913973 containerd[1635]: time="2026-01-14T13:52:27.912243779Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5677687b8b-zn66h,Uid:c72902a2-6254-4e22-9e4c-44098b40c0f6,Namespace:calico-system,Attempt:0,} returns sandbox id \"33111a5b67beb1963ac52a343dac01563ee0c2e3fdcc951af612a9f4c0163ece\"" Jan 14 13:52:27.921248 kubelet[2951]: E0114 13:52:27.919323 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:27.932870 containerd[1635]: time="2026-01-14T13:52:27.929979517Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\"" Jan 14 13:52:28.139332 systemd[1]: Started cri-containerd-14ebc5df29f97bd4359b3e55957c05fc18c0d3a2b470daa31b428fd78cb16055.scope - libcontainer container 14ebc5df29f97bd4359b3e55957c05fc18c0d3a2b470daa31b428fd78cb16055. Jan 14 13:52:28.331000 audit: BPF prog-id=154 op=LOAD Jan 14 13:52:28.334000 audit: BPF prog-id=155 op=LOAD Jan 14 13:52:28.334000 audit[3557]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=3539 pid=3557 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:28.334000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134656263356466323966393762643433353962336535353935376330 Jan 14 13:52:28.336000 audit: BPF prog-id=155 op=UNLOAD Jan 14 13:52:28.336000 audit[3557]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3539 pid=3557 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:28.336000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134656263356466323966393762643433353962336535353935376330 Jan 14 13:52:28.338000 audit: BPF prog-id=156 op=LOAD Jan 14 13:52:28.338000 audit[3557]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=3539 pid=3557 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:28.338000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134656263356466323966393762643433353962336535353935376330 Jan 14 13:52:28.340000 audit: BPF prog-id=157 op=LOAD Jan 14 13:52:28.340000 audit[3557]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=3539 pid=3557 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:28.340000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134656263356466323966393762643433353962336535353935376330 Jan 14 13:52:28.340000 audit: BPF prog-id=157 op=UNLOAD Jan 14 13:52:28.340000 audit[3557]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3539 pid=3557 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:28.340000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134656263356466323966393762643433353962336535353935376330 Jan 14 13:52:28.340000 audit: BPF prog-id=156 op=UNLOAD Jan 14 13:52:28.340000 audit[3557]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3539 pid=3557 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:28.340000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134656263356466323966393762643433353962336535353935376330 Jan 14 13:52:28.340000 audit: BPF prog-id=158 op=LOAD Jan 14 13:52:28.340000 audit[3557]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=3539 pid=3557 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:28.340000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134656263356466323966393762643433353962336535353935376330 Jan 14 13:52:28.588392 containerd[1635]: time="2026-01-14T13:52:28.587285001Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-mt5qd,Uid:e067985e-0631-4d20-b684-e59dafe9a54d,Namespace:calico-system,Attempt:0,} returns sandbox id \"14ebc5df29f97bd4359b3e55957c05fc18c0d3a2b470daa31b428fd78cb16055\"" Jan 14 13:52:28.594374 kubelet[2951]: E0114 13:52:28.594346 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:29.357473 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount56266804.mount: Deactivated successfully. Jan 14 13:52:29.383361 kubelet[2951]: E0114 13:52:29.382922 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:31.380921 kubelet[2951]: E0114 13:52:31.380246 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:33.381381 kubelet[2951]: E0114 13:52:33.380503 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:34.221856 containerd[1635]: time="2026-01-14T13:52:34.220983190Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:52:34.224259 containerd[1635]: time="2026-01-14T13:52:34.224216570Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.4: active requests=0, bytes read=35230631" Jan 14 13:52:34.233196 containerd[1635]: time="2026-01-14T13:52:34.232824077Z" level=info msg="ImageCreate event name:\"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:52:34.239901 containerd[1635]: time="2026-01-14T13:52:34.239858092Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:52:34.241416 containerd[1635]: time="2026-01-14T13:52:34.241375950Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.4\" with image id \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\", size \"35234482\" in 6.310904086s" Jan 14 13:52:34.242010 containerd[1635]: time="2026-01-14T13:52:34.241885438Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\" returns image reference \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\"" Jan 14 13:52:34.267208 containerd[1635]: time="2026-01-14T13:52:34.264294694Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\"" Jan 14 13:52:34.337871 containerd[1635]: time="2026-01-14T13:52:34.337449010Z" level=info msg="CreateContainer within sandbox \"33111a5b67beb1963ac52a343dac01563ee0c2e3fdcc951af612a9f4c0163ece\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Jan 14 13:52:34.429827 containerd[1635]: time="2026-01-14T13:52:34.420941407Z" level=info msg="Container 538cbab430da3f35472f6e1ca29a9b893186799d2e67c12263dad697c930a4b1: CDI devices from CRI Config.CDIDevices: []" Jan 14 13:52:34.509970 containerd[1635]: time="2026-01-14T13:52:34.502476417Z" level=info msg="CreateContainer within sandbox \"33111a5b67beb1963ac52a343dac01563ee0c2e3fdcc951af612a9f4c0163ece\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"538cbab430da3f35472f6e1ca29a9b893186799d2e67c12263dad697c930a4b1\"" Jan 14 13:52:34.517480 containerd[1635]: time="2026-01-14T13:52:34.516865208Z" level=info msg="StartContainer for \"538cbab430da3f35472f6e1ca29a9b893186799d2e67c12263dad697c930a4b1\"" Jan 14 13:52:34.529979 containerd[1635]: time="2026-01-14T13:52:34.528946263Z" level=info msg="connecting to shim 538cbab430da3f35472f6e1ca29a9b893186799d2e67c12263dad697c930a4b1" address="unix:///run/containerd/s/5d72708a39df1695ee7f76e1e0bd0d10bf4413a2936634aa81de4efbd141ed2e" protocol=ttrpc version=3 Jan 14 13:52:34.706408 systemd[1]: Started cri-containerd-538cbab430da3f35472f6e1ca29a9b893186799d2e67c12263dad697c930a4b1.scope - libcontainer container 538cbab430da3f35472f6e1ca29a9b893186799d2e67c12263dad697c930a4b1. Jan 14 13:52:34.815000 audit: BPF prog-id=159 op=LOAD Jan 14 13:52:34.851248 kernel: kauditd_printk_skb: 46 callbacks suppressed Jan 14 13:52:34.851347 kernel: audit: type=1334 audit(1768398754.815:543): prog-id=159 op=LOAD Jan 14 13:52:34.855693 kernel: audit: type=1334 audit(1768398754.818:544): prog-id=160 op=LOAD Jan 14 13:52:34.818000 audit: BPF prog-id=160 op=LOAD Jan 14 13:52:34.818000 audit[3591]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=3395 pid=3591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:34.947394 kernel: audit: type=1300 audit(1768398754.818:544): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=3395 pid=3591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:34.818000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533386362616234333064613366333534373266366531636132396139 Jan 14 13:52:35.017270 kernel: audit: type=1327 audit(1768398754.818:544): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533386362616234333064613366333534373266366531636132396139 Jan 14 13:52:34.818000 audit: BPF prog-id=160 op=UNLOAD Jan 14 13:52:34.818000 audit[3591]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3395 pid=3591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:35.103790 kernel: audit: type=1334 audit(1768398754.818:545): prog-id=160 op=UNLOAD Jan 14 13:52:35.103940 kernel: audit: type=1300 audit(1768398754.818:545): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3395 pid=3591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:34.818000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533386362616234333064613366333534373266366531636132396139 Jan 14 13:52:35.184219 kernel: audit: type=1327 audit(1768398754.818:545): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533386362616234333064613366333534373266366531636132396139 Jan 14 13:52:35.184370 kernel: audit: type=1334 audit(1768398754.818:546): prog-id=161 op=LOAD Jan 14 13:52:34.818000 audit: BPF prog-id=161 op=LOAD Jan 14 13:52:34.818000 audit[3591]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=3395 pid=3591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:35.271380 kernel: audit: type=1300 audit(1768398754.818:546): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=3395 pid=3591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:34.818000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533386362616234333064613366333534373266366531636132396139 Jan 14 13:52:35.277466 containerd[1635]: time="2026-01-14T13:52:35.277223519Z" level=info msg="StartContainer for \"538cbab430da3f35472f6e1ca29a9b893186799d2e67c12263dad697c930a4b1\" returns successfully" Jan 14 13:52:35.343176 kernel: audit: type=1327 audit(1768398754.818:546): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533386362616234333064613366333534373266366531636132396139 Jan 14 13:52:34.821000 audit: BPF prog-id=162 op=LOAD Jan 14 13:52:34.821000 audit[3591]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=3395 pid=3591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:34.821000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533386362616234333064613366333534373266366531636132396139 Jan 14 13:52:34.823000 audit: BPF prog-id=162 op=UNLOAD Jan 14 13:52:34.823000 audit[3591]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3395 pid=3591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:34.823000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533386362616234333064613366333534373266366531636132396139 Jan 14 13:52:34.823000 audit: BPF prog-id=161 op=UNLOAD Jan 14 13:52:34.823000 audit[3591]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3395 pid=3591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:34.823000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533386362616234333064613366333534373266366531636132396139 Jan 14 13:52:34.823000 audit: BPF prog-id=163 op=LOAD Jan 14 13:52:34.823000 audit[3591]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=3395 pid=3591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:34.823000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3533386362616234333064613366333534373266366531636132396139 Jan 14 13:52:35.384235 kubelet[2951]: E0114 13:52:35.383324 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:35.601840 kubelet[2951]: E0114 13:52:35.599232 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:35.664259 kubelet[2951]: I0114 13:52:35.660275 2951 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-5677687b8b-zn66h" podStartSLOduration=3.3417561669999998 podStartE2EDuration="9.660258709s" podCreationTimestamp="2026-01-14 13:52:26 +0000 UTC" firstStartedPulling="2026-01-14 13:52:27.929325759 +0000 UTC m=+55.177551028" lastFinishedPulling="2026-01-14 13:52:34.247828302 +0000 UTC m=+61.496053570" observedRunningTime="2026-01-14 13:52:35.645327236 +0000 UTC m=+62.893552515" watchObservedRunningTime="2026-01-14 13:52:35.660258709 +0000 UTC m=+62.908483977" Jan 14 13:52:35.679890 kubelet[2951]: E0114 13:52:35.679413 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.679890 kubelet[2951]: W0114 13:52:35.679436 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.679890 kubelet[2951]: E0114 13:52:35.679457 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.681489 kubelet[2951]: E0114 13:52:35.680323 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.681489 kubelet[2951]: W0114 13:52:35.680337 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.681489 kubelet[2951]: E0114 13:52:35.680348 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.681489 kubelet[2951]: E0114 13:52:35.681352 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.681489 kubelet[2951]: W0114 13:52:35.681361 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.681489 kubelet[2951]: E0114 13:52:35.681370 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.683195 kubelet[2951]: E0114 13:52:35.682255 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.683195 kubelet[2951]: W0114 13:52:35.682440 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.683195 kubelet[2951]: E0114 13:52:35.682454 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.684783 kubelet[2951]: E0114 13:52:35.684375 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.684783 kubelet[2951]: W0114 13:52:35.684400 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.684783 kubelet[2951]: E0114 13:52:35.684422 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.690222 kubelet[2951]: E0114 13:52:35.689235 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.690222 kubelet[2951]: W0114 13:52:35.689382 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.690222 kubelet[2951]: E0114 13:52:35.689396 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.693770 kubelet[2951]: E0114 13:52:35.692803 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.693770 kubelet[2951]: W0114 13:52:35.692944 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.693770 kubelet[2951]: E0114 13:52:35.692957 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.697419 kubelet[2951]: E0114 13:52:35.696852 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.697419 kubelet[2951]: W0114 13:52:35.696998 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.697419 kubelet[2951]: E0114 13:52:35.697011 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.701015 kubelet[2951]: E0114 13:52:35.700343 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.701015 kubelet[2951]: W0114 13:52:35.700488 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.701015 kubelet[2951]: E0114 13:52:35.700499 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.707461 kubelet[2951]: E0114 13:52:35.706427 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.707461 kubelet[2951]: W0114 13:52:35.707246 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.707461 kubelet[2951]: E0114 13:52:35.707263 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.711816 kubelet[2951]: E0114 13:52:35.710288 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.711816 kubelet[2951]: W0114 13:52:35.710302 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.711816 kubelet[2951]: E0114 13:52:35.710315 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.715265 kubelet[2951]: E0114 13:52:35.712419 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.715265 kubelet[2951]: W0114 13:52:35.714837 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.715265 kubelet[2951]: E0114 13:52:35.714853 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.717433 kubelet[2951]: E0114 13:52:35.717281 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.717433 kubelet[2951]: W0114 13:52:35.717296 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.717433 kubelet[2951]: E0114 13:52:35.717308 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.719311 kubelet[2951]: E0114 13:52:35.717799 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.719311 kubelet[2951]: W0114 13:52:35.717809 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.719311 kubelet[2951]: E0114 13:52:35.717819 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.719922 kubelet[2951]: E0114 13:52:35.719505 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.719922 kubelet[2951]: W0114 13:52:35.719779 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.719922 kubelet[2951]: E0114 13:52:35.719789 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.806946 kubelet[2951]: E0114 13:52:35.804890 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.806946 kubelet[2951]: W0114 13:52:35.804920 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.806946 kubelet[2951]: E0114 13:52:35.804946 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.808880 kubelet[2951]: E0114 13:52:35.807817 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.808880 kubelet[2951]: W0114 13:52:35.808345 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.808880 kubelet[2951]: E0114 13:52:35.808365 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.818487 kubelet[2951]: E0114 13:52:35.813953 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.818487 kubelet[2951]: W0114 13:52:35.814261 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.818487 kubelet[2951]: E0114 13:52:35.814279 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.818487 kubelet[2951]: E0114 13:52:35.815191 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.818487 kubelet[2951]: W0114 13:52:35.815202 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.818487 kubelet[2951]: E0114 13:52:35.815214 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.818487 kubelet[2951]: E0114 13:52:35.817449 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.818487 kubelet[2951]: W0114 13:52:35.817461 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.818487 kubelet[2951]: E0114 13:52:35.817473 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.818487 kubelet[2951]: E0114 13:52:35.818005 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.819307 kubelet[2951]: W0114 13:52:35.818015 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.819307 kubelet[2951]: E0114 13:52:35.818188 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.820838 kubelet[2951]: E0114 13:52:35.819863 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.820838 kubelet[2951]: W0114 13:52:35.819876 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.820838 kubelet[2951]: E0114 13:52:35.819887 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.826204 kubelet[2951]: E0114 13:52:35.825500 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.827449 kubelet[2951]: W0114 13:52:35.826502 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.829406 kubelet[2951]: E0114 13:52:35.828935 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.830862 kubelet[2951]: E0114 13:52:35.830237 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.830862 kubelet[2951]: W0114 13:52:35.830249 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.830862 kubelet[2951]: E0114 13:52:35.830259 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.832193 kubelet[2951]: E0114 13:52:35.831253 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.832193 kubelet[2951]: W0114 13:52:35.831421 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.832193 kubelet[2951]: E0114 13:52:35.831435 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.834333 kubelet[2951]: E0114 13:52:35.832415 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.834333 kubelet[2951]: W0114 13:52:35.832994 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.834333 kubelet[2951]: E0114 13:52:35.833010 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.834333 kubelet[2951]: E0114 13:52:35.834000 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.834333 kubelet[2951]: W0114 13:52:35.834011 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.834333 kubelet[2951]: E0114 13:52:35.834193 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.835205 kubelet[2951]: E0114 13:52:35.835000 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.835205 kubelet[2951]: W0114 13:52:35.835178 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.835205 kubelet[2951]: E0114 13:52:35.835188 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.848003 kubelet[2951]: E0114 13:52:35.845752 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.848003 kubelet[2951]: W0114 13:52:35.845771 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.848003 kubelet[2951]: E0114 13:52:35.845787 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.850201 kubelet[2951]: E0114 13:52:35.849396 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.850201 kubelet[2951]: W0114 13:52:35.849411 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.850201 kubelet[2951]: E0114 13:52:35.849422 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.855189 kubelet[2951]: E0114 13:52:35.853453 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.855189 kubelet[2951]: W0114 13:52:35.854222 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.855189 kubelet[2951]: E0114 13:52:35.854236 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.855189 kubelet[2951]: E0114 13:52:35.855008 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.855189 kubelet[2951]: W0114 13:52:35.855017 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.855189 kubelet[2951]: E0114 13:52:35.855187 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:35.856245 kubelet[2951]: E0114 13:52:35.855945 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:35.856245 kubelet[2951]: W0114 13:52:35.855956 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:35.856245 kubelet[2951]: E0114 13:52:35.855964 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.176841 containerd[1635]: time="2026-01-14T13:52:36.171370882Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:52:36.176841 containerd[1635]: time="2026-01-14T13:52:36.176439610Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4: active requests=0, bytes read=4442579" Jan 14 13:52:36.185229 containerd[1635]: time="2026-01-14T13:52:36.183201551Z" level=info msg="ImageCreate event name:\"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:52:36.212499 containerd[1635]: time="2026-01-14T13:52:36.209459077Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:52:36.212499 containerd[1635]: time="2026-01-14T13:52:36.211245931Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" with image id \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\", size \"5941314\" in 1.94691043s" Jan 14 13:52:36.212499 containerd[1635]: time="2026-01-14T13:52:36.211280997Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" returns image reference \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\"" Jan 14 13:52:36.270314 containerd[1635]: time="2026-01-14T13:52:36.269855231Z" level=info msg="CreateContainer within sandbox \"14ebc5df29f97bd4359b3e55957c05fc18c0d3a2b470daa31b428fd78cb16055\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Jan 14 13:52:36.348373 containerd[1635]: time="2026-01-14T13:52:36.347866142Z" level=info msg="Container e1bca9136139d957a53c8f2b5af8f00aa218d4df9c2d97f6c250dfbf053faea4: CDI devices from CRI Config.CDIDevices: []" Jan 14 13:52:36.409857 containerd[1635]: time="2026-01-14T13:52:36.408245381Z" level=info msg="CreateContainer within sandbox \"14ebc5df29f97bd4359b3e55957c05fc18c0d3a2b470daa31b428fd78cb16055\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"e1bca9136139d957a53c8f2b5af8f00aa218d4df9c2d97f6c250dfbf053faea4\"" Jan 14 13:52:36.412214 containerd[1635]: time="2026-01-14T13:52:36.410925060Z" level=info msg="StartContainer for \"e1bca9136139d957a53c8f2b5af8f00aa218d4df9c2d97f6c250dfbf053faea4\"" Jan 14 13:52:36.420460 containerd[1635]: time="2026-01-14T13:52:36.417992722Z" level=info msg="connecting to shim e1bca9136139d957a53c8f2b5af8f00aa218d4df9c2d97f6c250dfbf053faea4" address="unix:///run/containerd/s/aa914e9f8d13f41cdee1957eb5a9836fb7fc6d9f54222f9c02ed706a6d9468f8" protocol=ttrpc version=3 Jan 14 13:52:36.548217 systemd[1]: Started cri-containerd-e1bca9136139d957a53c8f2b5af8f00aa218d4df9c2d97f6c250dfbf053faea4.scope - libcontainer container e1bca9136139d957a53c8f2b5af8f00aa218d4df9c2d97f6c250dfbf053faea4. Jan 14 13:52:36.626865 kubelet[2951]: E0114 13:52:36.625871 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:36.644981 kubelet[2951]: E0114 13:52:36.643242 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.644981 kubelet[2951]: W0114 13:52:36.643266 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.644981 kubelet[2951]: E0114 13:52:36.643291 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.647249 kubelet[2951]: E0114 13:52:36.647006 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.647249 kubelet[2951]: W0114 13:52:36.647188 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.647249 kubelet[2951]: E0114 13:52:36.647211 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.653961 kubelet[2951]: E0114 13:52:36.652402 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.658907 kubelet[2951]: W0114 13:52:36.658780 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.658907 kubelet[2951]: E0114 13:52:36.658806 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.661434 kubelet[2951]: E0114 13:52:36.661249 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.661434 kubelet[2951]: W0114 13:52:36.661427 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.661836 kubelet[2951]: E0114 13:52:36.661446 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.677018 kubelet[2951]: E0114 13:52:36.675975 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.677018 kubelet[2951]: W0114 13:52:36.675991 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.677018 kubelet[2951]: E0114 13:52:36.676005 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.681969 kubelet[2951]: E0114 13:52:36.679374 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.681969 kubelet[2951]: W0114 13:52:36.679391 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.681969 kubelet[2951]: E0114 13:52:36.679408 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.681969 kubelet[2951]: E0114 13:52:36.680751 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.681969 kubelet[2951]: W0114 13:52:36.680760 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.681969 kubelet[2951]: E0114 13:52:36.680769 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.682345 kubelet[2951]: E0114 13:52:36.682272 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.682345 kubelet[2951]: W0114 13:52:36.682282 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.682345 kubelet[2951]: E0114 13:52:36.682292 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.690401 kubelet[2951]: E0114 13:52:36.688829 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.690401 kubelet[2951]: W0114 13:52:36.688984 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.690401 kubelet[2951]: E0114 13:52:36.688996 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.695837 kubelet[2951]: E0114 13:52:36.693424 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.695837 kubelet[2951]: W0114 13:52:36.694786 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.695837 kubelet[2951]: E0114 13:52:36.694800 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.697285 kubelet[2951]: E0114 13:52:36.696250 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.697285 kubelet[2951]: W0114 13:52:36.696437 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.697285 kubelet[2951]: E0114 13:52:36.696473 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.697285 kubelet[2951]: E0114 13:52:36.697271 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.697285 kubelet[2951]: W0114 13:52:36.697282 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.697453 kubelet[2951]: E0114 13:52:36.697299 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.698723 kubelet[2951]: E0114 13:52:36.697874 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.698723 kubelet[2951]: W0114 13:52:36.698354 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.698723 kubelet[2951]: E0114 13:52:36.698368 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.703387 kubelet[2951]: E0114 13:52:36.703204 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.703387 kubelet[2951]: W0114 13:52:36.703357 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.703387 kubelet[2951]: E0114 13:52:36.703375 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.711018 kubelet[2951]: E0114 13:52:36.709256 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.711018 kubelet[2951]: W0114 13:52:36.709271 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.711018 kubelet[2951]: E0114 13:52:36.709283 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.740426 kubelet[2951]: E0114 13:52:36.735844 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.740426 kubelet[2951]: W0114 13:52:36.736186 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.740426 kubelet[2951]: E0114 13:52:36.736221 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.740426 kubelet[2951]: E0114 13:52:36.737485 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.740426 kubelet[2951]: W0114 13:52:36.737497 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.740426 kubelet[2951]: E0114 13:52:36.737820 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.748807 kubelet[2951]: E0114 13:52:36.747489 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.748807 kubelet[2951]: W0114 13:52:36.747830 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.748807 kubelet[2951]: E0114 13:52:36.747851 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.759008 kubelet[2951]: E0114 13:52:36.753853 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.759008 kubelet[2951]: W0114 13:52:36.754005 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.759008 kubelet[2951]: E0114 13:52:36.754177 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.761963 kubelet[2951]: E0114 13:52:36.761791 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.772015 kubelet[2951]: W0114 13:52:36.761944 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.772015 kubelet[2951]: E0114 13:52:36.769966 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.772015 kubelet[2951]: E0114 13:52:36.771781 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.772015 kubelet[2951]: W0114 13:52:36.771793 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.772015 kubelet[2951]: E0114 13:52:36.771805 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.794859 kubelet[2951]: E0114 13:52:36.791983 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.794859 kubelet[2951]: W0114 13:52:36.792003 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.794859 kubelet[2951]: E0114 13:52:36.792021 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.799860 kubelet[2951]: E0114 13:52:36.798191 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.799860 kubelet[2951]: W0114 13:52:36.798348 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.799860 kubelet[2951]: E0114 13:52:36.798367 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.809421 kubelet[2951]: E0114 13:52:36.808826 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.809421 kubelet[2951]: W0114 13:52:36.808968 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.809421 kubelet[2951]: E0114 13:52:36.808989 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.817356 kubelet[2951]: E0114 13:52:36.814506 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.817356 kubelet[2951]: W0114 13:52:36.814940 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.817356 kubelet[2951]: E0114 13:52:36.814958 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.822190 kubelet[2951]: E0114 13:52:36.822000 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.822190 kubelet[2951]: W0114 13:52:36.822185 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.822285 kubelet[2951]: E0114 13:52:36.822205 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.831279 kubelet[2951]: E0114 13:52:36.830197 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.831279 kubelet[2951]: W0114 13:52:36.830354 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.831279 kubelet[2951]: E0114 13:52:36.830369 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.843355 kubelet[2951]: E0114 13:52:36.842866 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.843355 kubelet[2951]: W0114 13:52:36.843201 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.843355 kubelet[2951]: E0114 13:52:36.843223 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.852826 kubelet[2951]: E0114 13:52:36.851237 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.852826 kubelet[2951]: W0114 13:52:36.851251 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.852826 kubelet[2951]: E0114 13:52:36.851265 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.861346 kubelet[2951]: E0114 13:52:36.858981 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.861346 kubelet[2951]: W0114 13:52:36.859272 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.861346 kubelet[2951]: E0114 13:52:36.859292 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.868777 kubelet[2951]: E0114 13:52:36.867790 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.868777 kubelet[2951]: W0114 13:52:36.867934 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.868777 kubelet[2951]: E0114 13:52:36.867949 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.878399 kubelet[2951]: E0114 13:52:36.877812 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.878399 kubelet[2951]: W0114 13:52:36.877956 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.878399 kubelet[2951]: E0114 13:52:36.877970 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.885459 kubelet[2951]: E0114 13:52:36.885283 2951 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 13:52:36.885459 kubelet[2951]: W0114 13:52:36.885426 2951 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 13:52:36.885459 kubelet[2951]: E0114 13:52:36.885440 2951 plugins.go:697] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 13:52:36.991000 audit[3722]: NETFILTER_CFG table=filter:113 family=2 entries=21 op=nft_register_rule pid=3722 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:36.991000 audit[3722]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffd9886f210 a2=0 a3=7ffd9886f1fc items=0 ppid=3117 pid=3722 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:36.991000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:37.003000 audit[3722]: NETFILTER_CFG table=nat:114 family=2 entries=19 op=nft_register_chain pid=3722 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:52:37.003000 audit[3722]: SYSCALL arch=c000003e syscall=46 success=yes exit=6276 a0=3 a1=7ffd9886f210 a2=0 a3=7ffd9886f1fc items=0 ppid=3117 pid=3722 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:37.003000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:52:37.056000 audit: BPF prog-id=164 op=LOAD Jan 14 13:52:37.056000 audit[3668]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00009e488 a2=98 a3=0 items=0 ppid=3539 pid=3668 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:37.056000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531626361393133363133396439353761353363386632623561663866 Jan 14 13:52:37.057000 audit: BPF prog-id=165 op=LOAD Jan 14 13:52:37.057000 audit[3668]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00009e218 a2=98 a3=0 items=0 ppid=3539 pid=3668 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:37.057000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531626361393133363133396439353761353363386632623561663866 Jan 14 13:52:37.057000 audit: BPF prog-id=165 op=UNLOAD Jan 14 13:52:37.057000 audit[3668]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3539 pid=3668 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:37.057000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531626361393133363133396439353761353363386632623561663866 Jan 14 13:52:37.057000 audit: BPF prog-id=164 op=UNLOAD Jan 14 13:52:37.057000 audit[3668]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3539 pid=3668 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:37.057000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531626361393133363133396439353761353363386632623561663866 Jan 14 13:52:37.057000 audit: BPF prog-id=166 op=LOAD Jan 14 13:52:37.057000 audit[3668]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00009e6e8 a2=98 a3=0 items=0 ppid=3539 pid=3668 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:37.057000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531626361393133363133396439353761353363386632623561663866 Jan 14 13:52:37.277914 containerd[1635]: time="2026-01-14T13:52:37.277879079Z" level=info msg="StartContainer for \"e1bca9136139d957a53c8f2b5af8f00aa218d4df9c2d97f6c250dfbf053faea4\" returns successfully" Jan 14 13:52:37.344294 systemd[1]: cri-containerd-e1bca9136139d957a53c8f2b5af8f00aa218d4df9c2d97f6c250dfbf053faea4.scope: Deactivated successfully. Jan 14 13:52:37.350000 audit: BPF prog-id=166 op=UNLOAD Jan 14 13:52:37.357901 containerd[1635]: time="2026-01-14T13:52:37.356954073Z" level=info msg="received container exit event container_id:\"e1bca9136139d957a53c8f2b5af8f00aa218d4df9c2d97f6c250dfbf053faea4\" id:\"e1bca9136139d957a53c8f2b5af8f00aa218d4df9c2d97f6c250dfbf053faea4\" pid:3680 exited_at:{seconds:1768398757 nanos:348205990}" Jan 14 13:52:37.380412 kubelet[2951]: E0114 13:52:37.379228 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:37.642700 kubelet[2951]: E0114 13:52:37.642385 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:37.663198 kubelet[2951]: E0114 13:52:37.659390 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:37.672253 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e1bca9136139d957a53c8f2b5af8f00aa218d4df9c2d97f6c250dfbf053faea4-rootfs.mount: Deactivated successfully. Jan 14 13:52:38.651926 kubelet[2951]: E0114 13:52:38.651832 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:38.657928 containerd[1635]: time="2026-01-14T13:52:38.657845908Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\"" Jan 14 13:52:39.383391 kubelet[2951]: E0114 13:52:39.380419 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:41.380436 kubelet[2951]: E0114 13:52:41.379965 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:43.384985 kubelet[2951]: E0114 13:52:43.384941 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:45.381032 kubelet[2951]: E0114 13:52:45.380393 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:47.382659 kubelet[2951]: E0114 13:52:47.381396 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:49.384371 kubelet[2951]: E0114 13:52:49.384326 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:50.573872 containerd[1635]: time="2026-01-14T13:52:50.567977348Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:52:50.577495 containerd[1635]: time="2026-01-14T13:52:50.577304860Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.4: active requests=0, bytes read=70442291" Jan 14 13:52:50.586879 containerd[1635]: time="2026-01-14T13:52:50.581021287Z" level=info msg="ImageCreate event name:\"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:52:50.591334 containerd[1635]: time="2026-01-14T13:52:50.588753152Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:52:50.591334 containerd[1635]: time="2026-01-14T13:52:50.589974122Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.4\" with image id \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\", size \"71941459\" in 11.93208315s" Jan 14 13:52:50.591334 containerd[1635]: time="2026-01-14T13:52:50.589996734Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\" returns image reference \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\"" Jan 14 13:52:50.642281 containerd[1635]: time="2026-01-14T13:52:50.640444541Z" level=info msg="CreateContainer within sandbox \"14ebc5df29f97bd4359b3e55957c05fc18c0d3a2b470daa31b428fd78cb16055\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jan 14 13:52:50.681863 containerd[1635]: time="2026-01-14T13:52:50.681821945Z" level=info msg="Container 2124af40fb40daacf0ca94d0b849f9c74bf93dbb83b75e21d7aa58b2851e4cf4: CDI devices from CRI Config.CDIDevices: []" Jan 14 13:52:50.744466 containerd[1635]: time="2026-01-14T13:52:50.744428673Z" level=info msg="CreateContainer within sandbox \"14ebc5df29f97bd4359b3e55957c05fc18c0d3a2b470daa31b428fd78cb16055\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"2124af40fb40daacf0ca94d0b849f9c74bf93dbb83b75e21d7aa58b2851e4cf4\"" Jan 14 13:52:50.758906 containerd[1635]: time="2026-01-14T13:52:50.758871453Z" level=info msg="StartContainer for \"2124af40fb40daacf0ca94d0b849f9c74bf93dbb83b75e21d7aa58b2851e4cf4\"" Jan 14 13:52:50.768059 containerd[1635]: time="2026-01-14T13:52:50.766301856Z" level=info msg="connecting to shim 2124af40fb40daacf0ca94d0b849f9c74bf93dbb83b75e21d7aa58b2851e4cf4" address="unix:///run/containerd/s/aa914e9f8d13f41cdee1957eb5a9836fb7fc6d9f54222f9c02ed706a6d9468f8" protocol=ttrpc version=3 Jan 14 13:52:51.000265 systemd[1]: Started cri-containerd-2124af40fb40daacf0ca94d0b849f9c74bf93dbb83b75e21d7aa58b2851e4cf4.scope - libcontainer container 2124af40fb40daacf0ca94d0b849f9c74bf93dbb83b75e21d7aa58b2851e4cf4. Jan 14 13:52:51.306000 audit: BPF prog-id=167 op=LOAD Jan 14 13:52:51.342451 kernel: kauditd_printk_skb: 34 callbacks suppressed Jan 14 13:52:51.343283 kernel: audit: type=1334 audit(1768398771.306:559): prog-id=167 op=LOAD Jan 14 13:52:51.306000 audit[3765]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001b0488 a2=98 a3=0 items=0 ppid=3539 pid=3765 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:51.379860 kubelet[2951]: E0114 13:52:51.379322 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:51.415030 kernel: audit: type=1300 audit(1768398771.306:559): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001b0488 a2=98 a3=0 items=0 ppid=3539 pid=3765 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:51.306000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231323461663430666234306461616366306361393464306238343966 Jan 14 13:52:51.486748 kernel: audit: type=1327 audit(1768398771.306:559): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231323461663430666234306461616366306361393464306238343966 Jan 14 13:52:51.486865 kernel: audit: type=1334 audit(1768398771.306:560): prog-id=168 op=LOAD Jan 14 13:52:51.306000 audit: BPF prog-id=168 op=LOAD Jan 14 13:52:51.306000 audit[3765]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001b0218 a2=98 a3=0 items=0 ppid=3539 pid=3765 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:51.564442 kernel: audit: type=1300 audit(1768398771.306:560): arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001b0218 a2=98 a3=0 items=0 ppid=3539 pid=3765 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:51.565302 kernel: audit: type=1327 audit(1768398771.306:560): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231323461663430666234306461616366306361393464306238343966 Jan 14 13:52:51.306000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231323461663430666234306461616366306361393464306238343966 Jan 14 13:52:51.631275 kernel: audit: type=1334 audit(1768398771.306:561): prog-id=168 op=UNLOAD Jan 14 13:52:51.306000 audit: BPF prog-id=168 op=UNLOAD Jan 14 13:52:51.668485 kernel: audit: type=1300 audit(1768398771.306:561): arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3539 pid=3765 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:51.306000 audit[3765]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3539 pid=3765 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:51.702028 containerd[1635]: time="2026-01-14T13:52:51.693991653Z" level=info msg="StartContainer for \"2124af40fb40daacf0ca94d0b849f9c74bf93dbb83b75e21d7aa58b2851e4cf4\" returns successfully" Jan 14 13:52:51.722288 kernel: audit: type=1327 audit(1768398771.306:561): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231323461663430666234306461616366306361393464306238343966 Jan 14 13:52:51.306000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231323461663430666234306461616366306361393464306238343966 Jan 14 13:52:51.306000 audit: BPF prog-id=167 op=UNLOAD Jan 14 13:52:51.823250 kernel: audit: type=1334 audit(1768398771.306:562): prog-id=167 op=UNLOAD Jan 14 13:52:51.306000 audit[3765]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3539 pid=3765 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:51.306000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231323461663430666234306461616366306361393464306238343966 Jan 14 13:52:51.306000 audit: BPF prog-id=169 op=LOAD Jan 14 13:52:51.306000 audit[3765]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001b06e8 a2=98 a3=0 items=0 ppid=3539 pid=3765 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:52:51.306000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231323461663430666234306461616366306361393464306238343966 Jan 14 13:52:51.840435 kubelet[2951]: E0114 13:52:51.838846 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:52.856061 kubelet[2951]: E0114 13:52:52.854901 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:53.379948 kubelet[2951]: E0114 13:52:53.379089 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:55.386377 kubelet[2951]: E0114 13:52:55.384486 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:57.383464 kubelet[2951]: E0114 13:52:57.381023 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:52:57.850076 systemd[1]: cri-containerd-2124af40fb40daacf0ca94d0b849f9c74bf93dbb83b75e21d7aa58b2851e4cf4.scope: Deactivated successfully. Jan 14 13:52:57.855000 audit: BPF prog-id=169 op=UNLOAD Jan 14 13:52:57.851857 systemd[1]: cri-containerd-2124af40fb40daacf0ca94d0b849f9c74bf93dbb83b75e21d7aa58b2851e4cf4.scope: Consumed 4.887s CPU time, 181.3M memory peak, 3.3M read from disk, 171.3M written to disk. Jan 14 13:52:57.874062 kernel: kauditd_printk_skb: 5 callbacks suppressed Jan 14 13:52:57.874324 kernel: audit: type=1334 audit(1768398777.855:564): prog-id=169 op=UNLOAD Jan 14 13:52:57.907086 containerd[1635]: time="2026-01-14T13:52:57.905837693Z" level=info msg="received container exit event container_id:\"2124af40fb40daacf0ca94d0b849f9c74bf93dbb83b75e21d7aa58b2851e4cf4\" id:\"2124af40fb40daacf0ca94d0b849f9c74bf93dbb83b75e21d7aa58b2851e4cf4\" pid:3778 exited_at:{seconds:1768398777 nanos:892091925}" Jan 14 13:52:58.062816 kubelet[2951]: I0114 13:52:58.059102 2951 kubelet_node_status.go:439] "Fast updating node status as it just became ready" Jan 14 13:52:58.190401 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2124af40fb40daacf0ca94d0b849f9c74bf93dbb83b75e21d7aa58b2851e4cf4-rootfs.mount: Deactivated successfully. Jan 14 13:52:58.377943 systemd[1]: Created slice kubepods-burstable-pode6501533_023e_4ab4_8991_e08f2d55a030.slice - libcontainer container kubepods-burstable-pode6501533_023e_4ab4_8991_e08f2d55a030.slice. Jan 14 13:52:58.403854 systemd[1]: Created slice kubepods-besteffort-pod2b372f30_5f27_467a_8d6a_b220df9d05b0.slice - libcontainer container kubepods-besteffort-pod2b372f30_5f27_467a_8d6a_b220df9d05b0.slice. Jan 14 13:52:58.421396 systemd[1]: Created slice kubepods-besteffort-podd362a068_c796_4380_9ca2_cf2fbba56f1d.slice - libcontainer container kubepods-besteffort-podd362a068_c796_4380_9ca2_cf2fbba56f1d.slice. Jan 14 13:52:58.455461 systemd[1]: Created slice kubepods-besteffort-pod3a917c9a_ea97_482d_9194_bbf322e47535.slice - libcontainer container kubepods-besteffort-pod3a917c9a_ea97_482d_9194_bbf322e47535.slice. Jan 14 13:52:58.467823 kubelet[2951]: I0114 13:52:58.466800 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6501533-023e-4ab4-8991-e08f2d55a030-config-volume\") pod \"coredns-66bc5c9577-r9sfd\" (UID: \"e6501533-023e-4ab4-8991-e08f2d55a030\") " pod="kube-system/coredns-66bc5c9577-r9sfd" Jan 14 13:52:58.467823 kubelet[2951]: I0114 13:52:58.467022 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8vcg\" (UniqueName: \"kubernetes.io/projected/e6501533-023e-4ab4-8991-e08f2d55a030-kube-api-access-g8vcg\") pod \"coredns-66bc5c9577-r9sfd\" (UID: \"e6501533-023e-4ab4-8991-e08f2d55a030\") " pod="kube-system/coredns-66bc5c9577-r9sfd" Jan 14 13:52:58.467823 kubelet[2951]: I0114 13:52:58.467056 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph7bz\" (UniqueName: \"kubernetes.io/projected/d362a068-c796-4380-9ca2-cf2fbba56f1d-kube-api-access-ph7bz\") pod \"whisker-6f5bbf559f-5j9gg\" (UID: \"d362a068-c796-4380-9ca2-cf2fbba56f1d\") " pod="calico-system/whisker-6f5bbf559f-5j9gg" Jan 14 13:52:58.467823 kubelet[2951]: I0114 13:52:58.467086 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/2b372f30-5f27-467a-8d6a-b220df9d05b0-goldmane-key-pair\") pod \"goldmane-7c778bb748-8wp8q\" (UID: \"2b372f30-5f27-467a-8d6a-b220df9d05b0\") " pod="calico-system/goldmane-7c778bb748-8wp8q" Jan 14 13:52:58.467823 kubelet[2951]: I0114 13:52:58.467107 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d362a068-c796-4380-9ca2-cf2fbba56f1d-whisker-ca-bundle\") pod \"whisker-6f5bbf559f-5j9gg\" (UID: \"d362a068-c796-4380-9ca2-cf2fbba56f1d\") " pod="calico-system/whisker-6f5bbf559f-5j9gg" Jan 14 13:52:58.469110 kubelet[2951]: I0114 13:52:58.467315 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b372f30-5f27-467a-8d6a-b220df9d05b0-config\") pod \"goldmane-7c778bb748-8wp8q\" (UID: \"2b372f30-5f27-467a-8d6a-b220df9d05b0\") " pod="calico-system/goldmane-7c778bb748-8wp8q" Jan 14 13:52:58.469110 kubelet[2951]: I0114 13:52:58.467342 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b372f30-5f27-467a-8d6a-b220df9d05b0-goldmane-ca-bundle\") pod \"goldmane-7c778bb748-8wp8q\" (UID: \"2b372f30-5f27-467a-8d6a-b220df9d05b0\") " pod="calico-system/goldmane-7c778bb748-8wp8q" Jan 14 13:52:58.469110 kubelet[2951]: I0114 13:52:58.467363 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/d362a068-c796-4380-9ca2-cf2fbba56f1d-whisker-backend-key-pair\") pod \"whisker-6f5bbf559f-5j9gg\" (UID: \"d362a068-c796-4380-9ca2-cf2fbba56f1d\") " pod="calico-system/whisker-6f5bbf559f-5j9gg" Jan 14 13:52:58.469110 kubelet[2951]: I0114 13:52:58.467390 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2q6d\" (UniqueName: \"kubernetes.io/projected/2b372f30-5f27-467a-8d6a-b220df9d05b0-kube-api-access-p2q6d\") pod \"goldmane-7c778bb748-8wp8q\" (UID: \"2b372f30-5f27-467a-8d6a-b220df9d05b0\") " pod="calico-system/goldmane-7c778bb748-8wp8q" Jan 14 13:52:58.495367 systemd[1]: Created slice kubepods-besteffort-podb158b3fd_5a6b_446c_92b0_d1c78d153ee9.slice - libcontainer container kubepods-besteffort-podb158b3fd_5a6b_446c_92b0_d1c78d153ee9.slice. Jan 14 13:52:58.559878 systemd[1]: Created slice kubepods-besteffort-pod539656bb_a7f6_491c_9c24_e13907f6a8fd.slice - libcontainer container kubepods-besteffort-pod539656bb_a7f6_491c_9c24_e13907f6a8fd.slice. Jan 14 13:52:58.568481 kubelet[2951]: I0114 13:52:58.567917 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4efafa71-2a7d-4e25-8295-f778af674237-config-volume\") pod \"coredns-66bc5c9577-srck5\" (UID: \"4efafa71-2a7d-4e25-8295-f778af674237\") " pod="kube-system/coredns-66bc5c9577-srck5" Jan 14 13:52:58.568481 kubelet[2951]: I0114 13:52:58.568115 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/539656bb-a7f6-491c-9c24-e13907f6a8fd-calico-apiserver-certs\") pod \"calico-apiserver-6cbf5bf74d-k9nzg\" (UID: \"539656bb-a7f6-491c-9c24-e13907f6a8fd\") " pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" Jan 14 13:52:58.568481 kubelet[2951]: I0114 13:52:58.568322 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/3a917c9a-ea97-482d-9194-bbf322e47535-calico-apiserver-certs\") pod \"calico-apiserver-6cbf5bf74d-6k99t\" (UID: \"3a917c9a-ea97-482d-9194-bbf322e47535\") " pod="calico-apiserver/calico-apiserver-6cbf5bf74d-6k99t" Jan 14 13:52:58.568481 kubelet[2951]: I0114 13:52:58.568349 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czzbm\" (UniqueName: \"kubernetes.io/projected/3a917c9a-ea97-482d-9194-bbf322e47535-kube-api-access-czzbm\") pod \"calico-apiserver-6cbf5bf74d-6k99t\" (UID: \"3a917c9a-ea97-482d-9194-bbf322e47535\") " pod="calico-apiserver/calico-apiserver-6cbf5bf74d-6k99t" Jan 14 13:52:58.568481 kubelet[2951]: I0114 13:52:58.568366 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn48q\" (UniqueName: \"kubernetes.io/projected/539656bb-a7f6-491c-9c24-e13907f6a8fd-kube-api-access-wn48q\") pod \"calico-apiserver-6cbf5bf74d-k9nzg\" (UID: \"539656bb-a7f6-491c-9c24-e13907f6a8fd\") " pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" Jan 14 13:52:58.571058 kubelet[2951]: I0114 13:52:58.568380 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvflb\" (UniqueName: \"kubernetes.io/projected/4efafa71-2a7d-4e25-8295-f778af674237-kube-api-access-rvflb\") pod \"coredns-66bc5c9577-srck5\" (UID: \"4efafa71-2a7d-4e25-8295-f778af674237\") " pod="kube-system/coredns-66bc5c9577-srck5" Jan 14 13:52:58.571058 kubelet[2951]: I0114 13:52:58.568400 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b158b3fd-5a6b-446c-92b0-d1c78d153ee9-tigera-ca-bundle\") pod \"calico-kube-controllers-6cdc99f4d-mck2c\" (UID: \"b158b3fd-5a6b-446c-92b0-d1c78d153ee9\") " pod="calico-system/calico-kube-controllers-6cdc99f4d-mck2c" Jan 14 13:52:58.571058 kubelet[2951]: I0114 13:52:58.568431 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52svl\" (UniqueName: \"kubernetes.io/projected/b158b3fd-5a6b-446c-92b0-d1c78d153ee9-kube-api-access-52svl\") pod \"calico-kube-controllers-6cdc99f4d-mck2c\" (UID: \"b158b3fd-5a6b-446c-92b0-d1c78d153ee9\") " pod="calico-system/calico-kube-controllers-6cdc99f4d-mck2c" Jan 14 13:52:58.577062 systemd[1]: Created slice kubepods-burstable-pod4efafa71_2a7d_4e25_8295_f778af674237.slice - libcontainer container kubepods-burstable-pod4efafa71_2a7d_4e25_8295_f778af674237.slice. Jan 14 13:52:58.809047 containerd[1635]: time="2026-01-14T13:52:58.804095549Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6f5bbf559f-5j9gg,Uid:d362a068-c796-4380-9ca2-cf2fbba56f1d,Namespace:calico-system,Attempt:0,}" Jan 14 13:52:58.945880 containerd[1635]: time="2026-01-14T13:52:58.945831151Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6cbf5bf74d-k9nzg,Uid:539656bb-a7f6-491c-9c24-e13907f6a8fd,Namespace:calico-apiserver,Attempt:0,}" Jan 14 13:52:59.017340 kubelet[2951]: E0114 13:52:59.015354 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:59.057378 containerd[1635]: time="2026-01-14T13:52:59.056284472Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-r9sfd,Uid:e6501533-023e-4ab4-8991-e08f2d55a030,Namespace:kube-system,Attempt:0,}" Jan 14 13:52:59.089457 containerd[1635]: time="2026-01-14T13:52:59.085505491Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7c778bb748-8wp8q,Uid:2b372f30-5f27-467a-8d6a-b220df9d05b0,Namespace:calico-system,Attempt:0,}" Jan 14 13:52:59.095095 kubelet[2951]: E0114 13:52:59.094422 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:59.106107 containerd[1635]: time="2026-01-14T13:52:59.106007107Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6cbf5bf74d-6k99t,Uid:3a917c9a-ea97-482d-9194-bbf322e47535,Namespace:calico-apiserver,Attempt:0,}" Jan 14 13:52:59.109894 containerd[1635]: time="2026-01-14T13:52:59.109114446Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\"" Jan 14 13:52:59.122074 containerd[1635]: time="2026-01-14T13:52:59.122042206Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6cdc99f4d-mck2c,Uid:b158b3fd-5a6b-446c-92b0-d1c78d153ee9,Namespace:calico-system,Attempt:0,}" Jan 14 13:52:59.246407 kubelet[2951]: E0114 13:52:59.245443 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:52:59.450467 containerd[1635]: time="2026-01-14T13:52:59.449927040Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-srck5,Uid:4efafa71-2a7d-4e25-8295-f778af674237,Namespace:kube-system,Attempt:0,}" Jan 14 13:52:59.454462 systemd[1]: Created slice kubepods-besteffort-pod967b7513_2769_43e5_a113_9344f17909c3.slice - libcontainer container kubepods-besteffort-pod967b7513_2769_43e5_a113_9344f17909c3.slice. Jan 14 13:52:59.596862 containerd[1635]: time="2026-01-14T13:52:59.596463309Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-pvxlq,Uid:967b7513-2769-43e5-a113-9344f17909c3,Namespace:calico-system,Attempt:0,}" Jan 14 13:53:01.135729 containerd[1635]: time="2026-01-14T13:53:01.131895824Z" level=error msg="Failed to destroy network for sandbox \"7078f3e4eb06252235d41e7183ef38224f80e183a0b6d89302c45cbdf706e1ab\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.144006 systemd[1]: run-netns-cni\x2d9a0cdedf\x2dacad\x2d30e6\x2da0b6\x2d8dbae678ec0c.mount: Deactivated successfully. Jan 14 13:53:01.255458 containerd[1635]: time="2026-01-14T13:53:01.237444748Z" level=error msg="Failed to destroy network for sandbox \"dfdbee912882a755fa712f3a8c1903a55a20ecef46b5c0dda71f2ab6e0022afa\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.247752 systemd[1]: run-netns-cni\x2ded4f59e9\x2dfe2c\x2d45b3\x2dbe75\x2d111454565ed2.mount: Deactivated successfully. Jan 14 13:53:01.310931 containerd[1635]: time="2026-01-14T13:53:01.309492490Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-r9sfd,Uid:e6501533-023e-4ab4-8991-e08f2d55a030,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"7078f3e4eb06252235d41e7183ef38224f80e183a0b6d89302c45cbdf706e1ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.365859 containerd[1635]: time="2026-01-14T13:53:01.323890269Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6cdc99f4d-mck2c,Uid:b158b3fd-5a6b-446c-92b0-d1c78d153ee9,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"dfdbee912882a755fa712f3a8c1903a55a20ecef46b5c0dda71f2ab6e0022afa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.367992 containerd[1635]: time="2026-01-14T13:53:01.364904786Z" level=error msg="Failed to destroy network for sandbox \"fa2b8ede67d3afe923fe125a0c6086e265790286c54b1c80bb5d1d49206367c6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.383998 containerd[1635]: time="2026-01-14T13:53:01.377290737Z" level=error msg="Failed to destroy network for sandbox \"fe841a1ee4fe1cef4a8a710b9981fae0c2cba8970588da311d388542fd02f735\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.383998 containerd[1635]: time="2026-01-14T13:53:01.378900285Z" level=error msg="Failed to destroy network for sandbox \"329aaffe9eca494100b5d6baf5be258760e35094385067db22402ac51b79bc64\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.379388 systemd[1]: run-netns-cni\x2d15fc35ba\x2dee98\x2dcb90\x2dfdda\x2d50047c5dac21.mount: Deactivated successfully. Jan 14 13:53:01.388782 kubelet[2951]: E0114 13:53:01.377313 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dfdbee912882a755fa712f3a8c1903a55a20ecef46b5c0dda71f2ab6e0022afa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.388782 kubelet[2951]: E0114 13:53:01.377393 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dfdbee912882a755fa712f3a8c1903a55a20ecef46b5c0dda71f2ab6e0022afa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6cdc99f4d-mck2c" Jan 14 13:53:01.388782 kubelet[2951]: E0114 13:53:01.377425 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dfdbee912882a755fa712f3a8c1903a55a20ecef46b5c0dda71f2ab6e0022afa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6cdc99f4d-mck2c" Jan 14 13:53:01.388111 systemd[1]: run-netns-cni\x2d8a268f50\x2da661\x2d3dec\x2dbc31\x2d84b7f3df5dd0.mount: Deactivated successfully. Jan 14 13:53:01.391996 kubelet[2951]: E0114 13:53:01.377492 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6cdc99f4d-mck2c_calico-system(b158b3fd-5a6b-446c-92b0-d1c78d153ee9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6cdc99f4d-mck2c_calico-system(b158b3fd-5a6b-446c-92b0-d1c78d153ee9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"dfdbee912882a755fa712f3a8c1903a55a20ecef46b5c0dda71f2ab6e0022afa\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6cdc99f4d-mck2c" podUID="b158b3fd-5a6b-446c-92b0-d1c78d153ee9" Jan 14 13:53:01.391996 kubelet[2951]: E0114 13:53:01.378095 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7078f3e4eb06252235d41e7183ef38224f80e183a0b6d89302c45cbdf706e1ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.391996 kubelet[2951]: E0114 13:53:01.378130 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7078f3e4eb06252235d41e7183ef38224f80e183a0b6d89302c45cbdf706e1ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-r9sfd" Jan 14 13:53:01.388416 systemd[1]: run-netns-cni\x2dce3a744a\x2d8b6c\x2d07ec\x2d5985\x2d0fafec07007e.mount: Deactivated successfully. Jan 14 13:53:01.398305 containerd[1635]: time="2026-01-14T13:53:01.394925104Z" level=error msg="Failed to destroy network for sandbox \"bcdbbc108eb86398e7edb22eedccfeaa4869ca6fabc2e8e4894e34c2449eb68e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.398358 kubelet[2951]: E0114 13:53:01.378328 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7078f3e4eb06252235d41e7183ef38224f80e183a0b6d89302c45cbdf706e1ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-r9sfd" Jan 14 13:53:01.398358 kubelet[2951]: E0114 13:53:01.378375 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-r9sfd_kube-system(e6501533-023e-4ab4-8991-e08f2d55a030)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-r9sfd_kube-system(e6501533-023e-4ab4-8991-e08f2d55a030)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7078f3e4eb06252235d41e7183ef38224f80e183a0b6d89302c45cbdf706e1ab\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-r9sfd" podUID="e6501533-023e-4ab4-8991-e08f2d55a030" Jan 14 13:53:01.405337 systemd[1]: run-netns-cni\x2d189a4f35\x2d3d37\x2d9bcb\x2d2265\x2d2b66d6b49688.mount: Deactivated successfully. Jan 14 13:53:01.433947 containerd[1635]: time="2026-01-14T13:53:01.429390012Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-srck5,Uid:4efafa71-2a7d-4e25-8295-f778af674237,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa2b8ede67d3afe923fe125a0c6086e265790286c54b1c80bb5d1d49206367c6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.434914 kubelet[2951]: E0114 13:53:01.430903 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa2b8ede67d3afe923fe125a0c6086e265790286c54b1c80bb5d1d49206367c6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.434914 kubelet[2951]: E0114 13:53:01.430963 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa2b8ede67d3afe923fe125a0c6086e265790286c54b1c80bb5d1d49206367c6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-srck5" Jan 14 13:53:01.434914 kubelet[2951]: E0114 13:53:01.430990 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa2b8ede67d3afe923fe125a0c6086e265790286c54b1c80bb5d1d49206367c6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-srck5" Jan 14 13:53:01.441923 kubelet[2951]: E0114 13:53:01.431057 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-srck5_kube-system(4efafa71-2a7d-4e25-8295-f778af674237)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-srck5_kube-system(4efafa71-2a7d-4e25-8295-f778af674237)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"fa2b8ede67d3afe923fe125a0c6086e265790286c54b1c80bb5d1d49206367c6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-srck5" podUID="4efafa71-2a7d-4e25-8295-f778af674237" Jan 14 13:53:01.444461 containerd[1635]: time="2026-01-14T13:53:01.443963909Z" level=error msg="Failed to destroy network for sandbox \"b9552245dedbbadebe8ee6f611cbbf44efef8b3089711fbabc3e00b12444aa13\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.452091 containerd[1635]: time="2026-01-14T13:53:01.452044089Z" level=error msg="Failed to destroy network for sandbox \"12cfef14fe6d2aa786e27e3b2849062a423c603d1360e07437a83f2a343fb24f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.463053 containerd[1635]: time="2026-01-14T13:53:01.463002341Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7c778bb748-8wp8q,Uid:2b372f30-5f27-467a-8d6a-b220df9d05b0,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"bcdbbc108eb86398e7edb22eedccfeaa4869ca6fabc2e8e4894e34c2449eb68e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.464807 kubelet[2951]: E0114 13:53:01.464756 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bcdbbc108eb86398e7edb22eedccfeaa4869ca6fabc2e8e4894e34c2449eb68e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.464973 kubelet[2951]: E0114 13:53:01.464945 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bcdbbc108eb86398e7edb22eedccfeaa4869ca6fabc2e8e4894e34c2449eb68e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7c778bb748-8wp8q" Jan 14 13:53:01.465074 kubelet[2951]: E0114 13:53:01.465051 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bcdbbc108eb86398e7edb22eedccfeaa4869ca6fabc2e8e4894e34c2449eb68e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7c778bb748-8wp8q" Jan 14 13:53:01.468048 kubelet[2951]: E0114 13:53:01.468018 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-7c778bb748-8wp8q_calico-system(2b372f30-5f27-467a-8d6a-b220df9d05b0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-7c778bb748-8wp8q_calico-system(2b372f30-5f27-467a-8d6a-b220df9d05b0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"bcdbbc108eb86398e7edb22eedccfeaa4869ca6fabc2e8e4894e34c2449eb68e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-7c778bb748-8wp8q" podUID="2b372f30-5f27-467a-8d6a-b220df9d05b0" Jan 14 13:53:01.538501 containerd[1635]: time="2026-01-14T13:53:01.535892233Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-pvxlq,Uid:967b7513-2769-43e5-a113-9344f17909c3,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"329aaffe9eca494100b5d6baf5be258760e35094385067db22402ac51b79bc64\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.553461 kubelet[2951]: E0114 13:53:01.537905 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"329aaffe9eca494100b5d6baf5be258760e35094385067db22402ac51b79bc64\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.553461 kubelet[2951]: E0114 13:53:01.537968 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"329aaffe9eca494100b5d6baf5be258760e35094385067db22402ac51b79bc64\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-pvxlq" Jan 14 13:53:01.553461 kubelet[2951]: E0114 13:53:01.538001 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"329aaffe9eca494100b5d6baf5be258760e35094385067db22402ac51b79bc64\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-pvxlq" Jan 14 13:53:01.553968 kubelet[2951]: E0114 13:53:01.538070 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-pvxlq_calico-system(967b7513-2769-43e5-a113-9344f17909c3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-pvxlq_calico-system(967b7513-2769-43e5-a113-9344f17909c3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"329aaffe9eca494100b5d6baf5be258760e35094385067db22402ac51b79bc64\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:53:01.558042 containerd[1635]: time="2026-01-14T13:53:01.554771719Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6cbf5bf74d-k9nzg,Uid:539656bb-a7f6-491c-9c24-e13907f6a8fd,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"fe841a1ee4fe1cef4a8a710b9981fae0c2cba8970588da311d388542fd02f735\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.558839 kubelet[2951]: E0114 13:53:01.557878 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fe841a1ee4fe1cef4a8a710b9981fae0c2cba8970588da311d388542fd02f735\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.558839 kubelet[2951]: E0114 13:53:01.557921 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fe841a1ee4fe1cef4a8a710b9981fae0c2cba8970588da311d388542fd02f735\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" Jan 14 13:53:01.558839 kubelet[2951]: E0114 13:53:01.557937 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fe841a1ee4fe1cef4a8a710b9981fae0c2cba8970588da311d388542fd02f735\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" Jan 14 13:53:01.558920 kubelet[2951]: E0114 13:53:01.557977 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6cbf5bf74d-k9nzg_calico-apiserver(539656bb-a7f6-491c-9c24-e13907f6a8fd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6cbf5bf74d-k9nzg_calico-apiserver(539656bb-a7f6-491c-9c24-e13907f6a8fd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"fe841a1ee4fe1cef4a8a710b9981fae0c2cba8970588da311d388542fd02f735\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" podUID="539656bb-a7f6-491c-9c24-e13907f6a8fd" Jan 14 13:53:01.563907 containerd[1635]: time="2026-01-14T13:53:01.562963138Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6f5bbf559f-5j9gg,Uid:d362a068-c796-4380-9ca2-cf2fbba56f1d,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"b9552245dedbbadebe8ee6f611cbbf44efef8b3089711fbabc3e00b12444aa13\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.564393 kubelet[2951]: E0114 13:53:01.564107 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b9552245dedbbadebe8ee6f611cbbf44efef8b3089711fbabc3e00b12444aa13\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.564393 kubelet[2951]: E0114 13:53:01.564310 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b9552245dedbbadebe8ee6f611cbbf44efef8b3089711fbabc3e00b12444aa13\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-6f5bbf559f-5j9gg" Jan 14 13:53:01.564393 kubelet[2951]: E0114 13:53:01.564334 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b9552245dedbbadebe8ee6f611cbbf44efef8b3089711fbabc3e00b12444aa13\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-6f5bbf559f-5j9gg" Jan 14 13:53:01.564478 kubelet[2951]: E0114 13:53:01.564383 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-6f5bbf559f-5j9gg_calico-system(d362a068-c796-4380-9ca2-cf2fbba56f1d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-6f5bbf559f-5j9gg_calico-system(d362a068-c796-4380-9ca2-cf2fbba56f1d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b9552245dedbbadebe8ee6f611cbbf44efef8b3089711fbabc3e00b12444aa13\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-6f5bbf559f-5j9gg" podUID="d362a068-c796-4380-9ca2-cf2fbba56f1d" Jan 14 13:53:01.576096 containerd[1635]: time="2026-01-14T13:53:01.574804596Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6cbf5bf74d-6k99t,Uid:3a917c9a-ea97-482d-9194-bbf322e47535,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"12cfef14fe6d2aa786e27e3b2849062a423c603d1360e07437a83f2a343fb24f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.576462 kubelet[2951]: E0114 13:53:01.575015 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"12cfef14fe6d2aa786e27e3b2849062a423c603d1360e07437a83f2a343fb24f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:01.576462 kubelet[2951]: E0114 13:53:01.575073 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"12cfef14fe6d2aa786e27e3b2849062a423c603d1360e07437a83f2a343fb24f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-6k99t" Jan 14 13:53:01.576462 kubelet[2951]: E0114 13:53:01.575094 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"12cfef14fe6d2aa786e27e3b2849062a423c603d1360e07437a83f2a343fb24f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-6k99t" Jan 14 13:53:01.576878 kubelet[2951]: E0114 13:53:01.575315 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6cbf5bf74d-6k99t_calico-apiserver(3a917c9a-ea97-482d-9194-bbf322e47535)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6cbf5bf74d-6k99t_calico-apiserver(3a917c9a-ea97-482d-9194-bbf322e47535)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"12cfef14fe6d2aa786e27e3b2849062a423c603d1360e07437a83f2a343fb24f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-6k99t" podUID="3a917c9a-ea97-482d-9194-bbf322e47535" Jan 14 13:53:02.146943 systemd[1]: run-netns-cni\x2db90a1db0\x2d9ae3\x2df8a1\x2df39b\x2d6411b107a952.mount: Deactivated successfully. Jan 14 13:53:02.147918 systemd[1]: run-netns-cni\x2dd9542d9d\x2d46e3\x2d7f9c\x2d30b6\x2dded040063f3b.mount: Deactivated successfully. Jan 14 13:53:14.387003 kubelet[2951]: E0114 13:53:14.384988 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:14.424376 containerd[1635]: time="2026-01-14T13:53:14.423161952Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-pvxlq,Uid:967b7513-2769-43e5-a113-9344f17909c3,Namespace:calico-system,Attempt:0,}" Jan 14 13:53:14.830010 containerd[1635]: time="2026-01-14T13:53:14.829811760Z" level=error msg="Failed to destroy network for sandbox \"c796e0d0e03b2d4424dba271e9d29ee3d39aa1a7750ec018068db0b5de2a6bcd\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:14.833049 systemd[1]: run-netns-cni\x2d4c75536d\x2d24b6\x2d7240\x2df75a\x2d8010a7fc9f8a.mount: Deactivated successfully. Jan 14 13:53:14.839389 containerd[1635]: time="2026-01-14T13:53:14.839060119Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-pvxlq,Uid:967b7513-2769-43e5-a113-9344f17909c3,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"c796e0d0e03b2d4424dba271e9d29ee3d39aa1a7750ec018068db0b5de2a6bcd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:14.839671 kubelet[2951]: E0114 13:53:14.839638 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c796e0d0e03b2d4424dba271e9d29ee3d39aa1a7750ec018068db0b5de2a6bcd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:14.839830 kubelet[2951]: E0114 13:53:14.839695 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c796e0d0e03b2d4424dba271e9d29ee3d39aa1a7750ec018068db0b5de2a6bcd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-pvxlq" Jan 14 13:53:14.839830 kubelet[2951]: E0114 13:53:14.839787 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c796e0d0e03b2d4424dba271e9d29ee3d39aa1a7750ec018068db0b5de2a6bcd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-pvxlq" Jan 14 13:53:14.839891 kubelet[2951]: E0114 13:53:14.839841 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-pvxlq_calico-system(967b7513-2769-43e5-a113-9344f17909c3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-pvxlq_calico-system(967b7513-2769-43e5-a113-9344f17909c3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c796e0d0e03b2d4424dba271e9d29ee3d39aa1a7750ec018068db0b5de2a6bcd\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:53:15.406327 containerd[1635]: time="2026-01-14T13:53:15.406222866Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6cbf5bf74d-k9nzg,Uid:539656bb-a7f6-491c-9c24-e13907f6a8fd,Namespace:calico-apiserver,Attempt:0,}" Jan 14 13:53:15.408910 containerd[1635]: time="2026-01-14T13:53:15.408196982Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6cbf5bf74d-6k99t,Uid:3a917c9a-ea97-482d-9194-bbf322e47535,Namespace:calico-apiserver,Attempt:0,}" Jan 14 13:53:15.409702 containerd[1635]: time="2026-01-14T13:53:15.409090120Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6cdc99f4d-mck2c,Uid:b158b3fd-5a6b-446c-92b0-d1c78d153ee9,Namespace:calico-system,Attempt:0,}" Jan 14 13:53:15.559074 containerd[1635]: time="2026-01-14T13:53:15.559022217Z" level=error msg="Failed to destroy network for sandbox \"880be5d5a30c796c6792be7fd2d430b68205dd1f8b1e72b94c22556f4ec7884f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:15.562309 systemd[1]: run-netns-cni\x2decf807a0\x2dada4\x2dc16f\x2d3b0b\x2de07398c15f69.mount: Deactivated successfully. Jan 14 13:53:15.564145 containerd[1635]: time="2026-01-14T13:53:15.564034510Z" level=error msg="Failed to destroy network for sandbox \"0de6062cfde76d3dd1f82b68d8b80e6acbc4d776087fdaa88a4ab9420d1b83dd\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:15.567497 systemd[1]: run-netns-cni\x2d2558011a\x2d9ae4\x2d0e85\x2dc72b\x2ddc37f7c07dc4.mount: Deactivated successfully. Jan 14 13:53:15.591123 containerd[1635]: time="2026-01-14T13:53:15.591022916Z" level=error msg="Failed to destroy network for sandbox \"4610588067e022af796544bd36aaa6437feae2989d725df87a7f72537b751830\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:15.593927 systemd[1]: run-netns-cni\x2d5c6478ef\x2dcef9\x2d2e8e\x2d11b6\x2d970dce6abc19.mount: Deactivated successfully. Jan 14 13:53:15.601866 containerd[1635]: time="2026-01-14T13:53:15.601743157Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6cbf5bf74d-k9nzg,Uid:539656bb-a7f6-491c-9c24-e13907f6a8fd,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"0de6062cfde76d3dd1f82b68d8b80e6acbc4d776087fdaa88a4ab9420d1b83dd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:15.602843 kubelet[2951]: E0114 13:53:15.602676 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0de6062cfde76d3dd1f82b68d8b80e6acbc4d776087fdaa88a4ab9420d1b83dd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:15.603153 kubelet[2951]: E0114 13:53:15.602853 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0de6062cfde76d3dd1f82b68d8b80e6acbc4d776087fdaa88a4ab9420d1b83dd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" Jan 14 13:53:15.603153 kubelet[2951]: E0114 13:53:15.602878 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0de6062cfde76d3dd1f82b68d8b80e6acbc4d776087fdaa88a4ab9420d1b83dd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" Jan 14 13:53:15.603153 kubelet[2951]: E0114 13:53:15.602932 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6cbf5bf74d-k9nzg_calico-apiserver(539656bb-a7f6-491c-9c24-e13907f6a8fd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6cbf5bf74d-k9nzg_calico-apiserver(539656bb-a7f6-491c-9c24-e13907f6a8fd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0de6062cfde76d3dd1f82b68d8b80e6acbc4d776087fdaa88a4ab9420d1b83dd\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" podUID="539656bb-a7f6-491c-9c24-e13907f6a8fd" Jan 14 13:53:15.604651 containerd[1635]: time="2026-01-14T13:53:15.604473392Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6cdc99f4d-mck2c,Uid:b158b3fd-5a6b-446c-92b0-d1c78d153ee9,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"880be5d5a30c796c6792be7fd2d430b68205dd1f8b1e72b94c22556f4ec7884f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:15.606432 kubelet[2951]: E0114 13:53:15.606388 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"880be5d5a30c796c6792be7fd2d430b68205dd1f8b1e72b94c22556f4ec7884f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:15.606978 kubelet[2951]: E0114 13:53:15.606775 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"880be5d5a30c796c6792be7fd2d430b68205dd1f8b1e72b94c22556f4ec7884f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6cdc99f4d-mck2c" Jan 14 13:53:15.606978 kubelet[2951]: E0114 13:53:15.606805 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"880be5d5a30c796c6792be7fd2d430b68205dd1f8b1e72b94c22556f4ec7884f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6cdc99f4d-mck2c" Jan 14 13:53:15.606978 kubelet[2951]: E0114 13:53:15.606874 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6cdc99f4d-mck2c_calico-system(b158b3fd-5a6b-446c-92b0-d1c78d153ee9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6cdc99f4d-mck2c_calico-system(b158b3fd-5a6b-446c-92b0-d1c78d153ee9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"880be5d5a30c796c6792be7fd2d430b68205dd1f8b1e72b94c22556f4ec7884f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6cdc99f4d-mck2c" podUID="b158b3fd-5a6b-446c-92b0-d1c78d153ee9" Jan 14 13:53:15.607340 containerd[1635]: time="2026-01-14T13:53:15.607304497Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6cbf5bf74d-6k99t,Uid:3a917c9a-ea97-482d-9194-bbf322e47535,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"4610588067e022af796544bd36aaa6437feae2989d725df87a7f72537b751830\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:15.607838 kubelet[2951]: E0114 13:53:15.607734 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4610588067e022af796544bd36aaa6437feae2989d725df87a7f72537b751830\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:15.607838 kubelet[2951]: E0114 13:53:15.607763 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4610588067e022af796544bd36aaa6437feae2989d725df87a7f72537b751830\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-6k99t" Jan 14 13:53:15.607838 kubelet[2951]: E0114 13:53:15.607778 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4610588067e022af796544bd36aaa6437feae2989d725df87a7f72537b751830\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-6k99t" Jan 14 13:53:15.607932 kubelet[2951]: E0114 13:53:15.607808 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6cbf5bf74d-6k99t_calico-apiserver(3a917c9a-ea97-482d-9194-bbf322e47535)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6cbf5bf74d-6k99t_calico-apiserver(3a917c9a-ea97-482d-9194-bbf322e47535)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4610588067e022af796544bd36aaa6437feae2989d725df87a7f72537b751830\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-6k99t" podUID="3a917c9a-ea97-482d-9194-bbf322e47535" Jan 14 13:53:16.488931 containerd[1635]: time="2026-01-14T13:53:16.488806268Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6f5bbf559f-5j9gg,Uid:d362a068-c796-4380-9ca2-cf2fbba56f1d,Namespace:calico-system,Attempt:0,}" Jan 14 13:53:16.494491 containerd[1635]: time="2026-01-14T13:53:16.494017756Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7c778bb748-8wp8q,Uid:2b372f30-5f27-467a-8d6a-b220df9d05b0,Namespace:calico-system,Attempt:0,}" Jan 14 13:53:16.495477 kubelet[2951]: E0114 13:53:16.495177 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:16.497455 containerd[1635]: time="2026-01-14T13:53:16.497352144Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-srck5,Uid:4efafa71-2a7d-4e25-8295-f778af674237,Namespace:kube-system,Attempt:0,}" Jan 14 13:53:16.500949 kubelet[2951]: E0114 13:53:16.499076 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:16.502322 containerd[1635]: time="2026-01-14T13:53:16.502178227Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-r9sfd,Uid:e6501533-023e-4ab4-8991-e08f2d55a030,Namespace:kube-system,Attempt:0,}" Jan 14 13:53:16.695907 containerd[1635]: time="2026-01-14T13:53:16.695855343Z" level=error msg="Failed to destroy network for sandbox \"db22bded2633f5fc21f45943f81ba16134949c601cf5dcfbba61da88fc615187\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:16.701218 systemd[1]: run-netns-cni\x2d163d1a2c\x2dc9c1\x2d586a\x2d2db7\x2de972764eefa6.mount: Deactivated successfully. Jan 14 13:53:16.704794 containerd[1635]: time="2026-01-14T13:53:16.704757175Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7c778bb748-8wp8q,Uid:2b372f30-5f27-467a-8d6a-b220df9d05b0,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"db22bded2633f5fc21f45943f81ba16134949c601cf5dcfbba61da88fc615187\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:16.705604 kubelet[2951]: E0114 13:53:16.705434 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"db22bded2633f5fc21f45943f81ba16134949c601cf5dcfbba61da88fc615187\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:16.707156 kubelet[2951]: E0114 13:53:16.705492 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"db22bded2633f5fc21f45943f81ba16134949c601cf5dcfbba61da88fc615187\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7c778bb748-8wp8q" Jan 14 13:53:16.707156 kubelet[2951]: E0114 13:53:16.705631 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"db22bded2633f5fc21f45943f81ba16134949c601cf5dcfbba61da88fc615187\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-7c778bb748-8wp8q" Jan 14 13:53:16.707156 kubelet[2951]: E0114 13:53:16.705734 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-7c778bb748-8wp8q_calico-system(2b372f30-5f27-467a-8d6a-b220df9d05b0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-7c778bb748-8wp8q_calico-system(2b372f30-5f27-467a-8d6a-b220df9d05b0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"db22bded2633f5fc21f45943f81ba16134949c601cf5dcfbba61da88fc615187\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-7c778bb748-8wp8q" podUID="2b372f30-5f27-467a-8d6a-b220df9d05b0" Jan 14 13:53:16.707914 containerd[1635]: time="2026-01-14T13:53:16.706203134Z" level=error msg="Failed to destroy network for sandbox \"d6351bd580ec7b67f5369c81739582cdf3a098da456c892bd52f8f335a838dba\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:16.711398 systemd[1]: run-netns-cni\x2db62251bb\x2d09f6\x2de085\x2d5d50\x2da9c4ae8fc5c5.mount: Deactivated successfully. Jan 14 13:53:16.716684 containerd[1635]: time="2026-01-14T13:53:16.716157438Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-srck5,Uid:4efafa71-2a7d-4e25-8295-f778af674237,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"d6351bd580ec7b67f5369c81739582cdf3a098da456c892bd52f8f335a838dba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:16.718460 kubelet[2951]: E0114 13:53:16.717933 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d6351bd580ec7b67f5369c81739582cdf3a098da456c892bd52f8f335a838dba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:16.719118 kubelet[2951]: E0114 13:53:16.718883 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d6351bd580ec7b67f5369c81739582cdf3a098da456c892bd52f8f335a838dba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-srck5" Jan 14 13:53:16.719118 kubelet[2951]: E0114 13:53:16.719084 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d6351bd580ec7b67f5369c81739582cdf3a098da456c892bd52f8f335a838dba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-srck5" Jan 14 13:53:16.719635 kubelet[2951]: E0114 13:53:16.719476 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-srck5_kube-system(4efafa71-2a7d-4e25-8295-f778af674237)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-srck5_kube-system(4efafa71-2a7d-4e25-8295-f778af674237)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d6351bd580ec7b67f5369c81739582cdf3a098da456c892bd52f8f335a838dba\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-srck5" podUID="4efafa71-2a7d-4e25-8295-f778af674237" Jan 14 13:53:16.736072 containerd[1635]: time="2026-01-14T13:53:16.735896024Z" level=error msg="Failed to destroy network for sandbox \"89233ed2a165b94747b81a10c4c7d92f2c7c14f2d16da9d363e111e9f5406b8b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:16.741494 systemd[1]: run-netns-cni\x2d51fcb0ce\x2d37e2\x2d2241\x2dc03b\x2d239796818f3e.mount: Deactivated successfully. Jan 14 13:53:16.745457 containerd[1635]: time="2026-01-14T13:53:16.744984925Z" level=error msg="Failed to destroy network for sandbox \"29501fc9061e44fc13c90129fbef7ec016aae7af7cfff0bede19581d710c9746\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:16.747834 containerd[1635]: time="2026-01-14T13:53:16.745667020Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6f5bbf559f-5j9gg,Uid:d362a068-c796-4380-9ca2-cf2fbba56f1d,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"89233ed2a165b94747b81a10c4c7d92f2c7c14f2d16da9d363e111e9f5406b8b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:16.747988 kubelet[2951]: E0114 13:53:16.746130 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"89233ed2a165b94747b81a10c4c7d92f2c7c14f2d16da9d363e111e9f5406b8b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:16.747988 kubelet[2951]: E0114 13:53:16.746182 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"89233ed2a165b94747b81a10c4c7d92f2c7c14f2d16da9d363e111e9f5406b8b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-6f5bbf559f-5j9gg" Jan 14 13:53:16.747988 kubelet[2951]: E0114 13:53:16.746201 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"89233ed2a165b94747b81a10c4c7d92f2c7c14f2d16da9d363e111e9f5406b8b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-6f5bbf559f-5j9gg" Jan 14 13:53:16.748083 kubelet[2951]: E0114 13:53:16.746320 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-6f5bbf559f-5j9gg_calico-system(d362a068-c796-4380-9ca2-cf2fbba56f1d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-6f5bbf559f-5j9gg_calico-system(d362a068-c796-4380-9ca2-cf2fbba56f1d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"89233ed2a165b94747b81a10c4c7d92f2c7c14f2d16da9d363e111e9f5406b8b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-6f5bbf559f-5j9gg" podUID="d362a068-c796-4380-9ca2-cf2fbba56f1d" Jan 14 13:53:16.750336 systemd[1]: run-netns-cni\x2d6485e170\x2dd683\x2d9fb6\x2df29b\x2dc75cf99c19b3.mount: Deactivated successfully. Jan 14 13:53:16.755239 containerd[1635]: time="2026-01-14T13:53:16.754971714Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-r9sfd,Uid:e6501533-023e-4ab4-8991-e08f2d55a030,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"29501fc9061e44fc13c90129fbef7ec016aae7af7cfff0bede19581d710c9746\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:16.756750 kubelet[2951]: E0114 13:53:16.755851 2951 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"29501fc9061e44fc13c90129fbef7ec016aae7af7cfff0bede19581d710c9746\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 13:53:16.756750 kubelet[2951]: E0114 13:53:16.755891 2951 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"29501fc9061e44fc13c90129fbef7ec016aae7af7cfff0bede19581d710c9746\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-r9sfd" Jan 14 13:53:16.756750 kubelet[2951]: E0114 13:53:16.755908 2951 kuberuntime_manager.go:1343] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"29501fc9061e44fc13c90129fbef7ec016aae7af7cfff0bede19581d710c9746\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-66bc5c9577-r9sfd" Jan 14 13:53:16.756901 kubelet[2951]: E0114 13:53:16.755948 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-66bc5c9577-r9sfd_kube-system(e6501533-023e-4ab4-8991-e08f2d55a030)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-66bc5c9577-r9sfd_kube-system(e6501533-023e-4ab4-8991-e08f2d55a030)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"29501fc9061e44fc13c90129fbef7ec016aae7af7cfff0bede19581d710c9746\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-66bc5c9577-r9sfd" podUID="e6501533-023e-4ab4-8991-e08f2d55a030" Jan 14 13:53:18.381204 kubelet[2951]: E0114 13:53:18.381059 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:18.433413 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4195497663.mount: Deactivated successfully. Jan 14 13:53:18.487422 containerd[1635]: time="2026-01-14T13:53:18.487066268Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:53:18.489496 containerd[1635]: time="2026-01-14T13:53:18.489152596Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.4: active requests=0, bytes read=156880025" Jan 14 13:53:18.491815 containerd[1635]: time="2026-01-14T13:53:18.491446849Z" level=info msg="ImageCreate event name:\"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:53:18.494737 containerd[1635]: time="2026-01-14T13:53:18.494665109Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 13:53:18.495725 containerd[1635]: time="2026-01-14T13:53:18.495480013Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.4\" with image id \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\", size \"156883537\" in 19.386136581s" Jan 14 13:53:18.495725 containerd[1635]: time="2026-01-14T13:53:18.495659557Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\" returns image reference \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\"" Jan 14 13:53:18.520152 containerd[1635]: time="2026-01-14T13:53:18.520040826Z" level=info msg="CreateContainer within sandbox \"14ebc5df29f97bd4359b3e55957c05fc18c0d3a2b470daa31b428fd78cb16055\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Jan 14 13:53:18.535760 containerd[1635]: time="2026-01-14T13:53:18.535433261Z" level=info msg="Container 04dcb583bb857d58f5ea9f7572acf5987acc33ea989175bbf88e0b0874f8f235: CDI devices from CRI Config.CDIDevices: []" Jan 14 13:53:18.641110 containerd[1635]: time="2026-01-14T13:53:18.640820047Z" level=info msg="CreateContainer within sandbox \"14ebc5df29f97bd4359b3e55957c05fc18c0d3a2b470daa31b428fd78cb16055\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"04dcb583bb857d58f5ea9f7572acf5987acc33ea989175bbf88e0b0874f8f235\"" Jan 14 13:53:18.642889 containerd[1635]: time="2026-01-14T13:53:18.642677527Z" level=info msg="StartContainer for \"04dcb583bb857d58f5ea9f7572acf5987acc33ea989175bbf88e0b0874f8f235\"" Jan 14 13:53:18.644654 containerd[1635]: time="2026-01-14T13:53:18.644429140Z" level=info msg="connecting to shim 04dcb583bb857d58f5ea9f7572acf5987acc33ea989175bbf88e0b0874f8f235" address="unix:///run/containerd/s/aa914e9f8d13f41cdee1957eb5a9836fb7fc6d9f54222f9c02ed706a6d9468f8" protocol=ttrpc version=3 Jan 14 13:53:18.691911 systemd[1]: Started cri-containerd-04dcb583bb857d58f5ea9f7572acf5987acc33ea989175bbf88e0b0874f8f235.scope - libcontainer container 04dcb583bb857d58f5ea9f7572acf5987acc33ea989175bbf88e0b0874f8f235. Jan 14 13:53:18.797000 audit: BPF prog-id=170 op=LOAD Jan 14 13:53:18.797000 audit[4346]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3539 pid=4346 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:18.822377 kernel: audit: type=1334 audit(1768398798.797:565): prog-id=170 op=LOAD Jan 14 13:53:18.822454 kernel: audit: type=1300 audit(1768398798.797:565): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3539 pid=4346 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:18.797000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034646362353833626238353764353866356561396637353732616366 Jan 14 13:53:18.797000 audit: BPF prog-id=171 op=LOAD Jan 14 13:53:18.848020 kernel: audit: type=1327 audit(1768398798.797:565): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034646362353833626238353764353866356561396637353732616366 Jan 14 13:53:18.848234 kernel: audit: type=1334 audit(1768398798.797:566): prog-id=171 op=LOAD Jan 14 13:53:18.848353 kernel: audit: type=1300 audit(1768398798.797:566): arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3539 pid=4346 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:18.797000 audit[4346]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3539 pid=4346 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:18.866701 kernel: audit: type=1327 audit(1768398798.797:566): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034646362353833626238353764353866356561396637353732616366 Jan 14 13:53:18.797000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034646362353833626238353764353866356561396637353732616366 Jan 14 13:53:18.884688 kernel: audit: type=1334 audit(1768398798.797:567): prog-id=171 op=UNLOAD Jan 14 13:53:18.797000 audit: BPF prog-id=171 op=UNLOAD Jan 14 13:53:18.889716 kernel: audit: type=1300 audit(1768398798.797:567): arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3539 pid=4346 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:18.797000 audit[4346]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3539 pid=4346 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:18.797000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034646362353833626238353764353866356561396637353732616366 Jan 14 13:53:18.923496 containerd[1635]: time="2026-01-14T13:53:18.923404171Z" level=info msg="StartContainer for \"04dcb583bb857d58f5ea9f7572acf5987acc33ea989175bbf88e0b0874f8f235\" returns successfully" Jan 14 13:53:18.926395 kernel: audit: type=1327 audit(1768398798.797:567): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034646362353833626238353764353866356561396637353732616366 Jan 14 13:53:18.797000 audit: BPF prog-id=170 op=UNLOAD Jan 14 13:53:18.797000 audit[4346]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3539 pid=4346 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:18.797000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034646362353833626238353764353866356561396637353732616366 Jan 14 13:53:18.797000 audit: BPF prog-id=172 op=LOAD Jan 14 13:53:18.797000 audit[4346]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=3539 pid=4346 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:18.797000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034646362353833626238353764353866356561396637353732616366 Jan 14 13:53:18.932692 kernel: audit: type=1334 audit(1768398798.797:568): prog-id=170 op=UNLOAD Jan 14 13:53:19.089631 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Jan 14 13:53:19.089938 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Jan 14 13:53:19.355422 kubelet[2951]: E0114 13:53:19.355165 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:19.364119 kubelet[2951]: I0114 13:53:19.363408 2951 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/d362a068-c796-4380-9ca2-cf2fbba56f1d-whisker-backend-key-pair\") pod \"d362a068-c796-4380-9ca2-cf2fbba56f1d\" (UID: \"d362a068-c796-4380-9ca2-cf2fbba56f1d\") " Jan 14 13:53:19.364119 kubelet[2951]: I0114 13:53:19.363479 2951 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph7bz\" (UniqueName: \"kubernetes.io/projected/d362a068-c796-4380-9ca2-cf2fbba56f1d-kube-api-access-ph7bz\") pod \"d362a068-c796-4380-9ca2-cf2fbba56f1d\" (UID: \"d362a068-c796-4380-9ca2-cf2fbba56f1d\") " Jan 14 13:53:19.364119 kubelet[2951]: I0114 13:53:19.363650 2951 reconciler_common.go:163] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d362a068-c796-4380-9ca2-cf2fbba56f1d-whisker-ca-bundle\") pod \"d362a068-c796-4380-9ca2-cf2fbba56f1d\" (UID: \"d362a068-c796-4380-9ca2-cf2fbba56f1d\") " Jan 14 13:53:19.366978 kubelet[2951]: I0114 13:53:19.366734 2951 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d362a068-c796-4380-9ca2-cf2fbba56f1d-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "d362a068-c796-4380-9ca2-cf2fbba56f1d" (UID: "d362a068-c796-4380-9ca2-cf2fbba56f1d"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Jan 14 13:53:19.382647 kubelet[2951]: I0114 13:53:19.382423 2951 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d362a068-c796-4380-9ca2-cf2fbba56f1d-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "d362a068-c796-4380-9ca2-cf2fbba56f1d" (UID: "d362a068-c796-4380-9ca2-cf2fbba56f1d"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Jan 14 13:53:19.386255 kubelet[2951]: I0114 13:53:19.385903 2951 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d362a068-c796-4380-9ca2-cf2fbba56f1d-kube-api-access-ph7bz" (OuterVolumeSpecName: "kube-api-access-ph7bz") pod "d362a068-c796-4380-9ca2-cf2fbba56f1d" (UID: "d362a068-c796-4380-9ca2-cf2fbba56f1d"). InnerVolumeSpecName "kube-api-access-ph7bz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Jan 14 13:53:19.404991 kubelet[2951]: I0114 13:53:19.404763 2951 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-mt5qd" podStartSLOduration=3.507815055 podStartE2EDuration="53.404747997s" podCreationTimestamp="2026-01-14 13:52:26 +0000 UTC" firstStartedPulling="2026-01-14 13:52:28.600240707 +0000 UTC m=+55.848465975" lastFinishedPulling="2026-01-14 13:53:18.497173648 +0000 UTC m=+105.745398917" observedRunningTime="2026-01-14 13:53:19.401066419 +0000 UTC m=+106.649291697" watchObservedRunningTime="2026-01-14 13:53:19.404747997 +0000 UTC m=+106.652973266" Jan 14 13:53:19.436198 systemd[1]: var-lib-kubelet-pods-d362a068\x2dc796\x2d4380\x2d9ca2\x2dcf2fbba56f1d-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Jan 14 13:53:19.436639 systemd[1]: var-lib-kubelet-pods-d362a068\x2dc796\x2d4380\x2d9ca2\x2dcf2fbba56f1d-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dph7bz.mount: Deactivated successfully. Jan 14 13:53:19.465458 kubelet[2951]: I0114 13:53:19.465182 2951 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/d362a068-c796-4380-9ca2-cf2fbba56f1d-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Jan 14 13:53:19.465458 kubelet[2951]: I0114 13:53:19.465361 2951 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ph7bz\" (UniqueName: \"kubernetes.io/projected/d362a068-c796-4380-9ca2-cf2fbba56f1d-kube-api-access-ph7bz\") on node \"localhost\" DevicePath \"\"" Jan 14 13:53:19.465458 kubelet[2951]: I0114 13:53:19.465380 2951 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d362a068-c796-4380-9ca2-cf2fbba56f1d-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Jan 14 13:53:19.669346 systemd[1]: Removed slice kubepods-besteffort-podd362a068_c796_4380_9ca2_cf2fbba56f1d.slice - libcontainer container kubepods-besteffort-podd362a068_c796_4380_9ca2_cf2fbba56f1d.slice. Jan 14 13:53:19.783644 systemd[1]: Created slice kubepods-besteffort-pod0b696939_2eb2_437f_b107_24730b99d30f.slice - libcontainer container kubepods-besteffort-pod0b696939_2eb2_437f_b107_24730b99d30f.slice. Jan 14 13:53:19.870198 kubelet[2951]: I0114 13:53:19.869833 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/0b696939-2eb2-437f-b107-24730b99d30f-whisker-backend-key-pair\") pod \"whisker-868d8bc54c-7mv9t\" (UID: \"0b696939-2eb2-437f-b107-24730b99d30f\") " pod="calico-system/whisker-868d8bc54c-7mv9t" Jan 14 13:53:19.870198 kubelet[2951]: I0114 13:53:19.870089 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b696939-2eb2-437f-b107-24730b99d30f-whisker-ca-bundle\") pod \"whisker-868d8bc54c-7mv9t\" (UID: \"0b696939-2eb2-437f-b107-24730b99d30f\") " pod="calico-system/whisker-868d8bc54c-7mv9t" Jan 14 13:53:19.870198 kubelet[2951]: I0114 13:53:19.870109 2951 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh846\" (UniqueName: \"kubernetes.io/projected/0b696939-2eb2-437f-b107-24730b99d30f-kube-api-access-hh846\") pod \"whisker-868d8bc54c-7mv9t\" (UID: \"0b696939-2eb2-437f-b107-24730b99d30f\") " pod="calico-system/whisker-868d8bc54c-7mv9t" Jan 14 13:53:20.093144 containerd[1635]: time="2026-01-14T13:53:20.092770498Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-868d8bc54c-7mv9t,Uid:0b696939-2eb2-437f-b107-24730b99d30f,Namespace:calico-system,Attempt:0,}" Jan 14 13:53:20.356961 kubelet[2951]: E0114 13:53:20.356888 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:20.396862 kubelet[2951]: I0114 13:53:20.396497 2951 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d362a068-c796-4380-9ca2-cf2fbba56f1d" path="/var/lib/kubelet/pods/d362a068-c796-4380-9ca2-cf2fbba56f1d/volumes" Jan 14 13:53:20.461272 systemd-networkd[1527]: cali31dce086f57: Link UP Jan 14 13:53:20.461889 systemd-networkd[1527]: cali31dce086f57: Gained carrier Jan 14 13:53:20.515207 containerd[1635]: 2026-01-14 13:53:20.159 [INFO][4437] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 14 13:53:20.515207 containerd[1635]: 2026-01-14 13:53:20.194 [INFO][4437] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--868d8bc54c--7mv9t-eth0 whisker-868d8bc54c- calico-system 0b696939-2eb2-437f-b107-24730b99d30f 1129 0 2026-01-14 13:53:19 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:868d8bc54c projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-868d8bc54c-7mv9t eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali31dce086f57 [] [] }} ContainerID="050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" Namespace="calico-system" Pod="whisker-868d8bc54c-7mv9t" WorkloadEndpoint="localhost-k8s-whisker--868d8bc54c--7mv9t-" Jan 14 13:53:20.515207 containerd[1635]: 2026-01-14 13:53:20.194 [INFO][4437] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" Namespace="calico-system" Pod="whisker-868d8bc54c-7mv9t" WorkloadEndpoint="localhost-k8s-whisker--868d8bc54c--7mv9t-eth0" Jan 14 13:53:20.515207 containerd[1635]: 2026-01-14 13:53:20.348 [INFO][4452] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" HandleID="k8s-pod-network.050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" Workload="localhost-k8s-whisker--868d8bc54c--7mv9t-eth0" Jan 14 13:53:20.517729 containerd[1635]: 2026-01-14 13:53:20.350 [INFO][4452] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" HandleID="k8s-pod-network.050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" Workload="localhost-k8s-whisker--868d8bc54c--7mv9t-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001a4a40), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-868d8bc54c-7mv9t", "timestamp":"2026-01-14 13:53:20.34896807 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 13:53:20.517729 containerd[1635]: 2026-01-14 13:53:20.350 [INFO][4452] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 13:53:20.517729 containerd[1635]: 2026-01-14 13:53:20.351 [INFO][4452] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 13:53:20.517729 containerd[1635]: 2026-01-14 13:53:20.352 [INFO][4452] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 14 13:53:20.517729 containerd[1635]: 2026-01-14 13:53:20.370 [INFO][4452] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" host="localhost" Jan 14 13:53:20.517729 containerd[1635]: 2026-01-14 13:53:20.385 [INFO][4452] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 14 13:53:20.517729 containerd[1635]: 2026-01-14 13:53:20.396 [INFO][4452] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 14 13:53:20.517729 containerd[1635]: 2026-01-14 13:53:20.399 [INFO][4452] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:20.517729 containerd[1635]: 2026-01-14 13:53:20.403 [INFO][4452] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:20.517729 containerd[1635]: 2026-01-14 13:53:20.403 [INFO][4452] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" host="localhost" Jan 14 13:53:20.518032 containerd[1635]: 2026-01-14 13:53:20.408 [INFO][4452] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd Jan 14 13:53:20.518032 containerd[1635]: 2026-01-14 13:53:20.417 [INFO][4452] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" host="localhost" Jan 14 13:53:20.518032 containerd[1635]: 2026-01-14 13:53:20.431 [INFO][4452] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" host="localhost" Jan 14 13:53:20.518032 containerd[1635]: 2026-01-14 13:53:20.432 [INFO][4452] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" host="localhost" Jan 14 13:53:20.518032 containerd[1635]: 2026-01-14 13:53:20.432 [INFO][4452] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 13:53:20.518032 containerd[1635]: 2026-01-14 13:53:20.432 [INFO][4452] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" HandleID="k8s-pod-network.050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" Workload="localhost-k8s-whisker--868d8bc54c--7mv9t-eth0" Jan 14 13:53:20.518147 containerd[1635]: 2026-01-14 13:53:20.438 [INFO][4437] cni-plugin/k8s.go 418: Populated endpoint ContainerID="050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" Namespace="calico-system" Pod="whisker-868d8bc54c-7mv9t" WorkloadEndpoint="localhost-k8s-whisker--868d8bc54c--7mv9t-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--868d8bc54c--7mv9t-eth0", GenerateName:"whisker-868d8bc54c-", Namespace:"calico-system", SelfLink:"", UID:"0b696939-2eb2-437f-b107-24730b99d30f", ResourceVersion:"1129", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 53, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"868d8bc54c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-868d8bc54c-7mv9t", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali31dce086f57", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:20.518147 containerd[1635]: 2026-01-14 13:53:20.439 [INFO][4437] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" Namespace="calico-system" Pod="whisker-868d8bc54c-7mv9t" WorkloadEndpoint="localhost-k8s-whisker--868d8bc54c--7mv9t-eth0" Jan 14 13:53:20.518358 containerd[1635]: 2026-01-14 13:53:20.439 [INFO][4437] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali31dce086f57 ContainerID="050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" Namespace="calico-system" Pod="whisker-868d8bc54c-7mv9t" WorkloadEndpoint="localhost-k8s-whisker--868d8bc54c--7mv9t-eth0" Jan 14 13:53:20.518358 containerd[1635]: 2026-01-14 13:53:20.463 [INFO][4437] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" Namespace="calico-system" Pod="whisker-868d8bc54c-7mv9t" WorkloadEndpoint="localhost-k8s-whisker--868d8bc54c--7mv9t-eth0" Jan 14 13:53:20.521123 containerd[1635]: 2026-01-14 13:53:20.464 [INFO][4437] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" Namespace="calico-system" Pod="whisker-868d8bc54c-7mv9t" WorkloadEndpoint="localhost-k8s-whisker--868d8bc54c--7mv9t-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--868d8bc54c--7mv9t-eth0", GenerateName:"whisker-868d8bc54c-", Namespace:"calico-system", SelfLink:"", UID:"0b696939-2eb2-437f-b107-24730b99d30f", ResourceVersion:"1129", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 53, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"868d8bc54c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd", Pod:"whisker-868d8bc54c-7mv9t", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali31dce086f57", MAC:"d2:44:86:13:24:81", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:20.523208 containerd[1635]: 2026-01-14 13:53:20.494 [INFO][4437] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" Namespace="calico-system" Pod="whisker-868d8bc54c-7mv9t" WorkloadEndpoint="localhost-k8s-whisker--868d8bc54c--7mv9t-eth0" Jan 14 13:53:20.646974 containerd[1635]: time="2026-01-14T13:53:20.646763208Z" level=info msg="connecting to shim 050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd" address="unix:///run/containerd/s/5e9a5dfb6bf2b0facfc0e3d684c34576875fbab42d4544e9c45ddad7bd0c13c0" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:53:20.744823 systemd[1]: Started cri-containerd-050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd.scope - libcontainer container 050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd. Jan 14 13:53:20.762000 audit: BPF prog-id=173 op=LOAD Jan 14 13:53:20.763000 audit: BPF prog-id=174 op=LOAD Jan 14 13:53:20.763000 audit[4515]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=4504 pid=4515 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:20.763000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035306635386364343832663433663033323832643663613661613262 Jan 14 13:53:20.764000 audit: BPF prog-id=174 op=UNLOAD Jan 14 13:53:20.764000 audit[4515]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4504 pid=4515 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:20.764000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035306635386364343832663433663033323832643663613661613262 Jan 14 13:53:20.764000 audit: BPF prog-id=175 op=LOAD Jan 14 13:53:20.764000 audit[4515]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=4504 pid=4515 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:20.764000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035306635386364343832663433663033323832643663613661613262 Jan 14 13:53:20.764000 audit: BPF prog-id=176 op=LOAD Jan 14 13:53:20.764000 audit[4515]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=4504 pid=4515 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:20.764000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035306635386364343832663433663033323832643663613661613262 Jan 14 13:53:20.764000 audit: BPF prog-id=176 op=UNLOAD Jan 14 13:53:20.764000 audit[4515]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=4504 pid=4515 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:20.764000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035306635386364343832663433663033323832643663613661613262 Jan 14 13:53:20.764000 audit: BPF prog-id=175 op=UNLOAD Jan 14 13:53:20.764000 audit[4515]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4504 pid=4515 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:20.764000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035306635386364343832663433663033323832643663613661613262 Jan 14 13:53:20.764000 audit: BPF prog-id=177 op=LOAD Jan 14 13:53:20.764000 audit[4515]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=4504 pid=4515 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:20.764000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3035306635386364343832663433663033323832643663613661613262 Jan 14 13:53:20.767502 systemd-resolved[1297]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 14 13:53:20.830895 containerd[1635]: time="2026-01-14T13:53:20.830810588Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-868d8bc54c-7mv9t,Uid:0b696939-2eb2-437f-b107-24730b99d30f,Namespace:calico-system,Attempt:0,} returns sandbox id \"050f58cd482f43f03282d6ca6aa2b15d21b7d37a69d4ab71c8bfb9cb014475cd\"" Jan 14 13:53:20.833986 containerd[1635]: time="2026-01-14T13:53:20.833951830Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 14 13:53:20.930093 containerd[1635]: time="2026-01-14T13:53:20.929862324Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:20.932796 containerd[1635]: time="2026-01-14T13:53:20.932709779Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 14 13:53:20.932864 containerd[1635]: time="2026-01-14T13:53:20.932837777Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:20.934358 kubelet[2951]: E0114 13:53:20.934056 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 14 13:53:20.934358 kubelet[2951]: E0114 13:53:20.934166 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 14 13:53:20.935158 kubelet[2951]: E0114 13:53:20.934899 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container whisker start failed in pod whisker-868d8bc54c-7mv9t_calico-system(0b696939-2eb2-437f-b107-24730b99d30f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:20.945724 containerd[1635]: time="2026-01-14T13:53:20.945679056Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 14 13:53:21.025458 containerd[1635]: time="2026-01-14T13:53:21.025407690Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:21.029027 containerd[1635]: time="2026-01-14T13:53:21.028733125Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 14 13:53:21.029027 containerd[1635]: time="2026-01-14T13:53:21.028860389Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:21.030060 kubelet[2951]: E0114 13:53:21.029832 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 14 13:53:21.030060 kubelet[2951]: E0114 13:53:21.029937 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 14 13:53:21.030060 kubelet[2951]: E0114 13:53:21.030015 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container whisker-backend start failed in pod whisker-868d8bc54c-7mv9t_calico-system(0b696939-2eb2-437f-b107-24730b99d30f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:21.030169 kubelet[2951]: E0114 13:53:21.030062 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-868d8bc54c-7mv9t" podUID="0b696939-2eb2-437f-b107-24730b99d30f" Jan 14 13:53:21.181000 audit: BPF prog-id=178 op=LOAD Jan 14 13:53:21.181000 audit[4650]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd02a28c60 a2=98 a3=1fffffffffffffff items=0 ppid=4562 pid=4650 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.181000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 14 13:53:21.181000 audit: BPF prog-id=178 op=UNLOAD Jan 14 13:53:21.181000 audit[4650]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffd02a28c30 a3=0 items=0 ppid=4562 pid=4650 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.181000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 14 13:53:21.181000 audit: BPF prog-id=179 op=LOAD Jan 14 13:53:21.181000 audit[4650]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd02a28b40 a2=94 a3=3 items=0 ppid=4562 pid=4650 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.181000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 14 13:53:21.181000 audit: BPF prog-id=179 op=UNLOAD Jan 14 13:53:21.181000 audit[4650]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffd02a28b40 a2=94 a3=3 items=0 ppid=4562 pid=4650 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.181000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 14 13:53:21.181000 audit: BPF prog-id=180 op=LOAD Jan 14 13:53:21.181000 audit[4650]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd02a28b80 a2=94 a3=7ffd02a28d60 items=0 ppid=4562 pid=4650 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.181000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 14 13:53:21.181000 audit: BPF prog-id=180 op=UNLOAD Jan 14 13:53:21.181000 audit[4650]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffd02a28b80 a2=94 a3=7ffd02a28d60 items=0 ppid=4562 pid=4650 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.181000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 14 13:53:21.198000 audit: BPF prog-id=181 op=LOAD Jan 14 13:53:21.198000 audit[4651]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe4ac60880 a2=98 a3=3 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.198000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.199000 audit: BPF prog-id=181 op=UNLOAD Jan 14 13:53:21.199000 audit[4651]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffe4ac60850 a3=0 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.199000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.200000 audit: BPF prog-id=182 op=LOAD Jan 14 13:53:21.200000 audit[4651]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe4ac60670 a2=94 a3=54428f items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.200000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.200000 audit: BPF prog-id=182 op=UNLOAD Jan 14 13:53:21.200000 audit[4651]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe4ac60670 a2=94 a3=54428f items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.200000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.200000 audit: BPF prog-id=183 op=LOAD Jan 14 13:53:21.200000 audit[4651]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe4ac606a0 a2=94 a3=2 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.200000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.200000 audit: BPF prog-id=183 op=UNLOAD Jan 14 13:53:21.200000 audit[4651]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe4ac606a0 a2=0 a3=2 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.200000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.365754 kubelet[2951]: E0114 13:53:21.365472 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-868d8bc54c-7mv9t" podUID="0b696939-2eb2-437f-b107-24730b99d30f" Jan 14 13:53:21.383699 kubelet[2951]: E0114 13:53:21.382044 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:21.406000 audit[4676]: NETFILTER_CFG table=filter:115 family=2 entries=20 op=nft_register_rule pid=4676 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:21.406000 audit[4676]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffc920e4570 a2=0 a3=7ffc920e455c items=0 ppid=3117 pid=4676 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.406000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:53:21.414000 audit[4676]: NETFILTER_CFG table=nat:116 family=2 entries=14 op=nft_register_rule pid=4676 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:21.414000 audit[4676]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffc920e4570 a2=0 a3=0 items=0 ppid=3117 pid=4676 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.414000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:53:21.465000 audit: BPF prog-id=184 op=LOAD Jan 14 13:53:21.465000 audit[4651]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe4ac60560 a2=94 a3=1 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.465000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.465000 audit: BPF prog-id=184 op=UNLOAD Jan 14 13:53:21.465000 audit[4651]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe4ac60560 a2=94 a3=1 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.465000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.476000 audit: BPF prog-id=185 op=LOAD Jan 14 13:53:21.476000 audit[4651]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe4ac60550 a2=94 a3=4 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.476000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.476000 audit: BPF prog-id=185 op=UNLOAD Jan 14 13:53:21.476000 audit[4651]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffe4ac60550 a2=0 a3=4 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.476000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.476000 audit: BPF prog-id=186 op=LOAD Jan 14 13:53:21.476000 audit[4651]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffe4ac603b0 a2=94 a3=5 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.476000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.477000 audit: BPF prog-id=186 op=UNLOAD Jan 14 13:53:21.477000 audit[4651]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffe4ac603b0 a2=0 a3=5 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.477000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.477000 audit: BPF prog-id=187 op=LOAD Jan 14 13:53:21.477000 audit[4651]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe4ac605d0 a2=94 a3=6 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.477000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.477000 audit: BPF prog-id=187 op=UNLOAD Jan 14 13:53:21.477000 audit[4651]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffe4ac605d0 a2=0 a3=6 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.477000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.477000 audit: BPF prog-id=188 op=LOAD Jan 14 13:53:21.477000 audit[4651]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe4ac5fd80 a2=94 a3=88 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.477000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.477000 audit: BPF prog-id=189 op=LOAD Jan 14 13:53:21.477000 audit[4651]: SYSCALL arch=c000003e syscall=321 success=yes exit=7 a0=5 a1=7ffe4ac5fc00 a2=94 a3=2 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.477000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.477000 audit: BPF prog-id=189 op=UNLOAD Jan 14 13:53:21.477000 audit[4651]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=7 a1=7ffe4ac5fc30 a2=0 a3=7ffe4ac5fd30 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.477000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.478000 audit: BPF prog-id=188 op=UNLOAD Jan 14 13:53:21.478000 audit[4651]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=3f618d10 a2=0 a3=5139b65dba1584a5 items=0 ppid=4562 pid=4651 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.478000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 13:53:21.499000 audit: BPF prog-id=190 op=LOAD Jan 14 13:53:21.499000 audit[4679]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe30cf8330 a2=98 a3=1999999999999999 items=0 ppid=4562 pid=4679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.499000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 14 13:53:21.499000 audit: BPF prog-id=190 op=UNLOAD Jan 14 13:53:21.499000 audit[4679]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffe30cf8300 a3=0 items=0 ppid=4562 pid=4679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.499000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 14 13:53:21.499000 audit: BPF prog-id=191 op=LOAD Jan 14 13:53:21.499000 audit[4679]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe30cf8210 a2=94 a3=ffff items=0 ppid=4562 pid=4679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.499000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 14 13:53:21.499000 audit: BPF prog-id=191 op=UNLOAD Jan 14 13:53:21.499000 audit[4679]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffe30cf8210 a2=94 a3=ffff items=0 ppid=4562 pid=4679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.499000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 14 13:53:21.499000 audit: BPF prog-id=192 op=LOAD Jan 14 13:53:21.499000 audit[4679]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe30cf8250 a2=94 a3=7ffe30cf8430 items=0 ppid=4562 pid=4679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.499000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 14 13:53:21.499000 audit: BPF prog-id=192 op=UNLOAD Jan 14 13:53:21.499000 audit[4679]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffe30cf8250 a2=94 a3=7ffe30cf8430 items=0 ppid=4562 pid=4679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.499000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 14 13:53:21.660504 systemd-networkd[1527]: vxlan.calico: Link UP Jan 14 13:53:21.660775 systemd-networkd[1527]: vxlan.calico: Gained carrier Jan 14 13:53:21.702000 audit: BPF prog-id=193 op=LOAD Jan 14 13:53:21.702000 audit[4704]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffcea14cb20 a2=98 a3=0 items=0 ppid=4562 pid=4704 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.702000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 13:53:21.702000 audit: BPF prog-id=193 op=UNLOAD Jan 14 13:53:21.702000 audit[4704]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffcea14caf0 a3=0 items=0 ppid=4562 pid=4704 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.702000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 13:53:21.702000 audit: BPF prog-id=194 op=LOAD Jan 14 13:53:21.702000 audit[4704]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffcea14c930 a2=94 a3=54428f items=0 ppid=4562 pid=4704 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.702000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 13:53:21.702000 audit: BPF prog-id=194 op=UNLOAD Jan 14 13:53:21.702000 audit[4704]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffcea14c930 a2=94 a3=54428f items=0 ppid=4562 pid=4704 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.702000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 13:53:21.702000 audit: BPF prog-id=195 op=LOAD Jan 14 13:53:21.702000 audit[4704]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffcea14c960 a2=94 a3=2 items=0 ppid=4562 pid=4704 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.702000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 13:53:21.702000 audit: BPF prog-id=195 op=UNLOAD Jan 14 13:53:21.702000 audit[4704]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffcea14c960 a2=0 a3=2 items=0 ppid=4562 pid=4704 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.702000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 13:53:21.702000 audit: BPF prog-id=196 op=LOAD Jan 14 13:53:21.702000 audit[4704]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffcea14c710 a2=94 a3=4 items=0 ppid=4562 pid=4704 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.702000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 13:53:21.702000 audit: BPF prog-id=196 op=UNLOAD Jan 14 13:53:21.702000 audit[4704]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffcea14c710 a2=94 a3=4 items=0 ppid=4562 pid=4704 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.702000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 13:53:21.702000 audit: BPF prog-id=197 op=LOAD Jan 14 13:53:21.702000 audit[4704]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffcea14c810 a2=94 a3=7ffcea14c990 items=0 ppid=4562 pid=4704 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.702000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 13:53:21.702000 audit: BPF prog-id=197 op=UNLOAD Jan 14 13:53:21.702000 audit[4704]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffcea14c810 a2=0 a3=7ffcea14c990 items=0 ppid=4562 pid=4704 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.702000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 13:53:21.704000 audit: BPF prog-id=198 op=LOAD Jan 14 13:53:21.704000 audit[4704]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffcea14bf40 a2=94 a3=2 items=0 ppid=4562 pid=4704 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.704000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 13:53:21.704000 audit: BPF prog-id=198 op=UNLOAD Jan 14 13:53:21.704000 audit[4704]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffcea14bf40 a2=0 a3=2 items=0 ppid=4562 pid=4704 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.704000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 13:53:21.704000 audit: BPF prog-id=199 op=LOAD Jan 14 13:53:21.704000 audit[4704]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffcea14c040 a2=94 a3=30 items=0 ppid=4562 pid=4704 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.704000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 13:53:21.719000 audit: BPF prog-id=200 op=LOAD Jan 14 13:53:21.719000 audit[4711]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7fffff202ab0 a2=98 a3=0 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.719000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.719000 audit: BPF prog-id=200 op=UNLOAD Jan 14 13:53:21.719000 audit[4711]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7fffff202a80 a3=0 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.719000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.719000 audit: BPF prog-id=201 op=LOAD Jan 14 13:53:21.719000 audit[4711]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fffff2028a0 a2=94 a3=54428f items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.719000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.719000 audit: BPF prog-id=201 op=UNLOAD Jan 14 13:53:21.719000 audit[4711]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7fffff2028a0 a2=94 a3=54428f items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.719000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.719000 audit: BPF prog-id=202 op=LOAD Jan 14 13:53:21.719000 audit[4711]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fffff2028d0 a2=94 a3=2 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.719000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.719000 audit: BPF prog-id=202 op=UNLOAD Jan 14 13:53:21.719000 audit[4711]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7fffff2028d0 a2=0 a3=2 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.719000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.923000 audit: BPF prog-id=203 op=LOAD Jan 14 13:53:21.923000 audit[4711]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fffff202790 a2=94 a3=1 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.923000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.923000 audit: BPF prog-id=203 op=UNLOAD Jan 14 13:53:21.923000 audit[4711]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7fffff202790 a2=94 a3=1 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.923000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.933000 audit: BPF prog-id=204 op=LOAD Jan 14 13:53:21.933000 audit[4711]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7fffff202780 a2=94 a3=4 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.933000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.933000 audit: BPF prog-id=204 op=UNLOAD Jan 14 13:53:21.933000 audit[4711]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7fffff202780 a2=0 a3=4 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.933000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.933000 audit: BPF prog-id=205 op=LOAD Jan 14 13:53:21.933000 audit[4711]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fffff2025e0 a2=94 a3=5 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.933000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.934000 audit: BPF prog-id=205 op=UNLOAD Jan 14 13:53:21.934000 audit[4711]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7fffff2025e0 a2=0 a3=5 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.934000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.934000 audit: BPF prog-id=206 op=LOAD Jan 14 13:53:21.934000 audit[4711]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7fffff202800 a2=94 a3=6 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.934000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.934000 audit: BPF prog-id=206 op=UNLOAD Jan 14 13:53:21.934000 audit[4711]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7fffff202800 a2=0 a3=6 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.934000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.934000 audit: BPF prog-id=207 op=LOAD Jan 14 13:53:21.934000 audit[4711]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7fffff201fb0 a2=94 a3=88 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.934000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.934000 audit: BPF prog-id=208 op=LOAD Jan 14 13:53:21.934000 audit[4711]: SYSCALL arch=c000003e syscall=321 success=yes exit=7 a0=5 a1=7fffff201e30 a2=94 a3=2 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.934000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.934000 audit: BPF prog-id=208 op=UNLOAD Jan 14 13:53:21.934000 audit[4711]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=7 a1=7fffff201e60 a2=0 a3=7fffff201f60 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.934000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.935000 audit: BPF prog-id=207 op=UNLOAD Jan 14 13:53:21.935000 audit[4711]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=8e44d10 a2=0 a3=64cec89f6c39d7e5 items=0 ppid=4562 pid=4711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.935000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 13:53:21.949000 audit: BPF prog-id=199 op=UNLOAD Jan 14 13:53:21.949000 audit[4562]: SYSCALL arch=c000003e syscall=263 success=yes exit=0 a0=ffffffffffffff9c a1=c000ef8b80 a2=0 a3=0 items=0 ppid=4548 pid=4562 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="calico-node" exe="/usr/bin/calico-node" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:21.949000 audit: PROCTITLE proctitle=63616C69636F2D6E6F6465002D66656C6978 Jan 14 13:53:22.051000 audit[4740]: NETFILTER_CFG table=mangle:117 family=2 entries=16 op=nft_register_chain pid=4740 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 13:53:22.051000 audit[4740]: SYSCALL arch=c000003e syscall=46 success=yes exit=6868 a0=3 a1=7ffe21b88830 a2=0 a3=7ffe21b8881c items=0 ppid=4562 pid=4740 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:22.051000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 13:53:22.054000 audit[4733]: NETFILTER_CFG table=raw:118 family=2 entries=21 op=nft_register_chain pid=4733 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 13:53:22.054000 audit[4733]: SYSCALL arch=c000003e syscall=46 success=yes exit=8452 a0=3 a1=7ffd2b7af660 a2=0 a3=7ffd2b7af64c items=0 ppid=4562 pid=4733 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:22.054000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 13:53:22.057000 audit[4739]: NETFILTER_CFG table=nat:119 family=2 entries=15 op=nft_register_chain pid=4739 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 13:53:22.057000 audit[4739]: SYSCALL arch=c000003e syscall=46 success=yes exit=5084 a0=3 a1=7fff13ddf890 a2=0 a3=7fff13ddf87c items=0 ppid=4562 pid=4739 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:22.057000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 13:53:22.083000 audit[4743]: NETFILTER_CFG table=filter:120 family=2 entries=94 op=nft_register_chain pid=4743 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 13:53:22.083000 audit[4743]: SYSCALL arch=c000003e syscall=46 success=yes exit=53116 a0=3 a1=7ffecba44340 a2=0 a3=7ffecba4432c items=0 ppid=4562 pid=4743 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:22.083000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 13:53:22.202965 systemd-networkd[1527]: cali31dce086f57: Gained IPv6LL Jan 14 13:53:22.370612 kubelet[2951]: E0114 13:53:22.370258 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-868d8bc54c-7mv9t" podUID="0b696939-2eb2-437f-b107-24730b99d30f" Jan 14 13:53:22.714861 systemd-networkd[1527]: vxlan.calico: Gained IPv6LL Jan 14 13:53:26.389672 containerd[1635]: time="2026-01-14T13:53:26.389414346Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-pvxlq,Uid:967b7513-2769-43e5-a113-9344f17909c3,Namespace:calico-system,Attempt:0,}" Jan 14 13:53:26.631011 systemd-networkd[1527]: cali04ae1750e31: Link UP Jan 14 13:53:26.632740 systemd-networkd[1527]: cali04ae1750e31: Gained carrier Jan 14 13:53:26.661953 containerd[1635]: 2026-01-14 13:53:26.517 [INFO][4753] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--pvxlq-eth0 csi-node-driver- calico-system 967b7513-2769-43e5-a113-9344f17909c3 857 0 2026-01-14 13:52:26 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:9d99788f7 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-pvxlq eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali04ae1750e31 [] [] }} ContainerID="042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" Namespace="calico-system" Pod="csi-node-driver-pvxlq" WorkloadEndpoint="localhost-k8s-csi--node--driver--pvxlq-" Jan 14 13:53:26.661953 containerd[1635]: 2026-01-14 13:53:26.517 [INFO][4753] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" Namespace="calico-system" Pod="csi-node-driver-pvxlq" WorkloadEndpoint="localhost-k8s-csi--node--driver--pvxlq-eth0" Jan 14 13:53:26.661953 containerd[1635]: 2026-01-14 13:53:26.565 [INFO][4766] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" HandleID="k8s-pod-network.042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" Workload="localhost-k8s-csi--node--driver--pvxlq-eth0" Jan 14 13:53:26.662241 containerd[1635]: 2026-01-14 13:53:26.566 [INFO][4766] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" HandleID="k8s-pod-network.042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" Workload="localhost-k8s-csi--node--driver--pvxlq-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001a4da0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-pvxlq", "timestamp":"2026-01-14 13:53:26.565915153 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 13:53:26.662241 containerd[1635]: 2026-01-14 13:53:26.566 [INFO][4766] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 13:53:26.662241 containerd[1635]: 2026-01-14 13:53:26.566 [INFO][4766] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 13:53:26.662241 containerd[1635]: 2026-01-14 13:53:26.566 [INFO][4766] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 14 13:53:26.662241 containerd[1635]: 2026-01-14 13:53:26.575 [INFO][4766] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" host="localhost" Jan 14 13:53:26.662241 containerd[1635]: 2026-01-14 13:53:26.584 [INFO][4766] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 14 13:53:26.662241 containerd[1635]: 2026-01-14 13:53:26.591 [INFO][4766] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 14 13:53:26.662241 containerd[1635]: 2026-01-14 13:53:26.594 [INFO][4766] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:26.662241 containerd[1635]: 2026-01-14 13:53:26.599 [INFO][4766] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:26.662241 containerd[1635]: 2026-01-14 13:53:26.599 [INFO][4766] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" host="localhost" Jan 14 13:53:26.663863 containerd[1635]: 2026-01-14 13:53:26.602 [INFO][4766] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d Jan 14 13:53:26.663863 containerd[1635]: 2026-01-14 13:53:26.608 [INFO][4766] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" host="localhost" Jan 14 13:53:26.663863 containerd[1635]: 2026-01-14 13:53:26.619 [INFO][4766] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" host="localhost" Jan 14 13:53:26.663863 containerd[1635]: 2026-01-14 13:53:26.619 [INFO][4766] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" host="localhost" Jan 14 13:53:26.663863 containerd[1635]: 2026-01-14 13:53:26.619 [INFO][4766] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 13:53:26.663863 containerd[1635]: 2026-01-14 13:53:26.619 [INFO][4766] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" HandleID="k8s-pod-network.042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" Workload="localhost-k8s-csi--node--driver--pvxlq-eth0" Jan 14 13:53:26.663979 containerd[1635]: 2026-01-14 13:53:26.623 [INFO][4753] cni-plugin/k8s.go 418: Populated endpoint ContainerID="042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" Namespace="calico-system" Pod="csi-node-driver-pvxlq" WorkloadEndpoint="localhost-k8s-csi--node--driver--pvxlq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--pvxlq-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"967b7513-2769-43e5-a113-9344f17909c3", ResourceVersion:"857", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 52, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"9d99788f7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-pvxlq", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali04ae1750e31", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:26.664152 containerd[1635]: 2026-01-14 13:53:26.623 [INFO][4753] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" Namespace="calico-system" Pod="csi-node-driver-pvxlq" WorkloadEndpoint="localhost-k8s-csi--node--driver--pvxlq-eth0" Jan 14 13:53:26.664152 containerd[1635]: 2026-01-14 13:53:26.623 [INFO][4753] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali04ae1750e31 ContainerID="042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" Namespace="calico-system" Pod="csi-node-driver-pvxlq" WorkloadEndpoint="localhost-k8s-csi--node--driver--pvxlq-eth0" Jan 14 13:53:26.664152 containerd[1635]: 2026-01-14 13:53:26.632 [INFO][4753] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" Namespace="calico-system" Pod="csi-node-driver-pvxlq" WorkloadEndpoint="localhost-k8s-csi--node--driver--pvxlq-eth0" Jan 14 13:53:26.664216 containerd[1635]: 2026-01-14 13:53:26.634 [INFO][4753] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" Namespace="calico-system" Pod="csi-node-driver-pvxlq" WorkloadEndpoint="localhost-k8s-csi--node--driver--pvxlq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--pvxlq-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"967b7513-2769-43e5-a113-9344f17909c3", ResourceVersion:"857", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 52, 26, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"9d99788f7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d", Pod:"csi-node-driver-pvxlq", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali04ae1750e31", MAC:"e2:6d:33:d7:7b:6a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:26.664455 containerd[1635]: 2026-01-14 13:53:26.652 [INFO][4753] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" Namespace="calico-system" Pod="csi-node-driver-pvxlq" WorkloadEndpoint="localhost-k8s-csi--node--driver--pvxlq-eth0" Jan 14 13:53:26.689000 audit[4786]: NETFILTER_CFG table=filter:121 family=2 entries=36 op=nft_register_chain pid=4786 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 13:53:26.714888 kernel: kauditd_printk_skb: 231 callbacks suppressed Jan 14 13:53:26.715011 kernel: audit: type=1325 audit(1768398806.689:646): table=filter:121 family=2 entries=36 op=nft_register_chain pid=4786 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 13:53:26.689000 audit[4786]: SYSCALL arch=c000003e syscall=46 success=yes exit=19576 a0=3 a1=7fff4c0679a0 a2=0 a3=7fff4c06798c items=0 ppid=4562 pid=4786 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:26.721698 containerd[1635]: time="2026-01-14T13:53:26.721471319Z" level=info msg="connecting to shim 042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d" address="unix:///run/containerd/s/2fff12a14d36d6f3868708dd92cc8044641319445dd353354982c9cf771fbb6a" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:53:26.754647 kernel: audit: type=1300 audit(1768398806.689:646): arch=c000003e syscall=46 success=yes exit=19576 a0=3 a1=7fff4c0679a0 a2=0 a3=7fff4c06798c items=0 ppid=4562 pid=4786 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:26.754722 kernel: audit: type=1327 audit(1768398806.689:646): proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 13:53:26.689000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 13:53:26.806836 systemd[1]: Started cri-containerd-042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d.scope - libcontainer container 042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d. Jan 14 13:53:26.834000 audit: BPF prog-id=209 op=LOAD Jan 14 13:53:26.835000 audit: BPF prog-id=210 op=LOAD Jan 14 13:53:26.844854 systemd-resolved[1297]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 14 13:53:26.852088 kernel: audit: type=1334 audit(1768398806.834:647): prog-id=209 op=LOAD Jan 14 13:53:26.852167 kernel: audit: type=1334 audit(1768398806.835:648): prog-id=210 op=LOAD Jan 14 13:53:26.852201 kernel: audit: type=1300 audit(1768398806.835:648): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b0238 a2=98 a3=0 items=0 ppid=4795 pid=4806 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:26.835000 audit[4806]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b0238 a2=98 a3=0 items=0 ppid=4795 pid=4806 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:26.835000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034326562366233383164613362393866303037383430333339653139 Jan 14 13:53:26.894906 kernel: audit: type=1327 audit(1768398806.835:648): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034326562366233383164613362393866303037383430333339653139 Jan 14 13:53:26.835000 audit: BPF prog-id=210 op=UNLOAD Jan 14 13:53:26.901126 kernel: audit: type=1334 audit(1768398806.835:649): prog-id=210 op=UNLOAD Jan 14 13:53:26.835000 audit[4806]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4795 pid=4806 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:26.835000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034326562366233383164613362393866303037383430333339653139 Jan 14 13:53:26.949308 kernel: audit: type=1300 audit(1768398806.835:649): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4795 pid=4806 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:26.949664 kernel: audit: type=1327 audit(1768398806.835:649): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034326562366233383164613362393866303037383430333339653139 Jan 14 13:53:26.835000 audit: BPF prog-id=211 op=LOAD Jan 14 13:53:26.835000 audit[4806]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b0488 a2=98 a3=0 items=0 ppid=4795 pid=4806 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:26.835000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034326562366233383164613362393866303037383430333339653139 Jan 14 13:53:26.835000 audit: BPF prog-id=212 op=LOAD Jan 14 13:53:26.835000 audit[4806]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001b0218 a2=98 a3=0 items=0 ppid=4795 pid=4806 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:26.835000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034326562366233383164613362393866303037383430333339653139 Jan 14 13:53:26.835000 audit: BPF prog-id=212 op=UNLOAD Jan 14 13:53:26.835000 audit[4806]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4795 pid=4806 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:26.835000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034326562366233383164613362393866303037383430333339653139 Jan 14 13:53:26.835000 audit: BPF prog-id=211 op=UNLOAD Jan 14 13:53:26.835000 audit[4806]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4795 pid=4806 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:26.835000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034326562366233383164613362393866303037383430333339653139 Jan 14 13:53:26.835000 audit: BPF prog-id=213 op=LOAD Jan 14 13:53:26.835000 audit[4806]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b06e8 a2=98 a3=0 items=0 ppid=4795 pid=4806 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:26.835000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034326562366233383164613362393866303037383430333339653139 Jan 14 13:53:26.961203 containerd[1635]: time="2026-01-14T13:53:26.960879827Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-pvxlq,Uid:967b7513-2769-43e5-a113-9344f17909c3,Namespace:calico-system,Attempt:0,} returns sandbox id \"042eb6b381da3b98f007840339e19dc835afbc0b15ca20d2dbadff54047abb0d\"" Jan 14 13:53:26.965470 containerd[1635]: time="2026-01-14T13:53:26.965451916Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 14 13:53:27.033122 containerd[1635]: time="2026-01-14T13:53:27.032710675Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:27.035394 containerd[1635]: time="2026-01-14T13:53:27.035066720Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 14 13:53:27.035394 containerd[1635]: time="2026-01-14T13:53:27.035132393Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:27.035831 kubelet[2951]: E0114 13:53:27.035796 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 14 13:53:27.036761 kubelet[2951]: E0114 13:53:27.036175 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 14 13:53:27.036761 kubelet[2951]: E0114 13:53:27.036248 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-csi start failed in pod csi-node-driver-pvxlq_calico-system(967b7513-2769-43e5-a113-9344f17909c3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:27.038153 containerd[1635]: time="2026-01-14T13:53:27.037916828Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 14 13:53:27.113145 containerd[1635]: time="2026-01-14T13:53:27.113021327Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:27.115468 containerd[1635]: time="2026-01-14T13:53:27.115262607Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 14 13:53:27.115468 containerd[1635]: time="2026-01-14T13:53:27.115460616Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:27.115833 kubelet[2951]: E0114 13:53:27.115724 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 14 13:53:27.115833 kubelet[2951]: E0114 13:53:27.115831 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 14 13:53:27.115913 kubelet[2951]: E0114 13:53:27.115894 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container csi-node-driver-registrar start failed in pod csi-node-driver-pvxlq_calico-system(967b7513-2769-43e5-a113-9344f17909c3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:27.116060 kubelet[2951]: E0114 13:53:27.115931 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:53:27.385229 kubelet[2951]: E0114 13:53:27.385117 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:27.386654 containerd[1635]: time="2026-01-14T13:53:27.386408271Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-r9sfd,Uid:e6501533-023e-4ab4-8991-e08f2d55a030,Namespace:kube-system,Attempt:0,}" Jan 14 13:53:27.405254 kubelet[2951]: E0114 13:53:27.405214 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:53:27.685762 systemd-networkd[1527]: cali10ddc5ace98: Link UP Jan 14 13:53:27.686819 systemd-networkd[1527]: cali10ddc5ace98: Gained carrier Jan 14 13:53:27.715644 containerd[1635]: 2026-01-14 13:53:27.489 [INFO][4838] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--66bc5c9577--r9sfd-eth0 coredns-66bc5c9577- kube-system e6501533-023e-4ab4-8991-e08f2d55a030 1015 0 2026-01-14 13:51:35 +0000 UTC map[k8s-app:kube-dns pod-template-hash:66bc5c9577 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-66bc5c9577-r9sfd eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali10ddc5ace98 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" Namespace="kube-system" Pod="coredns-66bc5c9577-r9sfd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--r9sfd-" Jan 14 13:53:27.715644 containerd[1635]: 2026-01-14 13:53:27.490 [INFO][4838] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" Namespace="kube-system" Pod="coredns-66bc5c9577-r9sfd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--r9sfd-eth0" Jan 14 13:53:27.715644 containerd[1635]: 2026-01-14 13:53:27.595 [INFO][4852] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" HandleID="k8s-pod-network.e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" Workload="localhost-k8s-coredns--66bc5c9577--r9sfd-eth0" Jan 14 13:53:27.716292 containerd[1635]: 2026-01-14 13:53:27.596 [INFO][4852] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" HandleID="k8s-pod-network.e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" Workload="localhost-k8s-coredns--66bc5c9577--r9sfd-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001a3720), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-66bc5c9577-r9sfd", "timestamp":"2026-01-14 13:53:27.595139688 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 13:53:27.716292 containerd[1635]: 2026-01-14 13:53:27.596 [INFO][4852] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 13:53:27.716292 containerd[1635]: 2026-01-14 13:53:27.596 [INFO][4852] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 13:53:27.716292 containerd[1635]: 2026-01-14 13:53:27.596 [INFO][4852] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 14 13:53:27.716292 containerd[1635]: 2026-01-14 13:53:27.609 [INFO][4852] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" host="localhost" Jan 14 13:53:27.716292 containerd[1635]: 2026-01-14 13:53:27.618 [INFO][4852] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 14 13:53:27.716292 containerd[1635]: 2026-01-14 13:53:27.631 [INFO][4852] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 14 13:53:27.716292 containerd[1635]: 2026-01-14 13:53:27.635 [INFO][4852] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:27.716292 containerd[1635]: 2026-01-14 13:53:27.645 [INFO][4852] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:27.716292 containerd[1635]: 2026-01-14 13:53:27.646 [INFO][4852] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" host="localhost" Jan 14 13:53:27.716989 containerd[1635]: 2026-01-14 13:53:27.650 [INFO][4852] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38 Jan 14 13:53:27.716989 containerd[1635]: 2026-01-14 13:53:27.663 [INFO][4852] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" host="localhost" Jan 14 13:53:27.716989 containerd[1635]: 2026-01-14 13:53:27.673 [INFO][4852] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" host="localhost" Jan 14 13:53:27.716989 containerd[1635]: 2026-01-14 13:53:27.674 [INFO][4852] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" host="localhost" Jan 14 13:53:27.716989 containerd[1635]: 2026-01-14 13:53:27.674 [INFO][4852] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 13:53:27.716989 containerd[1635]: 2026-01-14 13:53:27.674 [INFO][4852] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" HandleID="k8s-pod-network.e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" Workload="localhost-k8s-coredns--66bc5c9577--r9sfd-eth0" Jan 14 13:53:27.717412 containerd[1635]: 2026-01-14 13:53:27.679 [INFO][4838] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" Namespace="kube-system" Pod="coredns-66bc5c9577-r9sfd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--r9sfd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--66bc5c9577--r9sfd-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"e6501533-023e-4ab4-8991-e08f2d55a030", ResourceVersion:"1015", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 51, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-66bc5c9577-r9sfd", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali10ddc5ace98", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:27.717412 containerd[1635]: 2026-01-14 13:53:27.679 [INFO][4838] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" Namespace="kube-system" Pod="coredns-66bc5c9577-r9sfd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--r9sfd-eth0" Jan 14 13:53:27.717412 containerd[1635]: 2026-01-14 13:53:27.679 [INFO][4838] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali10ddc5ace98 ContainerID="e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" Namespace="kube-system" Pod="coredns-66bc5c9577-r9sfd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--r9sfd-eth0" Jan 14 13:53:27.717412 containerd[1635]: 2026-01-14 13:53:27.688 [INFO][4838] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" Namespace="kube-system" Pod="coredns-66bc5c9577-r9sfd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--r9sfd-eth0" Jan 14 13:53:27.717412 containerd[1635]: 2026-01-14 13:53:27.689 [INFO][4838] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" Namespace="kube-system" Pod="coredns-66bc5c9577-r9sfd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--r9sfd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--66bc5c9577--r9sfd-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"e6501533-023e-4ab4-8991-e08f2d55a030", ResourceVersion:"1015", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 51, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38", Pod:"coredns-66bc5c9577-r9sfd", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali10ddc5ace98", MAC:"72:d7:52:a3:0a:ba", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:27.717412 containerd[1635]: 2026-01-14 13:53:27.709 [INFO][4838] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" Namespace="kube-system" Pod="coredns-66bc5c9577-r9sfd" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--r9sfd-eth0" Jan 14 13:53:27.764000 audit[4871]: NETFILTER_CFG table=filter:122 family=2 entries=46 op=nft_register_chain pid=4871 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 13:53:27.764000 audit[4871]: SYSCALL arch=c000003e syscall=46 success=yes exit=23740 a0=3 a1=7ffdc8b6d690 a2=0 a3=7ffdc8b6d67c items=0 ppid=4562 pid=4871 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:27.764000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 13:53:27.792823 containerd[1635]: time="2026-01-14T13:53:27.792706941Z" level=info msg="connecting to shim e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38" address="unix:///run/containerd/s/fc393a093e68a5b60e3afe7dbc748a1950755d7244ce5de84bd2180e2a5d2b28" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:53:27.860023 systemd[1]: Started cri-containerd-e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38.scope - libcontainer container e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38. Jan 14 13:53:27.884000 audit: BPF prog-id=214 op=LOAD Jan 14 13:53:27.886000 audit: BPF prog-id=215 op=LOAD Jan 14 13:53:27.886000 audit[4893]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=4881 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:27.886000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6536356630653131656637343038353538653564396635396130363337 Jan 14 13:53:27.887000 audit: BPF prog-id=215 op=UNLOAD Jan 14 13:53:27.887000 audit[4893]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4881 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:27.887000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6536356630653131656637343038353538653564396635396130363337 Jan 14 13:53:27.887000 audit: BPF prog-id=216 op=LOAD Jan 14 13:53:27.887000 audit[4893]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=4881 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:27.887000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6536356630653131656637343038353538653564396635396130363337 Jan 14 13:53:27.887000 audit: BPF prog-id=217 op=LOAD Jan 14 13:53:27.887000 audit[4893]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=4881 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:27.887000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6536356630653131656637343038353538653564396635396130363337 Jan 14 13:53:27.888000 audit: BPF prog-id=217 op=UNLOAD Jan 14 13:53:27.888000 audit[4893]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=4881 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:27.888000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6536356630653131656637343038353538653564396635396130363337 Jan 14 13:53:27.888000 audit: BPF prog-id=216 op=UNLOAD Jan 14 13:53:27.888000 audit[4893]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4881 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:27.888000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6536356630653131656637343038353538653564396635396130363337 Jan 14 13:53:27.888000 audit: BPF prog-id=218 op=LOAD Jan 14 13:53:27.888000 audit[4893]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=4881 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:27.888000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6536356630653131656637343038353538653564396635396130363337 Jan 14 13:53:27.895499 systemd-resolved[1297]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 14 13:53:27.992003 containerd[1635]: time="2026-01-14T13:53:27.991763003Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-r9sfd,Uid:e6501533-023e-4ab4-8991-e08f2d55a030,Namespace:kube-system,Attempt:0,} returns sandbox id \"e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38\"" Jan 14 13:53:28.000009 kubelet[2951]: E0114 13:53:27.999902 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:28.017676 containerd[1635]: time="2026-01-14T13:53:28.017407666Z" level=info msg="CreateContainer within sandbox \"e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 14 13:53:28.070915 containerd[1635]: time="2026-01-14T13:53:28.068286232Z" level=info msg="Container 378de221c0188308e4bdeba7ca3b264d82fed1193189438312d05c523b2fb94e: CDI devices from CRI Config.CDIDevices: []" Jan 14 13:53:28.074865 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount51567563.mount: Deactivated successfully. Jan 14 13:53:28.082672 containerd[1635]: time="2026-01-14T13:53:28.082198450Z" level=info msg="CreateContainer within sandbox \"e65f0e11ef7408558e5d9f59a063777baa71e4d070369f368502e30b47009a38\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"378de221c0188308e4bdeba7ca3b264d82fed1193189438312d05c523b2fb94e\"" Jan 14 13:53:28.084689 containerd[1635]: time="2026-01-14T13:53:28.084310568Z" level=info msg="StartContainer for \"378de221c0188308e4bdeba7ca3b264d82fed1193189438312d05c523b2fb94e\"" Jan 14 13:53:28.087175 containerd[1635]: time="2026-01-14T13:53:28.087134378Z" level=info msg="connecting to shim 378de221c0188308e4bdeba7ca3b264d82fed1193189438312d05c523b2fb94e" address="unix:///run/containerd/s/fc393a093e68a5b60e3afe7dbc748a1950755d7244ce5de84bd2180e2a5d2b28" protocol=ttrpc version=3 Jan 14 13:53:28.148928 systemd[1]: Started cri-containerd-378de221c0188308e4bdeba7ca3b264d82fed1193189438312d05c523b2fb94e.scope - libcontainer container 378de221c0188308e4bdeba7ca3b264d82fed1193189438312d05c523b2fb94e. Jan 14 13:53:28.156909 systemd-networkd[1527]: cali04ae1750e31: Gained IPv6LL Jan 14 13:53:28.191000 audit: BPF prog-id=219 op=LOAD Jan 14 13:53:28.193000 audit: BPF prog-id=220 op=LOAD Jan 14 13:53:28.193000 audit[4918]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=4881 pid=4918 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.193000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3337386465323231633031383833303865346264656261376361336232 Jan 14 13:53:28.193000 audit: BPF prog-id=220 op=UNLOAD Jan 14 13:53:28.193000 audit[4918]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4881 pid=4918 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.193000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3337386465323231633031383833303865346264656261376361336232 Jan 14 13:53:28.194000 audit: BPF prog-id=221 op=LOAD Jan 14 13:53:28.194000 audit[4918]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=4881 pid=4918 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.194000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3337386465323231633031383833303865346264656261376361336232 Jan 14 13:53:28.194000 audit: BPF prog-id=222 op=LOAD Jan 14 13:53:28.194000 audit[4918]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=4881 pid=4918 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.194000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3337386465323231633031383833303865346264656261376361336232 Jan 14 13:53:28.194000 audit: BPF prog-id=222 op=UNLOAD Jan 14 13:53:28.194000 audit[4918]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4881 pid=4918 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.194000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3337386465323231633031383833303865346264656261376361336232 Jan 14 13:53:28.194000 audit: BPF prog-id=221 op=UNLOAD Jan 14 13:53:28.194000 audit[4918]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4881 pid=4918 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.194000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3337386465323231633031383833303865346264656261376361336232 Jan 14 13:53:28.194000 audit: BPF prog-id=223 op=LOAD Jan 14 13:53:28.194000 audit[4918]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=4881 pid=4918 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.194000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3337386465323231633031383833303865346264656261376361336232 Jan 14 13:53:28.252922 containerd[1635]: time="2026-01-14T13:53:28.252139666Z" level=info msg="StartContainer for \"378de221c0188308e4bdeba7ca3b264d82fed1193189438312d05c523b2fb94e\" returns successfully" Jan 14 13:53:28.387284 containerd[1635]: time="2026-01-14T13:53:28.387223867Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6cdc99f4d-mck2c,Uid:b158b3fd-5a6b-446c-92b0-d1c78d153ee9,Namespace:calico-system,Attempt:0,}" Jan 14 13:53:28.440499 kubelet[2951]: E0114 13:53:28.435961 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:28.440499 kubelet[2951]: E0114 13:53:28.437926 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:53:28.484285 kubelet[2951]: I0114 13:53:28.483774 2951 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-r9sfd" podStartSLOduration=113.48375873 podStartE2EDuration="1m53.48375873s" podCreationTimestamp="2026-01-14 13:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-14 13:53:28.483156398 +0000 UTC m=+115.731381665" watchObservedRunningTime="2026-01-14 13:53:28.48375873 +0000 UTC m=+115.731983998" Jan 14 13:53:28.521000 audit[4963]: NETFILTER_CFG table=filter:123 family=2 entries=20 op=nft_register_rule pid=4963 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:28.521000 audit[4963]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffce30b2be0 a2=0 a3=7ffce30b2bcc items=0 ppid=3117 pid=4963 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.521000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:53:28.528000 audit[4963]: NETFILTER_CFG table=nat:124 family=2 entries=14 op=nft_register_rule pid=4963 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:28.528000 audit[4963]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffce30b2be0 a2=0 a3=0 items=0 ppid=3117 pid=4963 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.528000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:53:28.715098 systemd-networkd[1527]: cali2a9f4cfc03c: Link UP Jan 14 13:53:28.716669 systemd-networkd[1527]: cali2a9f4cfc03c: Gained carrier Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.561 [INFO][4948] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--6cdc99f4d--mck2c-eth0 calico-kube-controllers-6cdc99f4d- calico-system b158b3fd-5a6b-446c-92b0-d1c78d153ee9 1022 0 2026-01-14 13:52:27 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:6cdc99f4d projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-6cdc99f4d-mck2c eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali2a9f4cfc03c [] [] }} ContainerID="d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" Namespace="calico-system" Pod="calico-kube-controllers-6cdc99f4d-mck2c" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6cdc99f4d--mck2c-" Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.563 [INFO][4948] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" Namespace="calico-system" Pod="calico-kube-controllers-6cdc99f4d-mck2c" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6cdc99f4d--mck2c-eth0" Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.627 [INFO][4968] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" HandleID="k8s-pod-network.d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" Workload="localhost-k8s-calico--kube--controllers--6cdc99f4d--mck2c-eth0" Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.628 [INFO][4968] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" HandleID="k8s-pod-network.d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" Workload="localhost-k8s-calico--kube--controllers--6cdc99f4d--mck2c-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000592430), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-6cdc99f4d-mck2c", "timestamp":"2026-01-14 13:53:28.627778231 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.628 [INFO][4968] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.628 [INFO][4968] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.628 [INFO][4968] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.644 [INFO][4968] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" host="localhost" Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.657 [INFO][4968] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.673 [INFO][4968] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.677 [INFO][4968] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.681 [INFO][4968] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.681 [INFO][4968] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" host="localhost" Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.685 [INFO][4968] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42 Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.694 [INFO][4968] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" host="localhost" Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.702 [INFO][4968] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" host="localhost" Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.702 [INFO][4968] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" host="localhost" Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.702 [INFO][4968] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 13:53:28.748020 containerd[1635]: 2026-01-14 13:53:28.703 [INFO][4968] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" HandleID="k8s-pod-network.d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" Workload="localhost-k8s-calico--kube--controllers--6cdc99f4d--mck2c-eth0" Jan 14 13:53:28.749449 containerd[1635]: 2026-01-14 13:53:28.709 [INFO][4948] cni-plugin/k8s.go 418: Populated endpoint ContainerID="d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" Namespace="calico-system" Pod="calico-kube-controllers-6cdc99f4d-mck2c" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6cdc99f4d--mck2c-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--6cdc99f4d--mck2c-eth0", GenerateName:"calico-kube-controllers-6cdc99f4d-", Namespace:"calico-system", SelfLink:"", UID:"b158b3fd-5a6b-446c-92b0-d1c78d153ee9", ResourceVersion:"1022", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 52, 27, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6cdc99f4d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-6cdc99f4d-mck2c", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali2a9f4cfc03c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:28.749449 containerd[1635]: 2026-01-14 13:53:28.709 [INFO][4948] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" Namespace="calico-system" Pod="calico-kube-controllers-6cdc99f4d-mck2c" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6cdc99f4d--mck2c-eth0" Jan 14 13:53:28.749449 containerd[1635]: 2026-01-14 13:53:28.710 [INFO][4948] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2a9f4cfc03c ContainerID="d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" Namespace="calico-system" Pod="calico-kube-controllers-6cdc99f4d-mck2c" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6cdc99f4d--mck2c-eth0" Jan 14 13:53:28.749449 containerd[1635]: 2026-01-14 13:53:28.716 [INFO][4948] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" Namespace="calico-system" Pod="calico-kube-controllers-6cdc99f4d-mck2c" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6cdc99f4d--mck2c-eth0" Jan 14 13:53:28.749449 containerd[1635]: 2026-01-14 13:53:28.717 [INFO][4948] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" Namespace="calico-system" Pod="calico-kube-controllers-6cdc99f4d-mck2c" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6cdc99f4d--mck2c-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--6cdc99f4d--mck2c-eth0", GenerateName:"calico-kube-controllers-6cdc99f4d-", Namespace:"calico-system", SelfLink:"", UID:"b158b3fd-5a6b-446c-92b0-d1c78d153ee9", ResourceVersion:"1022", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 52, 27, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6cdc99f4d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42", Pod:"calico-kube-controllers-6cdc99f4d-mck2c", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali2a9f4cfc03c", MAC:"a2:2c:5d:d0:61:a1", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:28.749449 containerd[1635]: 2026-01-14 13:53:28.734 [INFO][4948] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" Namespace="calico-system" Pod="calico-kube-controllers-6cdc99f4d-mck2c" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6cdc99f4d--mck2c-eth0" Jan 14 13:53:28.777000 audit[4984]: NETFILTER_CFG table=filter:125 family=2 entries=44 op=nft_register_chain pid=4984 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 13:53:28.777000 audit[4984]: SYSCALL arch=c000003e syscall=46 success=yes exit=21952 a0=3 a1=7ffed4ee4580 a2=0 a3=7ffed4ee456c items=0 ppid=4562 pid=4984 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.777000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 13:53:28.838498 containerd[1635]: time="2026-01-14T13:53:28.838262998Z" level=info msg="connecting to shim d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42" address="unix:///run/containerd/s/ddbe466f6712b3086bbf1643e96461004acdc9d7459282df43afe5df5856ab28" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:53:28.940816 systemd[1]: Started cri-containerd-d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42.scope - libcontainer container d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42. Jan 14 13:53:28.976000 audit: BPF prog-id=224 op=LOAD Jan 14 13:53:28.977000 audit: BPF prog-id=225 op=LOAD Jan 14 13:53:28.977000 audit[5006]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=4994 pid=5006 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.977000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431636334363634373432343438313933396231366361303864313430 Jan 14 13:53:28.977000 audit: BPF prog-id=225 op=UNLOAD Jan 14 13:53:28.977000 audit[5006]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4994 pid=5006 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.977000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431636334363634373432343438313933396231366361303864313430 Jan 14 13:53:28.977000 audit: BPF prog-id=226 op=LOAD Jan 14 13:53:28.977000 audit[5006]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=4994 pid=5006 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.977000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431636334363634373432343438313933396231366361303864313430 Jan 14 13:53:28.977000 audit: BPF prog-id=227 op=LOAD Jan 14 13:53:28.977000 audit[5006]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=4994 pid=5006 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.977000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431636334363634373432343438313933396231366361303864313430 Jan 14 13:53:28.977000 audit: BPF prog-id=227 op=UNLOAD Jan 14 13:53:28.977000 audit[5006]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4994 pid=5006 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.977000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431636334363634373432343438313933396231366361303864313430 Jan 14 13:53:28.978000 audit: BPF prog-id=226 op=UNLOAD Jan 14 13:53:28.978000 audit[5006]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4994 pid=5006 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.978000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431636334363634373432343438313933396231366361303864313430 Jan 14 13:53:28.978000 audit: BPF prog-id=228 op=LOAD Jan 14 13:53:28.978000 audit[5006]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=4994 pid=5006 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:28.978000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431636334363634373432343438313933396231366361303864313430 Jan 14 13:53:28.981825 systemd-resolved[1297]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 14 13:53:29.048110 containerd[1635]: time="2026-01-14T13:53:29.047724713Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6cdc99f4d-mck2c,Uid:b158b3fd-5a6b-446c-92b0-d1c78d153ee9,Namespace:calico-system,Attempt:0,} returns sandbox id \"d1cc46647424481939b16ca08d1402c60e627bc04f7f8402acb73cccac9d0d42\"" Jan 14 13:53:29.051251 containerd[1635]: time="2026-01-14T13:53:29.051231529Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 14 13:53:29.133230 containerd[1635]: time="2026-01-14T13:53:29.132838402Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:29.135454 containerd[1635]: time="2026-01-14T13:53:29.135186848Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 14 13:53:29.135454 containerd[1635]: time="2026-01-14T13:53:29.135284961Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:29.135878 kubelet[2951]: E0114 13:53:29.135705 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 14 13:53:29.135878 kubelet[2951]: E0114 13:53:29.135747 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 14 13:53:29.135878 kubelet[2951]: E0114 13:53:29.135821 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-kube-controllers start failed in pod calico-kube-controllers-6cdc99f4d-mck2c_calico-system(b158b3fd-5a6b-446c-92b0-d1c78d153ee9): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:29.135878 kubelet[2951]: E0114 13:53:29.135858 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-6cdc99f4d-mck2c" podUID="b158b3fd-5a6b-446c-92b0-d1c78d153ee9" Jan 14 13:53:29.371224 systemd-networkd[1527]: cali10ddc5ace98: Gained IPv6LL Jan 14 13:53:29.384178 containerd[1635]: time="2026-01-14T13:53:29.384106881Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6cbf5bf74d-6k99t,Uid:3a917c9a-ea97-482d-9194-bbf322e47535,Namespace:calico-apiserver,Attempt:0,}" Jan 14 13:53:29.390443 containerd[1635]: time="2026-01-14T13:53:29.390077101Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7c778bb748-8wp8q,Uid:2b372f30-5f27-467a-8d6a-b220df9d05b0,Namespace:calico-system,Attempt:0,}" Jan 14 13:53:29.393834 containerd[1635]: time="2026-01-14T13:53:29.393481525Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6cbf5bf74d-k9nzg,Uid:539656bb-a7f6-491c-9c24-e13907f6a8fd,Namespace:calico-apiserver,Attempt:0,}" Jan 14 13:53:29.441409 kubelet[2951]: E0114 13:53:29.441243 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:29.448444 kubelet[2951]: E0114 13:53:29.447931 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-6cdc99f4d-mck2c" podUID="b158b3fd-5a6b-446c-92b0-d1c78d153ee9" Jan 14 13:53:29.649000 audit[5076]: NETFILTER_CFG table=filter:126 family=2 entries=17 op=nft_register_rule pid=5076 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:29.649000 audit[5076]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffedc7b9af0 a2=0 a3=7ffedc7b9adc items=0 ppid=3117 pid=5076 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:29.649000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:53:29.661000 audit[5076]: NETFILTER_CFG table=nat:127 family=2 entries=35 op=nft_register_chain pid=5076 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:29.661000 audit[5076]: SYSCALL arch=c000003e syscall=46 success=yes exit=14196 a0=3 a1=7ffedc7b9af0 a2=0 a3=7ffedc7b9adc items=0 ppid=3117 pid=5076 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:29.661000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:53:29.884480 systemd-networkd[1527]: cali2a9f4cfc03c: Gained IPv6LL Jan 14 13:53:29.907295 systemd-networkd[1527]: cali6af684d9401: Link UP Jan 14 13:53:29.910008 systemd-networkd[1527]: cali6af684d9401: Gained carrier Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.618 [INFO][5032] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--6cbf5bf74d--6k99t-eth0 calico-apiserver-6cbf5bf74d- calico-apiserver 3a917c9a-ea97-482d-9194-bbf322e47535 1021 0 2026-01-14 13:52:18 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6cbf5bf74d projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-6cbf5bf74d-6k99t eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali6af684d9401 [] [] }} ContainerID="79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-6k99t" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--6k99t-" Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.620 [INFO][5032] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-6k99t" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--6k99t-eth0" Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.746 [INFO][5078] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" HandleID="k8s-pod-network.79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" Workload="localhost-k8s-calico--apiserver--6cbf5bf74d--6k99t-eth0" Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.746 [INFO][5078] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" HandleID="k8s-pod-network.79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" Workload="localhost-k8s-calico--apiserver--6cbf5bf74d--6k99t-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00050ca80), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-6cbf5bf74d-6k99t", "timestamp":"2026-01-14 13:53:29.746506136 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.747 [INFO][5078] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.747 [INFO][5078] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.747 [INFO][5078] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.763 [INFO][5078] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" host="localhost" Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.797 [INFO][5078] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.824 [INFO][5078] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.830 [INFO][5078] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.835 [INFO][5078] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.835 [INFO][5078] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" host="localhost" Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.840 [INFO][5078] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.852 [INFO][5078] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" host="localhost" Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.865 [INFO][5078] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" host="localhost" Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.866 [INFO][5078] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" host="localhost" Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.868 [INFO][5078] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 13:53:29.952015 containerd[1635]: 2026-01-14 13:53:29.869 [INFO][5078] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" HandleID="k8s-pod-network.79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" Workload="localhost-k8s-calico--apiserver--6cbf5bf74d--6k99t-eth0" Jan 14 13:53:29.960224 containerd[1635]: 2026-01-14 13:53:29.882 [INFO][5032] cni-plugin/k8s.go 418: Populated endpoint ContainerID="79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-6k99t" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--6k99t-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6cbf5bf74d--6k99t-eth0", GenerateName:"calico-apiserver-6cbf5bf74d-", Namespace:"calico-apiserver", SelfLink:"", UID:"3a917c9a-ea97-482d-9194-bbf322e47535", ResourceVersion:"1021", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 52, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6cbf5bf74d", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-6cbf5bf74d-6k99t", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali6af684d9401", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:29.960224 containerd[1635]: 2026-01-14 13:53:29.882 [INFO][5032] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-6k99t" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--6k99t-eth0" Jan 14 13:53:29.960224 containerd[1635]: 2026-01-14 13:53:29.884 [INFO][5032] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6af684d9401 ContainerID="79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-6k99t" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--6k99t-eth0" Jan 14 13:53:29.960224 containerd[1635]: 2026-01-14 13:53:29.906 [INFO][5032] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-6k99t" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--6k99t-eth0" Jan 14 13:53:29.960224 containerd[1635]: 2026-01-14 13:53:29.909 [INFO][5032] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-6k99t" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--6k99t-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6cbf5bf74d--6k99t-eth0", GenerateName:"calico-apiserver-6cbf5bf74d-", Namespace:"calico-apiserver", SelfLink:"", UID:"3a917c9a-ea97-482d-9194-bbf322e47535", ResourceVersion:"1021", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 52, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6cbf5bf74d", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e", Pod:"calico-apiserver-6cbf5bf74d-6k99t", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali6af684d9401", MAC:"1e:62:b4:56:1e:a3", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:29.960224 containerd[1635]: 2026-01-14 13:53:29.939 [INFO][5032] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-6k99t" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--6k99t-eth0" Jan 14 13:53:30.048781 systemd-networkd[1527]: cali2f3f740a52a: Link UP Jan 14 13:53:30.050173 systemd-networkd[1527]: cali2f3f740a52a: Gained carrier Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.618 [INFO][5055] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--6cbf5bf74d--k9nzg-eth0 calico-apiserver-6cbf5bf74d- calico-apiserver 539656bb-a7f6-491c-9c24-e13907f6a8fd 1023 0 2026-01-14 13:52:18 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6cbf5bf74d projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-6cbf5bf74d-k9nzg eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali2f3f740a52a [] [] }} ContainerID="3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-k9nzg" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--k9nzg-" Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.620 [INFO][5055] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-k9nzg" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--k9nzg-eth0" Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.798 [INFO][5079] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" HandleID="k8s-pod-network.3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" Workload="localhost-k8s-calico--apiserver--6cbf5bf74d--k9nzg-eth0" Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.806 [INFO][5079] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" HandleID="k8s-pod-network.3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" Workload="localhost-k8s-calico--apiserver--6cbf5bf74d--k9nzg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001396b0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-6cbf5bf74d-k9nzg", "timestamp":"2026-01-14 13:53:29.798289119 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.806 [INFO][5079] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.866 [INFO][5079] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.866 [INFO][5079] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.880 [INFO][5079] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" host="localhost" Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.906 [INFO][5079] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.939 [INFO][5079] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.955 [INFO][5079] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.966 [INFO][5079] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.966 [INFO][5079] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" host="localhost" Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.973 [INFO][5079] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1 Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:29.982 [INFO][5079] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" host="localhost" Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:30.001 [INFO][5079] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" host="localhost" Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:30.002 [INFO][5079] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" host="localhost" Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:30.002 [INFO][5079] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 13:53:30.096880 containerd[1635]: 2026-01-14 13:53:30.002 [INFO][5079] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" HandleID="k8s-pod-network.3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" Workload="localhost-k8s-calico--apiserver--6cbf5bf74d--k9nzg-eth0" Jan 14 13:53:30.099493 containerd[1635]: 2026-01-14 13:53:30.008 [INFO][5055] cni-plugin/k8s.go 418: Populated endpoint ContainerID="3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-k9nzg" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--k9nzg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6cbf5bf74d--k9nzg-eth0", GenerateName:"calico-apiserver-6cbf5bf74d-", Namespace:"calico-apiserver", SelfLink:"", UID:"539656bb-a7f6-491c-9c24-e13907f6a8fd", ResourceVersion:"1023", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 52, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6cbf5bf74d", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-6cbf5bf74d-k9nzg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali2f3f740a52a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:30.099493 containerd[1635]: 2026-01-14 13:53:30.008 [INFO][5055] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-k9nzg" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--k9nzg-eth0" Jan 14 13:53:30.099493 containerd[1635]: 2026-01-14 13:53:30.010 [INFO][5055] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2f3f740a52a ContainerID="3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-k9nzg" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--k9nzg-eth0" Jan 14 13:53:30.099493 containerd[1635]: 2026-01-14 13:53:30.049 [INFO][5055] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-k9nzg" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--k9nzg-eth0" Jan 14 13:53:30.099493 containerd[1635]: 2026-01-14 13:53:30.061 [INFO][5055] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-k9nzg" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--k9nzg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6cbf5bf74d--k9nzg-eth0", GenerateName:"calico-apiserver-6cbf5bf74d-", Namespace:"calico-apiserver", SelfLink:"", UID:"539656bb-a7f6-491c-9c24-e13907f6a8fd", ResourceVersion:"1023", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 52, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6cbf5bf74d", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1", Pod:"calico-apiserver-6cbf5bf74d-k9nzg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali2f3f740a52a", MAC:"ea:49:9d:e1:2c:8e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:30.099493 containerd[1635]: 2026-01-14 13:53:30.091 [INFO][5055] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" Namespace="calico-apiserver" Pod="calico-apiserver-6cbf5bf74d-k9nzg" WorkloadEndpoint="localhost-k8s-calico--apiserver--6cbf5bf74d--k9nzg-eth0" Jan 14 13:53:30.097000 audit[5128]: NETFILTER_CFG table=filter:128 family=2 entries=62 op=nft_register_chain pid=5128 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 13:53:30.102008 containerd[1635]: time="2026-01-14T13:53:30.101696285Z" level=info msg="connecting to shim 79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e" address="unix:///run/containerd/s/90e624c94c0057a91e1879c2ec84b017b64e564cd1c9c2fe0bcc37702e87e0d2" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:53:30.097000 audit[5128]: SYSCALL arch=c000003e syscall=46 success=yes exit=31772 a0=3 a1=7ffcbd52dde0 a2=0 a3=7ffcbd52ddcc items=0 ppid=4562 pid=5128 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.097000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 13:53:30.257442 systemd[1]: Started cri-containerd-79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e.scope - libcontainer container 79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e. Jan 14 13:53:30.274646 containerd[1635]: time="2026-01-14T13:53:30.274015103Z" level=info msg="connecting to shim 3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1" address="unix:///run/containerd/s/7694d5c03673bddcd3068f6e55cff448e022b0b329cb49face83aca97fad55ee" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:53:30.281000 audit[5148]: NETFILTER_CFG table=filter:129 family=2 entries=59 op=nft_register_chain pid=5148 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 13:53:30.281000 audit[5148]: SYSCALL arch=c000003e syscall=46 success=yes exit=29492 a0=3 a1=7ffdf7d95df0 a2=0 a3=7ffdf7d95ddc items=0 ppid=4562 pid=5148 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.281000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 13:53:30.328837 systemd-networkd[1527]: cali87c2aa0d73a: Link UP Jan 14 13:53:30.329100 systemd-networkd[1527]: cali87c2aa0d73a: Gained carrier Jan 14 13:53:30.384719 kubelet[2951]: E0114 13:53:30.384484 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:30.389882 kubelet[2951]: E0114 13:53:30.389864 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:30.391483 containerd[1635]: time="2026-01-14T13:53:30.390880864Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-srck5,Uid:4efafa71-2a7d-4e25-8295-f778af674237,Namespace:kube-system,Attempt:0,}" Jan 14 13:53:30.416000 audit[5196]: NETFILTER_CFG table=filter:130 family=2 entries=60 op=nft_register_chain pid=5196 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 13:53:30.416000 audit[5196]: SYSCALL arch=c000003e syscall=46 success=yes exit=29916 a0=3 a1=7fffee5bbd70 a2=0 a3=7fffee5bbd5c items=0 ppid=4562 pid=5196 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.416000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:29.632 [INFO][5041] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--7c778bb748--8wp8q-eth0 goldmane-7c778bb748- calico-system 2b372f30-5f27-467a-8d6a-b220df9d05b0 1019 0 2026-01-14 13:52:17 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:7c778bb748 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-7c778bb748-8wp8q eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali87c2aa0d73a [] [] }} ContainerID="51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" Namespace="calico-system" Pod="goldmane-7c778bb748-8wp8q" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--8wp8q-" Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:29.632 [INFO][5041] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" Namespace="calico-system" Pod="goldmane-7c778bb748-8wp8q" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--8wp8q-eth0" Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:29.824 [INFO][5088] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" HandleID="k8s-pod-network.51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" Workload="localhost-k8s-goldmane--7c778bb748--8wp8q-eth0" Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:29.827 [INFO][5088] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" HandleID="k8s-pod-network.51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" Workload="localhost-k8s-goldmane--7c778bb748--8wp8q-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000270210), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-7c778bb748-8wp8q", "timestamp":"2026-01-14 13:53:29.824168842 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:29.827 [INFO][5088] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:30.002 [INFO][5088] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:30.005 [INFO][5088] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:30.077 [INFO][5088] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" host="localhost" Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:30.118 [INFO][5088] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:30.133 [INFO][5088] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:30.145 [INFO][5088] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:30.176 [INFO][5088] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:30.180 [INFO][5088] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" host="localhost" Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:30.204 [INFO][5088] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3 Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:30.224 [INFO][5088] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" host="localhost" Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:30.257 [INFO][5088] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" host="localhost" Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:30.258 [INFO][5088] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" host="localhost" Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:30.258 [INFO][5088] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 13:53:30.443153 containerd[1635]: 2026-01-14 13:53:30.258 [INFO][5088] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" HandleID="k8s-pod-network.51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" Workload="localhost-k8s-goldmane--7c778bb748--8wp8q-eth0" Jan 14 13:53:30.444136 containerd[1635]: 2026-01-14 13:53:30.290 [INFO][5041] cni-plugin/k8s.go 418: Populated endpoint ContainerID="51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" Namespace="calico-system" Pod="goldmane-7c778bb748-8wp8q" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--8wp8q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--7c778bb748--8wp8q-eth0", GenerateName:"goldmane-7c778bb748-", Namespace:"calico-system", SelfLink:"", UID:"2b372f30-5f27-467a-8d6a-b220df9d05b0", ResourceVersion:"1019", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 52, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"7c778bb748", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-7c778bb748-8wp8q", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali87c2aa0d73a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:30.444136 containerd[1635]: 2026-01-14 13:53:30.290 [INFO][5041] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" Namespace="calico-system" Pod="goldmane-7c778bb748-8wp8q" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--8wp8q-eth0" Jan 14 13:53:30.444136 containerd[1635]: 2026-01-14 13:53:30.290 [INFO][5041] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali87c2aa0d73a ContainerID="51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" Namespace="calico-system" Pod="goldmane-7c778bb748-8wp8q" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--8wp8q-eth0" Jan 14 13:53:30.444136 containerd[1635]: 2026-01-14 13:53:30.315 [INFO][5041] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" Namespace="calico-system" Pod="goldmane-7c778bb748-8wp8q" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--8wp8q-eth0" Jan 14 13:53:30.444136 containerd[1635]: 2026-01-14 13:53:30.317 [INFO][5041] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" Namespace="calico-system" Pod="goldmane-7c778bb748-8wp8q" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--8wp8q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--7c778bb748--8wp8q-eth0", GenerateName:"goldmane-7c778bb748-", Namespace:"calico-system", SelfLink:"", UID:"2b372f30-5f27-467a-8d6a-b220df9d05b0", ResourceVersion:"1019", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 52, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"7c778bb748", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3", Pod:"goldmane-7c778bb748-8wp8q", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali87c2aa0d73a", MAC:"22:e1:3d:ce:e9:9f", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:30.444136 containerd[1635]: 2026-01-14 13:53:30.342 [INFO][5041] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" Namespace="calico-system" Pod="goldmane-7c778bb748-8wp8q" WorkloadEndpoint="localhost-k8s-goldmane--7c778bb748--8wp8q-eth0" Jan 14 13:53:30.445613 kubelet[2951]: E0114 13:53:30.445160 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:30.446793 kubelet[2951]: E0114 13:53:30.446758 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-6cdc99f4d-mck2c" podUID="b158b3fd-5a6b-446c-92b0-d1c78d153ee9" Jan 14 13:53:30.453000 audit: BPF prog-id=229 op=LOAD Jan 14 13:53:30.455000 audit: BPF prog-id=230 op=LOAD Jan 14 13:53:30.455000 audit[5142]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=5123 pid=5142 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.455000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3739333635313736383230343338663235386331386436623633643164 Jan 14 13:53:30.456000 audit: BPF prog-id=230 op=UNLOAD Jan 14 13:53:30.456000 audit[5142]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5123 pid=5142 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.456000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3739333635313736383230343338663235386331386436623633643164 Jan 14 13:53:30.458000 audit: BPF prog-id=231 op=LOAD Jan 14 13:53:30.458000 audit[5142]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=5123 pid=5142 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.458000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3739333635313736383230343338663235386331386436623633643164 Jan 14 13:53:30.458000 audit: BPF prog-id=232 op=LOAD Jan 14 13:53:30.458000 audit[5142]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=5123 pid=5142 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.458000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3739333635313736383230343338663235386331386436623633643164 Jan 14 13:53:30.458000 audit: BPF prog-id=232 op=UNLOAD Jan 14 13:53:30.458000 audit[5142]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5123 pid=5142 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.458000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3739333635313736383230343338663235386331386436623633643164 Jan 14 13:53:30.458000 audit: BPF prog-id=231 op=UNLOAD Jan 14 13:53:30.458000 audit[5142]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5123 pid=5142 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.458000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3739333635313736383230343338663235386331386436623633643164 Jan 14 13:53:30.459000 audit: BPF prog-id=233 op=LOAD Jan 14 13:53:30.459000 audit[5142]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=5123 pid=5142 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.459000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3739333635313736383230343338663235386331386436623633643164 Jan 14 13:53:30.479747 systemd-resolved[1297]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 14 13:53:30.503106 systemd[1]: Started cri-containerd-3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1.scope - libcontainer container 3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1. Jan 14 13:53:30.597000 audit: BPF prog-id=234 op=LOAD Jan 14 13:53:30.601000 audit: BPF prog-id=235 op=LOAD Jan 14 13:53:30.601000 audit[5176]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00020c238 a2=98 a3=0 items=0 ppid=5161 pid=5176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.601000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3362653264633334666563323063393934653363313839313563353031 Jan 14 13:53:30.601000 audit: BPF prog-id=235 op=UNLOAD Jan 14 13:53:30.601000 audit[5176]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=5161 pid=5176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.601000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3362653264633334666563323063393934653363313839313563353031 Jan 14 13:53:30.601000 audit: BPF prog-id=236 op=LOAD Jan 14 13:53:30.601000 audit[5176]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00020c488 a2=98 a3=0 items=0 ppid=5161 pid=5176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.601000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3362653264633334666563323063393934653363313839313563353031 Jan 14 13:53:30.602000 audit: BPF prog-id=237 op=LOAD Jan 14 13:53:30.602000 audit[5176]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00020c218 a2=98 a3=0 items=0 ppid=5161 pid=5176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.602000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3362653264633334666563323063393934653363313839313563353031 Jan 14 13:53:30.602000 audit: BPF prog-id=237 op=UNLOAD Jan 14 13:53:30.602000 audit[5176]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=5161 pid=5176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.602000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3362653264633334666563323063393934653363313839313563353031 Jan 14 13:53:30.602000 audit: BPF prog-id=236 op=UNLOAD Jan 14 13:53:30.602000 audit[5176]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=5161 pid=5176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.602000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3362653264633334666563323063393934653363313839313563353031 Jan 14 13:53:30.602000 audit: BPF prog-id=238 op=LOAD Jan 14 13:53:30.602000 audit[5176]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00020c6e8 a2=98 a3=0 items=0 ppid=5161 pid=5176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.602000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3362653264633334666563323063393934653363313839313563353031 Jan 14 13:53:30.606027 systemd-resolved[1297]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 14 13:53:30.606284 containerd[1635]: time="2026-01-14T13:53:30.606157757Z" level=info msg="connecting to shim 51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3" address="unix:///run/containerd/s/3b16a4be44e62f004c433d1ac84861691124781c638f8e28cdad816cd4af8a78" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:53:30.667071 containerd[1635]: time="2026-01-14T13:53:30.667027701Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6cbf5bf74d-6k99t,Uid:3a917c9a-ea97-482d-9194-bbf322e47535,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"79365176820438f258c18d6b63d1d401e93b556c0f38ffb31fed4a7913e4e26e\"" Jan 14 13:53:30.683766 containerd[1635]: time="2026-01-14T13:53:30.683421129Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 14 13:53:30.703199 systemd[1]: Started cri-containerd-51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3.scope - libcontainer container 51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3. Jan 14 13:53:30.753000 audit: BPF prog-id=239 op=LOAD Jan 14 13:53:30.754000 audit: BPF prog-id=240 op=LOAD Jan 14 13:53:30.754000 audit[5253]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000186238 a2=98 a3=0 items=0 ppid=5240 pid=5253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.754000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3531636662646137613736633166623032306133383065616334316332 Jan 14 13:53:30.754000 audit: BPF prog-id=240 op=UNLOAD Jan 14 13:53:30.754000 audit[5253]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5240 pid=5253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.754000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3531636662646137613736633166623032306133383065616334316332 Jan 14 13:53:30.755000 audit: BPF prog-id=241 op=LOAD Jan 14 13:53:30.755000 audit[5253]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000186488 a2=98 a3=0 items=0 ppid=5240 pid=5253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.755000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3531636662646137613736633166623032306133383065616334316332 Jan 14 13:53:30.755000 audit: BPF prog-id=242 op=LOAD Jan 14 13:53:30.755000 audit[5253]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000186218 a2=98 a3=0 items=0 ppid=5240 pid=5253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.755000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3531636662646137613736633166623032306133383065616334316332 Jan 14 13:53:30.755000 audit: BPF prog-id=242 op=UNLOAD Jan 14 13:53:30.755000 audit[5253]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5240 pid=5253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.755000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3531636662646137613736633166623032306133383065616334316332 Jan 14 13:53:30.755000 audit: BPF prog-id=241 op=UNLOAD Jan 14 13:53:30.755000 audit[5253]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5240 pid=5253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.755000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3531636662646137613736633166623032306133383065616334316332 Jan 14 13:53:30.755000 audit: BPF prog-id=243 op=LOAD Jan 14 13:53:30.755000 audit[5253]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001866e8 a2=98 a3=0 items=0 ppid=5240 pid=5253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.755000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3531636662646137613736633166623032306133383065616334316332 Jan 14 13:53:30.768023 systemd-resolved[1297]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 14 13:53:30.796406 containerd[1635]: time="2026-01-14T13:53:30.796216668Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6cbf5bf74d-k9nzg,Uid:539656bb-a7f6-491c-9c24-e13907f6a8fd,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"3be2dc34fec20c994e3c18915c501da6f04d0b7cfdcf0184ec505b233875a4a1\"" Jan 14 13:53:30.797183 containerd[1635]: time="2026-01-14T13:53:30.797003111Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:30.798910 containerd[1635]: time="2026-01-14T13:53:30.798806901Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 14 13:53:30.798955 containerd[1635]: time="2026-01-14T13:53:30.798936022Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:30.799845 kubelet[2951]: E0114 13:53:30.799322 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 13:53:30.799845 kubelet[2951]: E0114 13:53:30.799726 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 13:53:30.799960 kubelet[2951]: E0114 13:53:30.799921 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-apiserver start failed in pod calico-apiserver-6cbf5bf74d-6k99t_calico-apiserver(3a917c9a-ea97-482d-9194-bbf322e47535): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:30.800090 kubelet[2951]: E0114 13:53:30.799954 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-6k99t" podUID="3a917c9a-ea97-482d-9194-bbf322e47535" Jan 14 13:53:30.809023 containerd[1635]: time="2026-01-14T13:53:30.808983962Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 14 13:53:30.817197 systemd-networkd[1527]: cali70bae19fd6e: Link UP Jan 14 13:53:30.829447 systemd-networkd[1527]: cali70bae19fd6e: Gained carrier Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.562 [INFO][5192] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--66bc5c9577--srck5-eth0 coredns-66bc5c9577- kube-system 4efafa71-2a7d-4e25-8295-f778af674237 1024 0 2026-01-14 13:51:35 +0000 UTC map[k8s-app:kube-dns pod-template-hash:66bc5c9577 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-66bc5c9577-srck5 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali70bae19fd6e [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 } {liveness-probe TCP 8080 0 } {readiness-probe TCP 8181 0 }] [] }} ContainerID="c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" Namespace="kube-system" Pod="coredns-66bc5c9577-srck5" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--srck5-" Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.563 [INFO][5192] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" Namespace="kube-system" Pod="coredns-66bc5c9577-srck5" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--srck5-eth0" Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.645 [INFO][5231] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" HandleID="k8s-pod-network.c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" Workload="localhost-k8s-coredns--66bc5c9577--srck5-eth0" Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.645 [INFO][5231] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" HandleID="k8s-pod-network.c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" Workload="localhost-k8s-coredns--66bc5c9577--srck5-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0000c0d20), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-66bc5c9577-srck5", "timestamp":"2026-01-14 13:53:30.645140461 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.646 [INFO][5231] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.646 [INFO][5231] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.646 [INFO][5231] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.669 [INFO][5231] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" host="localhost" Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.699 [INFO][5231] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.718 [INFO][5231] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.725 [INFO][5231] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.734 [INFO][5231] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.734 [INFO][5231] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" host="localhost" Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.740 [INFO][5231] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459 Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.751 [INFO][5231] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" host="localhost" Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.780 [INFO][5231] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" host="localhost" Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.781 [INFO][5231] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" host="localhost" Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.782 [INFO][5231] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 13:53:30.860216 containerd[1635]: 2026-01-14 13:53:30.783 [INFO][5231] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" HandleID="k8s-pod-network.c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" Workload="localhost-k8s-coredns--66bc5c9577--srck5-eth0" Jan 14 13:53:30.861070 containerd[1635]: 2026-01-14 13:53:30.792 [INFO][5192] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" Namespace="kube-system" Pod="coredns-66bc5c9577-srck5" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--srck5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--66bc5c9577--srck5-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"4efafa71-2a7d-4e25-8295-f778af674237", ResourceVersion:"1024", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 51, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-66bc5c9577-srck5", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali70bae19fd6e", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:30.861070 containerd[1635]: 2026-01-14 13:53:30.792 [INFO][5192] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" Namespace="kube-system" Pod="coredns-66bc5c9577-srck5" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--srck5-eth0" Jan 14 13:53:30.861070 containerd[1635]: 2026-01-14 13:53:30.792 [INFO][5192] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali70bae19fd6e ContainerID="c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" Namespace="kube-system" Pod="coredns-66bc5c9577-srck5" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--srck5-eth0" Jan 14 13:53:30.861070 containerd[1635]: 2026-01-14 13:53:30.826 [INFO][5192] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" Namespace="kube-system" Pod="coredns-66bc5c9577-srck5" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--srck5-eth0" Jan 14 13:53:30.861070 containerd[1635]: 2026-01-14 13:53:30.832 [INFO][5192] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" Namespace="kube-system" Pod="coredns-66bc5c9577-srck5" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--srck5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--66bc5c9577--srck5-eth0", GenerateName:"coredns-66bc5c9577-", Namespace:"kube-system", SelfLink:"", UID:"4efafa71-2a7d-4e25-8295-f778af674237", ResourceVersion:"1024", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 13, 51, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"66bc5c9577", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459", Pod:"coredns-66bc5c9577-srck5", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali70bae19fd6e", MAC:"fe:1e:be:20:98:e2", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"liveness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1f90, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"readiness-probe", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x1ff5, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 13:53:30.861070 containerd[1635]: 2026-01-14 13:53:30.854 [INFO][5192] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" Namespace="kube-system" Pod="coredns-66bc5c9577-srck5" WorkloadEndpoint="localhost-k8s-coredns--66bc5c9577--srck5-eth0" Jan 14 13:53:30.904298 containerd[1635]: time="2026-01-14T13:53:30.903806981Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:30.907000 audit[5304]: NETFILTER_CFG table=filter:131 family=2 entries=52 op=nft_register_chain pid=5304 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 13:53:30.907000 audit[5304]: SYSCALL arch=c000003e syscall=46 success=yes exit=23892 a0=3 a1=7ffccd7474b0 a2=0 a3=7ffccd74749c items=0 ppid=4562 pid=5304 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:30.907000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 13:53:30.917084 containerd[1635]: time="2026-01-14T13:53:30.916303878Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 14 13:53:30.917084 containerd[1635]: time="2026-01-14T13:53:30.916785380Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:30.919292 kubelet[2951]: E0114 13:53:30.916852 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 13:53:30.919292 kubelet[2951]: E0114 13:53:30.916892 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 13:53:30.919292 kubelet[2951]: E0114 13:53:30.916974 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-apiserver start failed in pod calico-apiserver-6cbf5bf74d-k9nzg_calico-apiserver(539656bb-a7f6-491c-9c24-e13907f6a8fd): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:30.919292 kubelet[2951]: E0114 13:53:30.917003 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" podUID="539656bb-a7f6-491c-9c24-e13907f6a8fd" Jan 14 13:53:30.922036 containerd[1635]: time="2026-01-14T13:53:30.921658856Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-7c778bb748-8wp8q,Uid:2b372f30-5f27-467a-8d6a-b220df9d05b0,Namespace:calico-system,Attempt:0,} returns sandbox id \"51cfbda7a76c1fb020a380eac41c2ca441552c46746a2ad041deb39f020c8bc3\"" Jan 14 13:53:30.933300 containerd[1635]: time="2026-01-14T13:53:30.932923260Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 14 13:53:30.967148 containerd[1635]: time="2026-01-14T13:53:30.966968967Z" level=info msg="connecting to shim c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459" address="unix:///run/containerd/s/b77dcc423abd090239fa71e11588eda167708170eacc404db97c305d9ca83aaf" namespace=k8s.io protocol=ttrpc version=3 Jan 14 13:53:30.971756 systemd-networkd[1527]: cali6af684d9401: Gained IPv6LL Jan 14 13:53:31.043869 containerd[1635]: time="2026-01-14T13:53:31.043315190Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:31.047822 containerd[1635]: time="2026-01-14T13:53:31.047152178Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 14 13:53:31.049086 containerd[1635]: time="2026-01-14T13:53:31.047670996Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:31.050785 kubelet[2951]: E0114 13:53:31.050672 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 14 13:53:31.050785 kubelet[2951]: E0114 13:53:31.050773 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 14 13:53:31.052110 kubelet[2951]: E0114 13:53:31.051842 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container goldmane start failed in pod goldmane-7c778bb748-8wp8q_calico-system(2b372f30-5f27-467a-8d6a-b220df9d05b0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:31.052110 kubelet[2951]: E0114 13:53:31.051886 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-8wp8q" podUID="2b372f30-5f27-467a-8d6a-b220df9d05b0" Jan 14 13:53:31.054242 systemd[1]: Started cri-containerd-c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459.scope - libcontainer container c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459. Jan 14 13:53:31.111000 audit: BPF prog-id=244 op=LOAD Jan 14 13:53:31.115000 audit: BPF prog-id=245 op=LOAD Jan 14 13:53:31.115000 audit[5326]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=5315 pid=5326 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.115000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6338396235366231383834373339323733323966333735636665616130 Jan 14 13:53:31.115000 audit: BPF prog-id=245 op=UNLOAD Jan 14 13:53:31.115000 audit[5326]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5315 pid=5326 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.115000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6338396235366231383834373339323733323966333735636665616130 Jan 14 13:53:31.117000 audit: BPF prog-id=246 op=LOAD Jan 14 13:53:31.117000 audit[5326]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=5315 pid=5326 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.117000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6338396235366231383834373339323733323966333735636665616130 Jan 14 13:53:31.117000 audit: BPF prog-id=247 op=LOAD Jan 14 13:53:31.117000 audit[5326]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=5315 pid=5326 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.117000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6338396235366231383834373339323733323966333735636665616130 Jan 14 13:53:31.117000 audit: BPF prog-id=247 op=UNLOAD Jan 14 13:53:31.117000 audit[5326]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5315 pid=5326 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.117000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6338396235366231383834373339323733323966333735636665616130 Jan 14 13:53:31.118000 audit: BPF prog-id=246 op=UNLOAD Jan 14 13:53:31.118000 audit[5326]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5315 pid=5326 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.118000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6338396235366231383834373339323733323966333735636665616130 Jan 14 13:53:31.118000 audit: BPF prog-id=248 op=LOAD Jan 14 13:53:31.118000 audit[5326]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=5315 pid=5326 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.118000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6338396235366231383834373339323733323966333735636665616130 Jan 14 13:53:31.124298 systemd-resolved[1297]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 14 13:53:31.187000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.127:22-10.0.0.1:33066 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:31.188323 systemd[1]: Started sshd@7-10.0.0.127:22-10.0.0.1:33066.service - OpenSSH per-connection server daemon (10.0.0.1:33066). Jan 14 13:53:31.225966 containerd[1635]: time="2026-01-14T13:53:31.225805716Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-66bc5c9577-srck5,Uid:4efafa71-2a7d-4e25-8295-f778af674237,Namespace:kube-system,Attempt:0,} returns sandbox id \"c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459\"" Jan 14 13:53:31.227149 kubelet[2951]: E0114 13:53:31.227122 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:31.247956 containerd[1635]: time="2026-01-14T13:53:31.247899547Z" level=info msg="CreateContainer within sandbox \"c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 14 13:53:31.275310 containerd[1635]: time="2026-01-14T13:53:31.274630848Z" level=info msg="Container bbeda9494c56ccdd1e1d41f249b7234182082fcdc0d248d5dd61eaa70680c9f6: CDI devices from CRI Config.CDIDevices: []" Jan 14 13:53:31.309811 containerd[1635]: time="2026-01-14T13:53:31.309774931Z" level=info msg="CreateContainer within sandbox \"c89b56b188473927329f375cfeaa092fef744218b4c54e10f9db1e8a5ce60459\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"bbeda9494c56ccdd1e1d41f249b7234182082fcdc0d248d5dd61eaa70680c9f6\"" Jan 14 13:53:31.316017 containerd[1635]: time="2026-01-14T13:53:31.315916974Z" level=info msg="StartContainer for \"bbeda9494c56ccdd1e1d41f249b7234182082fcdc0d248d5dd61eaa70680c9f6\"" Jan 14 13:53:31.325802 containerd[1635]: time="2026-01-14T13:53:31.325733473Z" level=info msg="connecting to shim bbeda9494c56ccdd1e1d41f249b7234182082fcdc0d248d5dd61eaa70680c9f6" address="unix:///run/containerd/s/b77dcc423abd090239fa71e11588eda167708170eacc404db97c305d9ca83aaf" protocol=ttrpc version=3 Jan 14 13:53:31.469335 kubelet[2951]: E0114 13:53:31.466457 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-8wp8q" podUID="2b372f30-5f27-467a-8d6a-b220df9d05b0" Jan 14 13:53:31.471000 audit[5347]: USER_ACCT pid=5347 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:31.473000 audit[5347]: CRED_ACQ pid=5347 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:31.473000 audit[5347]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcdca59870 a2=3 a3=0 items=0 ppid=1 pid=5347 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=9 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.473000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:53:31.477004 sshd[5347]: Accepted publickey for core from 10.0.0.1 port 33066 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:53:31.478072 sshd-session[5347]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:53:31.511902 kubelet[2951]: E0114 13:53:31.511805 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" podUID="539656bb-a7f6-491c-9c24-e13907f6a8fd" Jan 14 13:53:31.536138 systemd[1]: Started cri-containerd-bbeda9494c56ccdd1e1d41f249b7234182082fcdc0d248d5dd61eaa70680c9f6.scope - libcontainer container bbeda9494c56ccdd1e1d41f249b7234182082fcdc0d248d5dd61eaa70680c9f6. Jan 14 13:53:31.548104 systemd-logind[1617]: New session 9 of user core. Jan 14 13:53:31.549907 systemd-networkd[1527]: cali2f3f740a52a: Gained IPv6LL Jan 14 13:53:31.553708 systemd[1]: Started session-9.scope - Session 9 of User core. Jan 14 13:53:31.562794 kubelet[2951]: E0114 13:53:31.561832 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-6k99t" podUID="3a917c9a-ea97-482d-9194-bbf322e47535" Jan 14 13:53:31.576000 audit[5347]: USER_START pid=5347 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:31.585000 audit[5379]: CRED_ACQ pid=5379 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:31.640000 audit: BPF prog-id=249 op=LOAD Jan 14 13:53:31.640000 audit: BPF prog-id=250 op=LOAD Jan 14 13:53:31.640000 audit[5356]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000124238 a2=98 a3=0 items=0 ppid=5315 pid=5356 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.640000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6262656461393439346335366363646431653164343166323439623732 Jan 14 13:53:31.640000 audit: BPF prog-id=250 op=UNLOAD Jan 14 13:53:31.640000 audit[5356]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5315 pid=5356 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.640000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6262656461393439346335366363646431653164343166323439623732 Jan 14 13:53:31.640000 audit: BPF prog-id=251 op=LOAD Jan 14 13:53:31.640000 audit[5356]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000124488 a2=98 a3=0 items=0 ppid=5315 pid=5356 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.640000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6262656461393439346335366363646431653164343166323439623732 Jan 14 13:53:31.640000 audit: BPF prog-id=252 op=LOAD Jan 14 13:53:31.640000 audit[5356]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000124218 a2=98 a3=0 items=0 ppid=5315 pid=5356 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.640000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6262656461393439346335366363646431653164343166323439623732 Jan 14 13:53:31.640000 audit: BPF prog-id=252 op=UNLOAD Jan 14 13:53:31.640000 audit[5356]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5315 pid=5356 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.640000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6262656461393439346335366363646431653164343166323439623732 Jan 14 13:53:31.640000 audit: BPF prog-id=251 op=UNLOAD Jan 14 13:53:31.640000 audit[5356]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5315 pid=5356 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.640000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6262656461393439346335366363646431653164343166323439623732 Jan 14 13:53:31.641000 audit: BPF prog-id=253 op=LOAD Jan 14 13:53:31.641000 audit[5356]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001246e8 a2=98 a3=0 items=0 ppid=5315 pid=5356 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.641000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6262656461393439346335366363646431653164343166323439623732 Jan 14 13:53:31.647000 audit[5386]: NETFILTER_CFG table=filter:132 family=2 entries=14 op=nft_register_rule pid=5386 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:31.647000 audit[5386]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffddc1de9c0 a2=0 a3=7ffddc1de9ac items=0 ppid=3117 pid=5386 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.647000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:53:31.655000 audit[5386]: NETFILTER_CFG table=nat:133 family=2 entries=20 op=nft_register_rule pid=5386 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:31.655000 audit[5386]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffddc1de9c0 a2=0 a3=7ffddc1de9ac items=0 ppid=3117 pid=5386 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:31.655000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:53:31.695142 containerd[1635]: time="2026-01-14T13:53:31.694494900Z" level=info msg="StartContainer for \"bbeda9494c56ccdd1e1d41f249b7234182082fcdc0d248d5dd61eaa70680c9f6\" returns successfully" Jan 14 13:53:31.832116 sshd[5379]: Connection closed by 10.0.0.1 port 33066 Jan 14 13:53:31.832314 sshd-session[5347]: pam_unix(sshd:session): session closed for user core Jan 14 13:53:31.834000 audit[5347]: USER_END pid=5347 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:31.838862 systemd[1]: sshd@7-10.0.0.127:22-10.0.0.1:33066.service: Deactivated successfully. Jan 14 13:53:31.841476 kernel: kauditd_printk_skb: 235 callbacks suppressed Jan 14 13:53:31.841888 kernel: audit: type=1106 audit(1768398811.834:737): pid=5347 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:31.842442 systemd[1]: session-9.scope: Deactivated successfully. Jan 14 13:53:31.847671 systemd-logind[1617]: Session 9 logged out. Waiting for processes to exit. Jan 14 13:53:31.849409 systemd-logind[1617]: Removed session 9. Jan 14 13:53:31.834000 audit[5347]: CRED_DISP pid=5347 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:31.886068 kernel: audit: type=1104 audit(1768398811.834:738): pid=5347 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:31.886132 kernel: audit: type=1131 audit(1768398811.839:739): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.127:22-10.0.0.1:33066 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:31.839000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.127:22-10.0.0.1:33066 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:31.931287 systemd-networkd[1527]: cali87c2aa0d73a: Gained IPv6LL Jan 14 13:53:32.542397 kubelet[2951]: E0114 13:53:32.541945 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:32.546036 kubelet[2951]: E0114 13:53:32.545736 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-8wp8q" podUID="2b372f30-5f27-467a-8d6a-b220df9d05b0" Jan 14 13:53:32.546036 kubelet[2951]: E0114 13:53:32.545807 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" podUID="539656bb-a7f6-491c-9c24-e13907f6a8fd" Jan 14 13:53:32.546416 kubelet[2951]: E0114 13:53:32.546193 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-6k99t" podUID="3a917c9a-ea97-482d-9194-bbf322e47535" Jan 14 13:53:32.638041 kubelet[2951]: I0114 13:53:32.637771 2951 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-66bc5c9577-srck5" podStartSLOduration=117.637757317 podStartE2EDuration="1m57.637757317s" podCreationTimestamp="2026-01-14 13:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-14 13:53:32.608251953 +0000 UTC m=+119.856477221" watchObservedRunningTime="2026-01-14 13:53:32.637757317 +0000 UTC m=+119.885982585" Jan 14 13:53:32.693000 audit[5417]: NETFILTER_CFG table=filter:134 family=2 entries=14 op=nft_register_rule pid=5417 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:32.716286 kernel: audit: type=1325 audit(1768398812.693:740): table=filter:134 family=2 entries=14 op=nft_register_rule pid=5417 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:32.716494 kernel: audit: type=1300 audit(1768398812.693:740): arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7fffbaa782d0 a2=0 a3=7fffbaa782bc items=0 ppid=3117 pid=5417 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:32.693000 audit[5417]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7fffbaa782d0 a2=0 a3=7fffbaa782bc items=0 ppid=3117 pid=5417 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:32.693000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:53:32.764969 kernel: audit: type=1327 audit(1768398812.693:740): proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:53:32.716000 audit[5417]: NETFILTER_CFG table=nat:135 family=2 entries=44 op=nft_register_rule pid=5417 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:32.782507 kernel: audit: type=1325 audit(1768398812.716:741): table=nat:135 family=2 entries=44 op=nft_register_rule pid=5417 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:32.782694 kernel: audit: type=1300 audit(1768398812.716:741): arch=c000003e syscall=46 success=yes exit=14196 a0=3 a1=7fffbaa782d0 a2=0 a3=7fffbaa782bc items=0 ppid=3117 pid=5417 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:32.716000 audit[5417]: SYSCALL arch=c000003e syscall=46 success=yes exit=14196 a0=3 a1=7fffbaa782d0 a2=0 a3=7fffbaa782bc items=0 ppid=3117 pid=5417 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:32.813762 kernel: audit: type=1327 audit(1768398812.716:741): proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:53:32.716000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:53:32.827449 systemd-networkd[1527]: cali70bae19fd6e: Gained IPv6LL Jan 14 13:53:33.547672 kubelet[2951]: E0114 13:53:33.547464 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:33.843000 audit[5421]: NETFILTER_CFG table=filter:136 family=2 entries=14 op=nft_register_rule pid=5421 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:33.843000 audit[5421]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffe52cf1b10 a2=0 a3=7ffe52cf1afc items=0 ppid=3117 pid=5421 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:33.858725 kernel: audit: type=1325 audit(1768398813.843:742): table=filter:136 family=2 entries=14 op=nft_register_rule pid=5421 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:33.843000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:53:33.872000 audit[5421]: NETFILTER_CFG table=nat:137 family=2 entries=56 op=nft_register_chain pid=5421 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:53:33.872000 audit[5421]: SYSCALL arch=c000003e syscall=46 success=yes exit=19860 a0=3 a1=7ffe52cf1b10 a2=0 a3=7ffe52cf1afc items=0 ppid=3117 pid=5421 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:33.872000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:53:34.551049 kubelet[2951]: E0114 13:53:34.550466 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:36.852141 systemd[1]: Started sshd@8-10.0.0.127:22-10.0.0.1:48960.service - OpenSSH per-connection server daemon (10.0.0.1:48960). Jan 14 13:53:36.850000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.127:22-10.0.0.1:48960 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:36.857781 kernel: kauditd_printk_skb: 5 callbacks suppressed Jan 14 13:53:36.857915 kernel: audit: type=1130 audit(1768398816.850:744): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.127:22-10.0.0.1:48960 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:36.968000 audit[5427]: USER_ACCT pid=5427 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:36.969876 sshd[5427]: Accepted publickey for core from 10.0.0.1 port 48960 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:53:36.972435 sshd-session[5427]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:53:36.987761 systemd-logind[1617]: New session 10 of user core. Jan 14 13:53:36.969000 audit[5427]: CRED_ACQ pid=5427 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:37.010875 kernel: audit: type=1101 audit(1768398816.968:745): pid=5427 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:37.010970 kernel: audit: type=1103 audit(1768398816.969:746): pid=5427 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:37.011014 kernel: audit: type=1006 audit(1768398816.969:747): pid=5427 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=10 res=1 Jan 14 13:53:36.969000 audit[5427]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe8abd7d50 a2=3 a3=0 items=0 ppid=1 pid=5427 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:37.058712 kernel: audit: type=1300 audit(1768398816.969:747): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe8abd7d50 a2=3 a3=0 items=0 ppid=1 pid=5427 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:37.058794 kernel: audit: type=1327 audit(1768398816.969:747): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:53:36.969000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:53:37.072245 systemd[1]: Started session-10.scope - Session 10 of User core. Jan 14 13:53:37.078000 audit[5427]: USER_START pid=5427 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:37.115770 kernel: audit: type=1105 audit(1768398817.078:748): pid=5427 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:37.116000 audit[5431]: CRED_ACQ pid=5431 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:37.143211 kernel: audit: type=1103 audit(1768398817.116:749): pid=5431 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:37.322797 sshd[5431]: Connection closed by 10.0.0.1 port 48960 Jan 14 13:53:37.323694 sshd-session[5427]: pam_unix(sshd:session): session closed for user core Jan 14 13:53:37.325000 audit[5427]: USER_END pid=5427 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:37.332907 systemd[1]: sshd@8-10.0.0.127:22-10.0.0.1:48960.service: Deactivated successfully. Jan 14 13:53:37.337662 systemd[1]: session-10.scope: Deactivated successfully. Jan 14 13:53:37.339305 systemd-logind[1617]: Session 10 logged out. Waiting for processes to exit. Jan 14 13:53:37.342748 systemd-logind[1617]: Removed session 10. Jan 14 13:53:37.327000 audit[5427]: CRED_DISP pid=5427 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:37.378089 kernel: audit: type=1106 audit(1768398817.325:750): pid=5427 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:37.379680 kernel: audit: type=1104 audit(1768398817.327:751): pid=5427 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:37.333000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.127:22-10.0.0.1:48960 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:37.384680 containerd[1635]: time="2026-01-14T13:53:37.383842935Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 14 13:53:37.472748 containerd[1635]: time="2026-01-14T13:53:37.472090819Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:37.474849 containerd[1635]: time="2026-01-14T13:53:37.474673100Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 14 13:53:37.474849 containerd[1635]: time="2026-01-14T13:53:37.474765452Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:37.475862 kubelet[2951]: E0114 13:53:37.475755 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 14 13:53:37.475862 kubelet[2951]: E0114 13:53:37.475803 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 14 13:53:37.477930 kubelet[2951]: E0114 13:53:37.475891 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container whisker start failed in pod whisker-868d8bc54c-7mv9t_calico-system(0b696939-2eb2-437f-b107-24730b99d30f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:37.479134 containerd[1635]: time="2026-01-14T13:53:37.478104748Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 14 13:53:37.558674 containerd[1635]: time="2026-01-14T13:53:37.558417665Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:37.561667 containerd[1635]: time="2026-01-14T13:53:37.561496895Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 14 13:53:37.561911 containerd[1635]: time="2026-01-14T13:53:37.561770571Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:37.562463 kubelet[2951]: E0114 13:53:37.562128 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 14 13:53:37.562463 kubelet[2951]: E0114 13:53:37.562251 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 14 13:53:37.562463 kubelet[2951]: E0114 13:53:37.562419 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container whisker-backend start failed in pod whisker-868d8bc54c-7mv9t_calico-system(0b696939-2eb2-437f-b107-24730b99d30f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:37.562998 kubelet[2951]: E0114 13:53:37.562480 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-868d8bc54c-7mv9t" podUID="0b696939-2eb2-437f-b107-24730b99d30f" Jan 14 13:53:39.855039 kubelet[2951]: E0114 13:53:39.852637 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:53:42.352987 systemd[1]: Started sshd@9-10.0.0.127:22-10.0.0.1:48972.service - OpenSSH per-connection server daemon (10.0.0.1:48972). Jan 14 13:53:42.354000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.127:22-10.0.0.1:48972 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:42.361939 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 14 13:53:42.362761 kernel: audit: type=1130 audit(1768398822.354:753): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.127:22-10.0.0.1:48972 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:42.399909 containerd[1635]: time="2026-01-14T13:53:42.399635925Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 14 13:53:42.478964 containerd[1635]: time="2026-01-14T13:53:42.478924620Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:42.481414 containerd[1635]: time="2026-01-14T13:53:42.481224545Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 14 13:53:42.481895 containerd[1635]: time="2026-01-14T13:53:42.481386232Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:42.482769 kubelet[2951]: E0114 13:53:42.482672 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 14 13:53:42.482769 kubelet[2951]: E0114 13:53:42.482716 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 14 13:53:42.483653 kubelet[2951]: E0114 13:53:42.482856 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-csi start failed in pod csi-node-driver-pvxlq_calico-system(967b7513-2769-43e5-a113-9344f17909c3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:42.485210 containerd[1635]: time="2026-01-14T13:53:42.484889468Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 14 13:53:42.520000 audit[5457]: USER_ACCT pid=5457 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.523070 sshd[5457]: Accepted publickey for core from 10.0.0.1 port 48972 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:53:42.526105 sshd-session[5457]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:53:42.539013 systemd-logind[1617]: New session 11 of user core. Jan 14 13:53:42.522000 audit[5457]: CRED_ACQ pid=5457 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.564410 kernel: audit: type=1101 audit(1768398822.520:754): pid=5457 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.564703 kernel: audit: type=1103 audit(1768398822.522:755): pid=5457 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.564825 kernel: audit: type=1006 audit(1768398822.522:756): pid=5457 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=11 res=1 Jan 14 13:53:42.522000 audit[5457]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fffd215da30 a2=3 a3=0 items=0 ppid=1 pid=5457 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=11 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:42.577024 systemd[1]: Started session-11.scope - Session 11 of User core. Jan 14 13:53:42.599041 kernel: audit: type=1300 audit(1768398822.522:756): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fffd215da30 a2=3 a3=0 items=0 ppid=1 pid=5457 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=11 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:42.599168 kernel: audit: type=1327 audit(1768398822.522:756): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:53:42.522000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:53:42.602998 containerd[1635]: time="2026-01-14T13:53:42.602887851Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:42.605209 containerd[1635]: time="2026-01-14T13:53:42.604452075Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 14 13:53:42.605209 containerd[1635]: time="2026-01-14T13:53:42.604770690Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:42.606093 kubelet[2951]: E0114 13:53:42.605658 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 14 13:53:42.606093 kubelet[2951]: E0114 13:53:42.605709 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 14 13:53:42.606093 kubelet[2951]: E0114 13:53:42.605778 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container csi-node-driver-registrar start failed in pod csi-node-driver-pvxlq_calico-system(967b7513-2769-43e5-a113-9344f17909c3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:42.606093 kubelet[2951]: E0114 13:53:42.605820 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:53:42.608790 kernel: audit: type=1105 audit(1768398822.581:757): pid=5457 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.581000 audit[5457]: USER_START pid=5457 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.584000 audit[5461]: CRED_ACQ pid=5461 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.652496 kernel: audit: type=1103 audit(1768398822.584:758): pid=5461 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.787417 sshd[5461]: Connection closed by 10.0.0.1 port 48972 Jan 14 13:53:42.788748 sshd-session[5457]: pam_unix(sshd:session): session closed for user core Jan 14 13:53:42.789000 audit[5457]: USER_END pid=5457 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.789000 audit[5457]: CRED_DISP pid=5457 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.831153 systemd[1]: sshd@9-10.0.0.127:22-10.0.0.1:48972.service: Deactivated successfully. Jan 14 13:53:42.835116 systemd[1]: session-11.scope: Deactivated successfully. Jan 14 13:53:42.837263 systemd-logind[1617]: Session 11 logged out. Waiting for processes to exit. Jan 14 13:53:42.844119 systemd[1]: Started sshd@10-10.0.0.127:22-10.0.0.1:48984.service - OpenSSH per-connection server daemon (10.0.0.1:48984). Jan 14 13:53:42.848614 systemd-logind[1617]: Removed session 11. Jan 14 13:53:42.851807 kernel: audit: type=1106 audit(1768398822.789:759): pid=5457 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.852051 kernel: audit: type=1104 audit(1768398822.789:760): pid=5457 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.831000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.127:22-10.0.0.1:48972 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:42.842000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.127:22-10.0.0.1:48984 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:42.937710 sshd[5476]: Accepted publickey for core from 10.0.0.1 port 48984 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:53:42.935000 audit[5476]: USER_ACCT pid=5476 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.937000 audit[5476]: CRED_ACQ pid=5476 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.937000 audit[5476]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffde4f5e1c0 a2=3 a3=0 items=0 ppid=1 pid=5476 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=12 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:42.937000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:53:42.940894 sshd-session[5476]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:53:42.951319 systemd-logind[1617]: New session 12 of user core. Jan 14 13:53:42.974891 systemd[1]: Started session-12.scope - Session 12 of User core. Jan 14 13:53:42.980000 audit[5476]: USER_START pid=5476 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:42.983000 audit[5480]: CRED_ACQ pid=5480 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:43.231812 sshd[5480]: Connection closed by 10.0.0.1 port 48984 Jan 14 13:53:43.235843 sshd-session[5476]: pam_unix(sshd:session): session closed for user core Jan 14 13:53:43.239000 audit[5476]: USER_END pid=5476 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:43.239000 audit[5476]: CRED_DISP pid=5476 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:43.248226 systemd[1]: sshd@10-10.0.0.127:22-10.0.0.1:48984.service: Deactivated successfully. Jan 14 13:53:43.248000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.127:22-10.0.0.1:48984 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:43.253299 systemd[1]: session-12.scope: Deactivated successfully. Jan 14 13:53:43.256904 systemd-logind[1617]: Session 12 logged out. Waiting for processes to exit. Jan 14 13:53:43.262700 systemd[1]: Started sshd@11-10.0.0.127:22-10.0.0.1:48994.service - OpenSSH per-connection server daemon (10.0.0.1:48994). Jan 14 13:53:43.262000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.127:22-10.0.0.1:48994 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:43.266884 systemd-logind[1617]: Removed session 12. Jan 14 13:53:43.356000 audit[5491]: USER_ACCT pid=5491 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:43.359883 sshd[5491]: Accepted publickey for core from 10.0.0.1 port 48994 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:53:43.359000 audit[5491]: CRED_ACQ pid=5491 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:43.359000 audit[5491]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe46ac99c0 a2=3 a3=0 items=0 ppid=1 pid=5491 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=13 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:43.359000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:53:43.361457 sshd-session[5491]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:53:43.371213 systemd-logind[1617]: New session 13 of user core. Jan 14 13:53:43.383888 systemd[1]: Started session-13.scope - Session 13 of User core. Jan 14 13:53:43.387000 audit[5491]: USER_START pid=5491 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:43.390000 audit[5495]: CRED_ACQ pid=5495 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:43.547029 sshd[5495]: Connection closed by 10.0.0.1 port 48994 Jan 14 13:53:43.549423 sshd-session[5491]: pam_unix(sshd:session): session closed for user core Jan 14 13:53:43.550000 audit[5491]: USER_END pid=5491 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:43.550000 audit[5491]: CRED_DISP pid=5491 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:43.556171 systemd[1]: sshd@11-10.0.0.127:22-10.0.0.1:48994.service: Deactivated successfully. Jan 14 13:53:43.555000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.127:22-10.0.0.1:48994 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:43.559915 systemd[1]: session-13.scope: Deactivated successfully. Jan 14 13:53:43.561940 systemd-logind[1617]: Session 13 logged out. Waiting for processes to exit. Jan 14 13:53:43.564008 systemd-logind[1617]: Removed session 13. Jan 14 13:53:44.382830 containerd[1635]: time="2026-01-14T13:53:44.382792835Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 14 13:53:44.459103 containerd[1635]: time="2026-01-14T13:53:44.458903667Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:44.461068 containerd[1635]: time="2026-01-14T13:53:44.460876582Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 14 13:53:44.461068 containerd[1635]: time="2026-01-14T13:53:44.461005411Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:44.461423 kubelet[2951]: E0114 13:53:44.461190 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 13:53:44.461423 kubelet[2951]: E0114 13:53:44.461229 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 13:53:44.461423 kubelet[2951]: E0114 13:53:44.461288 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-apiserver start failed in pod calico-apiserver-6cbf5bf74d-k9nzg_calico-apiserver(539656bb-a7f6-491c-9c24-e13907f6a8fd): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:44.461423 kubelet[2951]: E0114 13:53:44.461322 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" podUID="539656bb-a7f6-491c-9c24-e13907f6a8fd" Jan 14 13:53:45.383877 containerd[1635]: time="2026-01-14T13:53:45.383758160Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 14 13:53:45.490746 containerd[1635]: time="2026-01-14T13:53:45.490212918Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:45.493504 containerd[1635]: time="2026-01-14T13:53:45.493222661Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 14 13:53:45.493760 containerd[1635]: time="2026-01-14T13:53:45.493297824Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:45.494504 kubelet[2951]: E0114 13:53:45.494096 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 14 13:53:45.494504 kubelet[2951]: E0114 13:53:45.494213 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 14 13:53:45.494966 kubelet[2951]: E0114 13:53:45.494458 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container goldmane start failed in pod goldmane-7c778bb748-8wp8q_calico-system(2b372f30-5f27-467a-8d6a-b220df9d05b0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:45.494966 kubelet[2951]: E0114 13:53:45.494715 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-8wp8q" podUID="2b372f30-5f27-467a-8d6a-b220df9d05b0" Jan 14 13:53:45.496689 containerd[1635]: time="2026-01-14T13:53:45.496445009Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 14 13:53:45.578183 containerd[1635]: time="2026-01-14T13:53:45.577859345Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:45.580605 containerd[1635]: time="2026-01-14T13:53:45.580245356Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 14 13:53:45.580605 containerd[1635]: time="2026-01-14T13:53:45.580459449Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:45.581428 kubelet[2951]: E0114 13:53:45.580795 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 14 13:53:45.581428 kubelet[2951]: E0114 13:53:45.580839 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 14 13:53:45.581428 kubelet[2951]: E0114 13:53:45.580911 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-kube-controllers start failed in pod calico-kube-controllers-6cdc99f4d-mck2c_calico-system(b158b3fd-5a6b-446c-92b0-d1c78d153ee9): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:45.581428 kubelet[2951]: E0114 13:53:45.580945 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-6cdc99f4d-mck2c" podUID="b158b3fd-5a6b-446c-92b0-d1c78d153ee9" Jan 14 13:53:46.385313 containerd[1635]: time="2026-01-14T13:53:46.384090410Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 14 13:53:46.457191 containerd[1635]: time="2026-01-14T13:53:46.456997740Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:53:46.459206 containerd[1635]: time="2026-01-14T13:53:46.459020963Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 14 13:53:46.459206 containerd[1635]: time="2026-01-14T13:53:46.459086827Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 14 13:53:46.459661 kubelet[2951]: E0114 13:53:46.459483 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 13:53:46.460085 kubelet[2951]: E0114 13:53:46.459772 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 13:53:46.460085 kubelet[2951]: E0114 13:53:46.459950 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-apiserver start failed in pod calico-apiserver-6cbf5bf74d-6k99t_calico-apiserver(3a917c9a-ea97-482d-9194-bbf322e47535): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 14 13:53:46.460085 kubelet[2951]: E0114 13:53:46.459996 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-6k99t" podUID="3a917c9a-ea97-482d-9194-bbf322e47535" Jan 14 13:53:48.564000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.127:22-10.0.0.1:49040 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:48.565463 systemd[1]: Started sshd@12-10.0.0.127:22-10.0.0.1:49040.service - OpenSSH per-connection server daemon (10.0.0.1:49040). Jan 14 13:53:48.583654 kernel: kauditd_printk_skb: 23 callbacks suppressed Jan 14 13:53:48.583730 kernel: audit: type=1130 audit(1768398828.564:780): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.127:22-10.0.0.1:49040 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:48.657000 audit[5511]: USER_ACCT pid=5511 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:48.658749 sshd[5511]: Accepted publickey for core from 10.0.0.1 port 49040 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:53:48.660999 sshd-session[5511]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:53:48.671806 systemd-logind[1617]: New session 14 of user core. Jan 14 13:53:48.657000 audit[5511]: CRED_ACQ pid=5511 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:48.704775 kernel: audit: type=1101 audit(1768398828.657:781): pid=5511 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:48.704838 kernel: audit: type=1103 audit(1768398828.657:782): pid=5511 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:48.718844 kernel: audit: type=1006 audit(1768398828.657:783): pid=5511 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=14 res=1 Jan 14 13:53:48.718896 kernel: audit: type=1300 audit(1768398828.657:783): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc56ff6b20 a2=3 a3=0 items=0 ppid=1 pid=5511 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:48.657000 audit[5511]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc56ff6b20 a2=3 a3=0 items=0 ppid=1 pid=5511 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:48.744844 kernel: audit: type=1327 audit(1768398828.657:783): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:53:48.657000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:53:48.756035 systemd[1]: Started session-14.scope - Session 14 of User core. Jan 14 13:53:48.761000 audit[5511]: USER_START pid=5511 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:48.792718 kernel: audit: type=1105 audit(1768398828.761:784): pid=5511 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:48.761000 audit[5516]: CRED_ACQ pid=5516 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:48.815699 kernel: audit: type=1103 audit(1768398828.761:785): pid=5516 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:48.965511 sshd[5516]: Connection closed by 10.0.0.1 port 49040 Jan 14 13:53:48.966818 sshd-session[5511]: pam_unix(sshd:session): session closed for user core Jan 14 13:53:48.969000 audit[5511]: USER_END pid=5511 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:48.973483 systemd[1]: sshd@12-10.0.0.127:22-10.0.0.1:49040.service: Deactivated successfully. Jan 14 13:53:48.978104 systemd[1]: session-14.scope: Deactivated successfully. Jan 14 13:53:48.982029 systemd-logind[1617]: Session 14 logged out. Waiting for processes to exit. Jan 14 13:53:48.984250 systemd-logind[1617]: Removed session 14. Jan 14 13:53:48.969000 audit[5511]: CRED_DISP pid=5511 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:48.996740 kernel: audit: type=1106 audit(1768398828.969:786): pid=5511 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:48.997074 kernel: audit: type=1104 audit(1768398828.969:787): pid=5511 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:48.969000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.127:22-10.0.0.1:49040 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:52.386602 kubelet[2951]: E0114 13:53:52.386464 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-868d8bc54c-7mv9t" podUID="0b696939-2eb2-437f-b107-24730b99d30f" Jan 14 13:53:53.987208 systemd[1]: Started sshd@13-10.0.0.127:22-10.0.0.1:49044.service - OpenSSH per-connection server daemon (10.0.0.1:49044). Jan 14 13:53:53.987000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.127:22-10.0.0.1:49044 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:53.993780 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 14 13:53:53.993841 kernel: audit: type=1130 audit(1768398833.987:789): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.127:22-10.0.0.1:49044 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:54.093000 audit[5562]: USER_ACCT pid=5562 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:54.094485 sshd[5562]: Accepted publickey for core from 10.0.0.1 port 49044 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:53:54.098029 sshd-session[5562]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:53:54.109332 systemd-logind[1617]: New session 15 of user core. Jan 14 13:53:54.095000 audit[5562]: CRED_ACQ pid=5562 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:54.145733 kernel: audit: type=1101 audit(1768398834.093:790): pid=5562 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:54.145791 kernel: audit: type=1103 audit(1768398834.095:791): pid=5562 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:54.160493 kernel: audit: type=1006 audit(1768398834.096:792): pid=5562 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=15 res=1 Jan 14 13:53:54.160668 kernel: audit: type=1300 audit(1768398834.096:792): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff862e8d60 a2=3 a3=0 items=0 ppid=1 pid=5562 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:54.096000 audit[5562]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff862e8d60 a2=3 a3=0 items=0 ppid=1 pid=5562 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:54.186787 kernel: audit: type=1327 audit(1768398834.096:792): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:53:54.096000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:53:54.199042 systemd[1]: Started session-15.scope - Session 15 of User core. Jan 14 13:53:54.203000 audit[5562]: USER_START pid=5562 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:54.203000 audit[5566]: CRED_ACQ pid=5566 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:54.251183 kernel: audit: type=1105 audit(1768398834.203:793): pid=5562 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:54.251244 kernel: audit: type=1103 audit(1768398834.203:794): pid=5566 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:54.357015 sshd[5566]: Connection closed by 10.0.0.1 port 49044 Jan 14 13:53:54.358783 sshd-session[5562]: pam_unix(sshd:session): session closed for user core Jan 14 13:53:54.362000 audit[5562]: USER_END pid=5562 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:54.366897 systemd[1]: sshd@13-10.0.0.127:22-10.0.0.1:49044.service: Deactivated successfully. Jan 14 13:53:54.370149 systemd[1]: session-15.scope: Deactivated successfully. Jan 14 13:53:54.372451 systemd-logind[1617]: Session 15 logged out. Waiting for processes to exit. Jan 14 13:53:54.374229 systemd-logind[1617]: Removed session 15. Jan 14 13:53:54.395072 kernel: audit: type=1106 audit(1768398834.362:795): pid=5562 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:54.395756 kernel: audit: type=1104 audit(1768398834.362:796): pid=5562 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:54.362000 audit[5562]: CRED_DISP pid=5562 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:54.397254 kubelet[2951]: E0114 13:53:54.397151 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:53:54.367000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.127:22-10.0.0.1:49044 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:55.383238 kubelet[2951]: E0114 13:53:55.383154 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" podUID="539656bb-a7f6-491c-9c24-e13907f6a8fd" Jan 14 13:53:56.393983 kubelet[2951]: E0114 13:53:56.393788 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-8wp8q" podUID="2b372f30-5f27-467a-8d6a-b220df9d05b0" Jan 14 13:53:58.382964 kubelet[2951]: E0114 13:53:58.382281 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-6cdc99f4d-mck2c" podUID="b158b3fd-5a6b-446c-92b0-d1c78d153ee9" Jan 14 13:53:59.377999 systemd[1]: Started sshd@14-10.0.0.127:22-10.0.0.1:43494.service - OpenSSH per-connection server daemon (10.0.0.1:43494). Jan 14 13:53:59.379000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.127:22-10.0.0.1:43494 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:59.385789 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 14 13:53:59.386225 kernel: audit: type=1130 audit(1768398839.379:798): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.127:22-10.0.0.1:43494 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:53:59.482000 audit[5581]: USER_ACCT pid=5581 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:59.483989 sshd[5581]: Accepted publickey for core from 10.0.0.1 port 43494 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:53:59.487338 sshd-session[5581]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:53:59.496805 systemd-logind[1617]: New session 16 of user core. Jan 14 13:53:59.485000 audit[5581]: CRED_ACQ pid=5581 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:59.526698 kernel: audit: type=1101 audit(1768398839.482:799): pid=5581 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:59.526752 kernel: audit: type=1103 audit(1768398839.485:800): pid=5581 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:59.526771 kernel: audit: type=1006 audit(1768398839.485:801): pid=5581 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=16 res=1 Jan 14 13:53:59.485000 audit[5581]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff73e757c0 a2=3 a3=0 items=0 ppid=1 pid=5581 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:59.564630 kernel: audit: type=1300 audit(1768398839.485:801): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff73e757c0 a2=3 a3=0 items=0 ppid=1 pid=5581 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:53:59.564678 kernel: audit: type=1327 audit(1768398839.485:801): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:53:59.485000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:53:59.577140 systemd[1]: Started session-16.scope - Session 16 of User core. Jan 14 13:53:59.581000 audit[5581]: USER_START pid=5581 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:59.585000 audit[5585]: CRED_ACQ pid=5585 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:59.629066 kernel: audit: type=1105 audit(1768398839.581:802): pid=5581 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:59.629115 kernel: audit: type=1103 audit(1768398839.585:803): pid=5585 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:59.736845 sshd[5585]: Connection closed by 10.0.0.1 port 43494 Jan 14 13:53:59.739019 sshd-session[5581]: pam_unix(sshd:session): session closed for user core Jan 14 13:53:59.742000 audit[5581]: USER_END pid=5581 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:59.747016 systemd[1]: sshd@14-10.0.0.127:22-10.0.0.1:43494.service: Deactivated successfully. Jan 14 13:53:59.750766 systemd[1]: session-16.scope: Deactivated successfully. Jan 14 13:53:59.753984 systemd-logind[1617]: Session 16 logged out. Waiting for processes to exit. Jan 14 13:53:59.755707 systemd-logind[1617]: Removed session 16. Jan 14 13:53:59.742000 audit[5581]: CRED_DISP pid=5581 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:59.801799 kernel: audit: type=1106 audit(1768398839.742:804): pid=5581 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:59.801871 kernel: audit: type=1104 audit(1768398839.742:805): pid=5581 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:53:59.747000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.127:22-10.0.0.1:43494 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:01.382146 kubelet[2951]: E0114 13:54:01.382012 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-6k99t" podUID="3a917c9a-ea97-482d-9194-bbf322e47535" Jan 14 13:54:04.756000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.127:22-10.0.0.1:51628 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:04.756747 systemd[1]: Started sshd@15-10.0.0.127:22-10.0.0.1:51628.service - OpenSSH per-connection server daemon (10.0.0.1:51628). Jan 14 13:54:04.763426 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 14 13:54:04.763700 kernel: audit: type=1130 audit(1768398844.756:807): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.127:22-10.0.0.1:51628 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:04.855000 audit[5605]: USER_ACCT pid=5605 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:04.856689 sshd[5605]: Accepted publickey for core from 10.0.0.1 port 51628 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:54:04.860202 sshd-session[5605]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:54:04.871127 systemd-logind[1617]: New session 17 of user core. Jan 14 13:54:04.858000 audit[5605]: CRED_ACQ pid=5605 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:04.900199 kernel: audit: type=1101 audit(1768398844.855:808): pid=5605 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:04.900315 kernel: audit: type=1103 audit(1768398844.858:809): pid=5605 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:04.900349 kernel: audit: type=1006 audit(1768398844.858:810): pid=5605 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=17 res=1 Jan 14 13:54:04.858000 audit[5605]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcb71de350 a2=3 a3=0 items=0 ppid=1 pid=5605 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=17 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:04.937073 kernel: audit: type=1300 audit(1768398844.858:810): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcb71de350 a2=3 a3=0 items=0 ppid=1 pid=5605 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=17 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:04.937238 kernel: audit: type=1327 audit(1768398844.858:810): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:54:04.858000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:54:04.949121 systemd[1]: Started session-17.scope - Session 17 of User core. Jan 14 13:54:04.954000 audit[5605]: USER_START pid=5605 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:04.959000 audit[5609]: CRED_ACQ pid=5609 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:04.985900 kernel: audit: type=1105 audit(1768398844.954:811): pid=5605 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:04.986151 kernel: audit: type=1103 audit(1768398844.959:812): pid=5609 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:05.205411 sshd[5609]: Connection closed by 10.0.0.1 port 51628 Jan 14 13:54:05.206721 sshd-session[5605]: pam_unix(sshd:session): session closed for user core Jan 14 13:54:05.208000 audit[5605]: USER_END pid=5605 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:05.209000 audit[5605]: CRED_DISP pid=5605 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:05.260371 kernel: audit: type=1106 audit(1768398845.208:813): pid=5605 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:05.260983 kernel: audit: type=1104 audit(1768398845.209:814): pid=5605 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:05.266723 systemd[1]: sshd@15-10.0.0.127:22-10.0.0.1:51628.service: Deactivated successfully. Jan 14 13:54:05.266000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.127:22-10.0.0.1:51628 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:05.271143 systemd[1]: session-17.scope: Deactivated successfully. Jan 14 13:54:05.273918 systemd-logind[1617]: Session 17 logged out. Waiting for processes to exit. Jan 14 13:54:05.277000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.127:22-10.0.0.1:51640 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:05.278163 systemd[1]: Started sshd@16-10.0.0.127:22-10.0.0.1:51640.service - OpenSSH per-connection server daemon (10.0.0.1:51640). Jan 14 13:54:05.280328 systemd-logind[1617]: Removed session 17. Jan 14 13:54:05.383408 containerd[1635]: time="2026-01-14T13:54:05.382806929Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 14 13:54:05.398000 audit[5624]: USER_ACCT pid=5624 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:05.402005 sshd[5624]: Accepted publickey for core from 10.0.0.1 port 51640 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:54:05.402000 audit[5624]: CRED_ACQ pid=5624 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:05.402000 audit[5624]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcba452330 a2=3 a3=0 items=0 ppid=1 pid=5624 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=18 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:05.402000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:54:05.405190 sshd-session[5624]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:54:05.415015 systemd-logind[1617]: New session 18 of user core. Jan 14 13:54:05.426914 systemd[1]: Started session-18.scope - Session 18 of User core. Jan 14 13:54:05.433000 audit[5624]: USER_START pid=5624 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:05.436000 audit[5628]: CRED_ACQ pid=5628 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:05.478972 containerd[1635]: time="2026-01-14T13:54:05.477269884Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:54:05.481874 containerd[1635]: time="2026-01-14T13:54:05.481432964Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 14 13:54:05.481874 containerd[1635]: time="2026-01-14T13:54:05.481851664Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 14 13:54:05.482438 kubelet[2951]: E0114 13:54:05.482395 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 14 13:54:05.483104 kubelet[2951]: E0114 13:54:05.482652 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 14 13:54:05.483104 kubelet[2951]: E0114 13:54:05.482736 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container whisker start failed in pod whisker-868d8bc54c-7mv9t_calico-system(0b696939-2eb2-437f-b107-24730b99d30f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 14 13:54:05.486397 containerd[1635]: time="2026-01-14T13:54:05.486237044Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 14 13:54:05.568723 containerd[1635]: time="2026-01-14T13:54:05.568355480Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:54:05.578242 containerd[1635]: time="2026-01-14T13:54:05.577315691Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 14 13:54:05.578242 containerd[1635]: time="2026-01-14T13:54:05.577760170Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 14 13:54:05.578373 kubelet[2951]: E0114 13:54:05.578157 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 14 13:54:05.578373 kubelet[2951]: E0114 13:54:05.578207 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 14 13:54:05.578373 kubelet[2951]: E0114 13:54:05.578287 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container whisker-backend start failed in pod whisker-868d8bc54c-7mv9t_calico-system(0b696939-2eb2-437f-b107-24730b99d30f): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 14 13:54:05.579244 kubelet[2951]: E0114 13:54:05.578682 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-868d8bc54c-7mv9t" podUID="0b696939-2eb2-437f-b107-24730b99d30f" Jan 14 13:54:05.860260 sshd[5628]: Connection closed by 10.0.0.1 port 51640 Jan 14 13:54:05.860064 sshd-session[5624]: pam_unix(sshd:session): session closed for user core Jan 14 13:54:05.863000 audit[5624]: USER_END pid=5624 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:05.863000 audit[5624]: CRED_DISP pid=5624 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:05.870785 systemd[1]: sshd@16-10.0.0.127:22-10.0.0.1:51640.service: Deactivated successfully. Jan 14 13:54:05.871000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.127:22-10.0.0.1:51640 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:05.874786 systemd[1]: session-18.scope: Deactivated successfully. Jan 14 13:54:05.878902 systemd-logind[1617]: Session 18 logged out. Waiting for processes to exit. Jan 14 13:54:05.883000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-10.0.0.127:22-10.0.0.1:51648 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:05.883393 systemd[1]: Started sshd@17-10.0.0.127:22-10.0.0.1:51648.service - OpenSSH per-connection server daemon (10.0.0.1:51648). Jan 14 13:54:05.884787 systemd-logind[1617]: Removed session 18. Jan 14 13:54:06.009000 audit[5640]: USER_ACCT pid=5640 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:06.010114 sshd[5640]: Accepted publickey for core from 10.0.0.1 port 51648 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:54:06.011000 audit[5640]: CRED_ACQ pid=5640 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:06.012000 audit[5640]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe39d23860 a2=3 a3=0 items=0 ppid=1 pid=5640 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=19 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:06.012000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:54:06.015103 sshd-session[5640]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:54:06.026212 systemd-logind[1617]: New session 19 of user core. Jan 14 13:54:06.039146 systemd[1]: Started session-19.scope - Session 19 of User core. Jan 14 13:54:06.044000 audit[5640]: USER_START pid=5640 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:06.048000 audit[5644]: CRED_ACQ pid=5644 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:06.866357 sshd[5644]: Connection closed by 10.0.0.1 port 51648 Jan 14 13:54:06.867838 sshd-session[5640]: pam_unix(sshd:session): session closed for user core Jan 14 13:54:06.871000 audit[5640]: USER_END pid=5640 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:06.871000 audit[5640]: CRED_DISP pid=5640 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:06.883057 systemd[1]: sshd@17-10.0.0.127:22-10.0.0.1:51648.service: Deactivated successfully. Jan 14 13:54:06.883000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-10.0.0.127:22-10.0.0.1:51648 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:06.887363 systemd[1]: session-19.scope: Deactivated successfully. Jan 14 13:54:06.890322 systemd-logind[1617]: Session 19 logged out. Waiting for processes to exit. Jan 14 13:54:06.896981 systemd-logind[1617]: Removed session 19. Jan 14 13:54:06.904000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.127:22-10.0.0.1:51656 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:06.905100 systemd[1]: Started sshd@18-10.0.0.127:22-10.0.0.1:51656.service - OpenSSH per-connection server daemon (10.0.0.1:51656). Jan 14 13:54:06.947000 audit[5657]: NETFILTER_CFG table=filter:138 family=2 entries=26 op=nft_register_rule pid=5657 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:54:06.947000 audit[5657]: SYSCALL arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7ffcd369d6f0 a2=0 a3=7ffcd369d6dc items=0 ppid=3117 pid=5657 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:06.947000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:54:06.962000 audit[5657]: NETFILTER_CFG table=nat:139 family=2 entries=20 op=nft_register_rule pid=5657 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:54:06.962000 audit[5657]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffcd369d6f0 a2=0 a3=0 items=0 ppid=3117 pid=5657 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:06.962000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:54:06.997000 audit[5662]: USER_ACCT pid=5662 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:06.998231 sshd[5662]: Accepted publickey for core from 10.0.0.1 port 51656 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:54:06.999000 audit[5662]: CRED_ACQ pid=5662 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:06.999000 audit[5662]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffde1717e10 a2=3 a3=0 items=0 ppid=1 pid=5662 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=20 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:06.999000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:54:07.001824 sshd-session[5662]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:54:07.012888 systemd-logind[1617]: New session 20 of user core. Jan 14 13:54:07.028079 systemd[1]: Started session-20.scope - Session 20 of User core. Jan 14 13:54:07.032000 audit[5662]: USER_START pid=5662 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:07.035000 audit[5666]: CRED_ACQ pid=5666 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:07.454706 sshd[5666]: Connection closed by 10.0.0.1 port 51656 Jan 14 13:54:07.454423 sshd-session[5662]: pam_unix(sshd:session): session closed for user core Jan 14 13:54:07.457000 audit[5662]: USER_END pid=5662 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:07.457000 audit[5662]: CRED_DISP pid=5662 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:07.467944 systemd[1]: sshd@18-10.0.0.127:22-10.0.0.1:51656.service: Deactivated successfully. Jan 14 13:54:07.467000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.127:22-10.0.0.1:51656 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:07.473439 systemd[1]: session-20.scope: Deactivated successfully. Jan 14 13:54:07.476143 systemd-logind[1617]: Session 20 logged out. Waiting for processes to exit. Jan 14 13:54:07.481000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.127:22-10.0.0.1:51672 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:07.482087 systemd[1]: Started sshd@19-10.0.0.127:22-10.0.0.1:51672.service - OpenSSH per-connection server daemon (10.0.0.1:51672). Jan 14 13:54:07.485809 systemd-logind[1617]: Removed session 20. Jan 14 13:54:07.582000 audit[5678]: USER_ACCT pid=5678 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:07.583414 sshd[5678]: Accepted publickey for core from 10.0.0.1 port 51672 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:54:07.584000 audit[5678]: CRED_ACQ pid=5678 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:07.584000 audit[5678]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc28cd8d90 a2=3 a3=0 items=0 ppid=1 pid=5678 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=21 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:07.584000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:54:07.587874 sshd-session[5678]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:54:07.600414 systemd-logind[1617]: New session 21 of user core. Jan 14 13:54:07.613118 systemd[1]: Started session-21.scope - Session 21 of User core. Jan 14 13:54:07.618000 audit[5678]: USER_START pid=5678 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:07.621000 audit[5682]: CRED_ACQ pid=5682 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:07.767903 sshd[5682]: Connection closed by 10.0.0.1 port 51672 Jan 14 13:54:07.768841 sshd-session[5678]: pam_unix(sshd:session): session closed for user core Jan 14 13:54:07.771000 audit[5678]: USER_END pid=5678 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:07.772000 audit[5678]: CRED_DISP pid=5678 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:07.775333 systemd[1]: sshd@19-10.0.0.127:22-10.0.0.1:51672.service: Deactivated successfully. Jan 14 13:54:07.775000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.127:22-10.0.0.1:51672 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:07.778384 systemd[1]: session-21.scope: Deactivated successfully. Jan 14 13:54:07.781086 systemd-logind[1617]: Session 21 logged out. Waiting for processes to exit. Jan 14 13:54:07.784404 systemd-logind[1617]: Removed session 21. Jan 14 13:54:07.993000 audit[5695]: NETFILTER_CFG table=filter:140 family=2 entries=38 op=nft_register_rule pid=5695 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:54:07.993000 audit[5695]: SYSCALL arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7ffd9966f790 a2=0 a3=7ffd9966f77c items=0 ppid=3117 pid=5695 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:07.993000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:54:08.007000 audit[5695]: NETFILTER_CFG table=nat:141 family=2 entries=20 op=nft_register_rule pid=5695 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:54:08.007000 audit[5695]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffd9966f790 a2=0 a3=0 items=0 ppid=3117 pid=5695 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:08.007000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:54:09.384998 containerd[1635]: time="2026-01-14T13:54:09.384756296Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 14 13:54:09.457874 containerd[1635]: time="2026-01-14T13:54:09.457445381Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:54:09.460152 containerd[1635]: time="2026-01-14T13:54:09.459992985Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 14 13:54:09.460152 containerd[1635]: time="2026-01-14T13:54:09.460127947Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 14 13:54:09.460356 kubelet[2951]: E0114 13:54:09.460234 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 14 13:54:09.460356 kubelet[2951]: E0114 13:54:09.460267 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 14 13:54:09.460894 kubelet[2951]: E0114 13:54:09.460408 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-csi start failed in pod csi-node-driver-pvxlq_calico-system(967b7513-2769-43e5-a113-9344f17909c3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 14 13:54:09.461701 containerd[1635]: time="2026-01-14T13:54:09.461302080Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 14 13:54:09.532791 containerd[1635]: time="2026-01-14T13:54:09.532748296Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:54:09.536881 containerd[1635]: time="2026-01-14T13:54:09.536768066Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 14 13:54:09.536955 containerd[1635]: time="2026-01-14T13:54:09.536885364Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 14 13:54:09.537048 kubelet[2951]: E0114 13:54:09.537016 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 14 13:54:09.537091 kubelet[2951]: E0114 13:54:09.537058 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 14 13:54:09.537214 kubelet[2951]: E0114 13:54:09.537189 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container goldmane start failed in pod goldmane-7c778bb748-8wp8q_calico-system(2b372f30-5f27-467a-8d6a-b220df9d05b0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 14 13:54:09.537302 kubelet[2951]: E0114 13:54:09.537282 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-8wp8q" podUID="2b372f30-5f27-467a-8d6a-b220df9d05b0" Jan 14 13:54:09.537776 containerd[1635]: time="2026-01-14T13:54:09.537756147Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 14 13:54:09.617086 containerd[1635]: time="2026-01-14T13:54:09.616906829Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:54:09.619734 containerd[1635]: time="2026-01-14T13:54:09.619293305Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 14 13:54:09.619734 containerd[1635]: time="2026-01-14T13:54:09.619380937Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 14 13:54:09.620091 kubelet[2951]: E0114 13:54:09.619970 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 13:54:09.620091 kubelet[2951]: E0114 13:54:09.620018 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 13:54:09.620203 kubelet[2951]: E0114 13:54:09.620183 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-apiserver start failed in pod calico-apiserver-6cbf5bf74d-k9nzg_calico-apiserver(539656bb-a7f6-491c-9c24-e13907f6a8fd): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 14 13:54:09.620260 kubelet[2951]: E0114 13:54:09.620218 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" podUID="539656bb-a7f6-491c-9c24-e13907f6a8fd" Jan 14 13:54:09.621073 containerd[1635]: time="2026-01-14T13:54:09.620995394Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 14 13:54:09.698322 containerd[1635]: time="2026-01-14T13:54:09.697919577Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:54:09.701114 containerd[1635]: time="2026-01-14T13:54:09.700948275Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 14 13:54:09.701272 containerd[1635]: time="2026-01-14T13:54:09.701017367Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 14 13:54:09.701807 kubelet[2951]: E0114 13:54:09.701406 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 14 13:54:09.701894 kubelet[2951]: E0114 13:54:09.701850 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 14 13:54:09.702172 kubelet[2951]: E0114 13:54:09.702020 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container csi-node-driver-registrar start failed in pod csi-node-driver-pvxlq_calico-system(967b7513-2769-43e5-a113-9344f17909c3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 14 13:54:09.702172 kubelet[2951]: E0114 13:54:09.702145 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:54:12.787802 systemd[1]: Started sshd@20-10.0.0.127:22-10.0.0.1:51674.service - OpenSSH per-connection server daemon (10.0.0.1:51674). Jan 14 13:54:12.787000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.127:22-10.0.0.1:51674 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:12.793912 kernel: kauditd_printk_skb: 57 callbacks suppressed Jan 14 13:54:12.794079 kernel: audit: type=1130 audit(1768398852.787:856): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.127:22-10.0.0.1:51674 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:12.897000 audit[5699]: USER_ACCT pid=5699 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:12.899769 sshd[5699]: Accepted publickey for core from 10.0.0.1 port 51674 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:54:12.903776 sshd-session[5699]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:54:12.914172 systemd-logind[1617]: New session 22 of user core. Jan 14 13:54:12.900000 audit[5699]: CRED_ACQ pid=5699 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:12.942894 kernel: audit: type=1101 audit(1768398852.897:857): pid=5699 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:12.942961 kernel: audit: type=1103 audit(1768398852.900:858): pid=5699 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:12.956363 kernel: audit: type=1006 audit(1768398852.901:859): pid=5699 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=22 res=1 Jan 14 13:54:12.956416 kernel: audit: type=1300 audit(1768398852.901:859): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe1fe5fc00 a2=3 a3=0 items=0 ppid=1 pid=5699 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=22 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:12.901000 audit[5699]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe1fe5fc00 a2=3 a3=0 items=0 ppid=1 pid=5699 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=22 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:12.981759 kernel: audit: type=1327 audit(1768398852.901:859): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:54:12.901000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:54:12.993049 systemd[1]: Started session-22.scope - Session 22 of User core. Jan 14 13:54:13.001000 audit[5699]: USER_START pid=5699 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:13.036795 kernel: audit: type=1105 audit(1768398853.001:860): pid=5699 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:13.004000 audit[5703]: CRED_ACQ pid=5703 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:13.059931 kernel: audit: type=1103 audit(1768398853.004:861): pid=5703 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:13.185310 sshd[5703]: Connection closed by 10.0.0.1 port 51674 Jan 14 13:54:13.186314 sshd-session[5699]: pam_unix(sshd:session): session closed for user core Jan 14 13:54:13.189000 audit[5699]: USER_END pid=5699 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:13.194338 systemd[1]: sshd@20-10.0.0.127:22-10.0.0.1:51674.service: Deactivated successfully. Jan 14 13:54:13.199302 systemd[1]: session-22.scope: Deactivated successfully. Jan 14 13:54:13.201891 systemd-logind[1617]: Session 22 logged out. Waiting for processes to exit. Jan 14 13:54:13.206342 systemd-logind[1617]: Removed session 22. Jan 14 13:54:13.189000 audit[5699]: CRED_DISP pid=5699 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:13.243188 kernel: audit: type=1106 audit(1768398853.189:862): pid=5699 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:13.243340 kernel: audit: type=1104 audit(1768398853.189:863): pid=5699 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:13.194000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.127:22-10.0.0.1:51674 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:13.385722 containerd[1635]: time="2026-01-14T13:54:13.385689101Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 14 13:54:13.487756 containerd[1635]: time="2026-01-14T13:54:13.487213197Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:54:13.490849 containerd[1635]: time="2026-01-14T13:54:13.490752952Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 14 13:54:13.490849 containerd[1635]: time="2026-01-14T13:54:13.490825167Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 14 13:54:13.491176 kubelet[2951]: E0114 13:54:13.491143 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 14 13:54:13.492852 kubelet[2951]: E0114 13:54:13.492206 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 14 13:54:13.492852 kubelet[2951]: E0114 13:54:13.492307 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-kube-controllers start failed in pod calico-kube-controllers-6cdc99f4d-mck2c_calico-system(b158b3fd-5a6b-446c-92b0-d1c78d153ee9): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 14 13:54:13.492852 kubelet[2951]: E0114 13:54:13.492338 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-6cdc99f4d-mck2c" podUID="b158b3fd-5a6b-446c-92b0-d1c78d153ee9" Jan 14 13:54:13.553000 audit[5718]: NETFILTER_CFG table=filter:142 family=2 entries=26 op=nft_register_rule pid=5718 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:54:13.553000 audit[5718]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7fff586c6430 a2=0 a3=7fff586c641c items=0 ppid=3117 pid=5718 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:13.553000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:54:13.594000 audit[5718]: NETFILTER_CFG table=nat:143 family=2 entries=104 op=nft_register_chain pid=5718 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 13:54:13.594000 audit[5718]: SYSCALL arch=c000003e syscall=46 success=yes exit=48684 a0=3 a1=7fff586c6430 a2=0 a3=7fff586c641c items=0 ppid=3117 pid=5718 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:13.594000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 13:54:16.382976 containerd[1635]: time="2026-01-14T13:54:16.382926329Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 14 13:54:16.485675 containerd[1635]: time="2026-01-14T13:54:16.485241830Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 13:54:16.487742 containerd[1635]: time="2026-01-14T13:54:16.487691314Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 14 13:54:16.487802 containerd[1635]: time="2026-01-14T13:54:16.487770292Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 14 13:54:16.488322 kubelet[2951]: E0114 13:54:16.488184 2951 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 13:54:16.488322 kubelet[2951]: E0114 13:54:16.488307 2951 kuberuntime_image.go:43] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 13:54:16.488899 kubelet[2951]: E0114 13:54:16.488395 2951 kuberuntime_manager.go:1449] "Unhandled Error" err="container calico-apiserver start failed in pod calico-apiserver-6cbf5bf74d-6k99t_calico-apiserver(3a917c9a-ea97-482d-9194-bbf322e47535): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 14 13:54:16.488899 kubelet[2951]: E0114 13:54:16.488437 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-6k99t" podUID="3a917c9a-ea97-482d-9194-bbf322e47535" Jan 14 13:54:18.202434 systemd[1]: Started sshd@21-10.0.0.127:22-10.0.0.1:53848.service - OpenSSH per-connection server daemon (10.0.0.1:53848). Jan 14 13:54:18.202000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.127:22-10.0.0.1:53848 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:18.210170 kernel: kauditd_printk_skb: 7 callbacks suppressed Jan 14 13:54:18.210255 kernel: audit: type=1130 audit(1768398858.202:867): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.127:22-10.0.0.1:53848 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:18.326000 audit[5720]: USER_ACCT pid=5720 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:18.334187 sshd-session[5720]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:54:18.350805 sshd[5720]: Accepted publickey for core from 10.0.0.1 port 53848 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:54:18.328000 audit[5720]: CRED_ACQ pid=5720 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:18.375923 systemd-logind[1617]: New session 23 of user core. Jan 14 13:54:18.380806 kernel: audit: type=1101 audit(1768398858.326:868): pid=5720 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:18.380863 kernel: audit: type=1103 audit(1768398858.328:869): pid=5720 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:18.391414 kubelet[2951]: E0114 13:54:18.388078 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-868d8bc54c-7mv9t" podUID="0b696939-2eb2-437f-b107-24730b99d30f" Jan 14 13:54:18.399046 kernel: audit: type=1006 audit(1768398858.328:870): pid=5720 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=23 res=1 Jan 14 13:54:18.399117 kernel: audit: type=1300 audit(1768398858.328:870): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc1b87b950 a2=3 a3=0 items=0 ppid=1 pid=5720 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=23 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:18.328000 audit[5720]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc1b87b950 a2=3 a3=0 items=0 ppid=1 pid=5720 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=23 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:18.328000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:54:18.427103 systemd[1]: Started session-23.scope - Session 23 of User core. Jan 14 13:54:18.435691 kernel: audit: type=1327 audit(1768398858.328:870): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:54:18.434000 audit[5720]: USER_START pid=5720 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:18.462315 kernel: audit: type=1105 audit(1768398858.434:871): pid=5720 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:18.462388 kernel: audit: type=1103 audit(1768398858.439:872): pid=5730 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:18.439000 audit[5730]: CRED_ACQ pid=5730 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:18.595233 sshd[5730]: Connection closed by 10.0.0.1 port 53848 Jan 14 13:54:18.595806 sshd-session[5720]: pam_unix(sshd:session): session closed for user core Jan 14 13:54:18.597000 audit[5720]: USER_END pid=5720 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:18.601132 systemd[1]: sshd@21-10.0.0.127:22-10.0.0.1:53848.service: Deactivated successfully. Jan 14 13:54:18.604837 systemd[1]: session-23.scope: Deactivated successfully. Jan 14 13:54:18.607792 systemd-logind[1617]: Session 23 logged out. Waiting for processes to exit. Jan 14 13:54:18.611436 systemd-logind[1617]: Removed session 23. Jan 14 13:54:18.597000 audit[5720]: CRED_DISP pid=5720 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:18.650404 kernel: audit: type=1106 audit(1768398858.597:873): pid=5720 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:18.650783 kernel: audit: type=1104 audit(1768398858.597:874): pid=5720 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:18.601000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.127:22-10.0.0.1:53848 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:20.584263 kubelet[2951]: E0114 13:54:20.584232 2951 dns.go:154] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 14 13:54:21.380432 kubelet[2951]: E0114 13:54:21.380247 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-6cbf5bf74d-k9nzg" podUID="539656bb-a7f6-491c-9c24-e13907f6a8fd" Jan 14 13:54:22.383690 kubelet[2951]: E0114 13:54:22.383042 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-7c778bb748-8wp8q" podUID="2b372f30-5f27-467a-8d6a-b220df9d05b0" Jan 14 13:54:23.384027 kubelet[2951]: E0114 13:54:23.383809 2951 pod_workers.go:1324] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-pvxlq" podUID="967b7513-2769-43e5-a113-9344f17909c3" Jan 14 13:54:23.613264 systemd[1]: Started sshd@22-10.0.0.127:22-10.0.0.1:53854.service - OpenSSH per-connection server daemon (10.0.0.1:53854). Jan 14 13:54:23.612000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.127:22-10.0.0.1:53854 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:23.618717 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 14 13:54:23.618821 kernel: audit: type=1130 audit(1768398863.612:876): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.127:22-10.0.0.1:53854 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 13:54:23.709000 audit[5777]: USER_ACCT pid=5777 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:23.711333 sshd[5777]: Accepted publickey for core from 10.0.0.1 port 53854 ssh2: RSA SHA256:6ImhlCg2Y75dQ4DnaE2aO9dHLur/A4YXKF0wGnkswcQ Jan 14 13:54:23.713834 sshd-session[5777]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 13:54:23.722912 systemd-logind[1617]: New session 24 of user core. Jan 14 13:54:23.711000 audit[5777]: CRED_ACQ pid=5777 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:23.757108 kernel: audit: type=1101 audit(1768398863.709:877): pid=5777 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:23.757694 kernel: audit: type=1103 audit(1768398863.711:878): pid=5777 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:23.711000 audit[5777]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcb2d47e80 a2=3 a3=0 items=0 ppid=1 pid=5777 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=24 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:23.797945 kernel: audit: type=1006 audit(1768398863.711:879): pid=5777 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=24 res=1 Jan 14 13:54:23.798010 kernel: audit: type=1300 audit(1768398863.711:879): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcb2d47e80 a2=3 a3=0 items=0 ppid=1 pid=5777 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=24 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 13:54:23.798144 kernel: audit: type=1327 audit(1768398863.711:879): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:54:23.711000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 13:54:23.799057 systemd[1]: Started session-24.scope - Session 24 of User core. Jan 14 13:54:23.803000 audit[5777]: USER_START pid=5777 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:23.837444 kernel: audit: type=1105 audit(1768398863.803:880): pid=5777 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:23.807000 audit[5781]: CRED_ACQ pid=5781 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:23.862924 kernel: audit: type=1103 audit(1768398863.807:881): pid=5781 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:23.974840 sshd[5781]: Connection closed by 10.0.0.1 port 53854 Jan 14 13:54:23.975083 sshd-session[5777]: pam_unix(sshd:session): session closed for user core Jan 14 13:54:23.978000 audit[5777]: USER_END pid=5777 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:23.984304 systemd[1]: sshd@22-10.0.0.127:22-10.0.0.1:53854.service: Deactivated successfully. Jan 14 13:54:23.988892 systemd[1]: session-24.scope: Deactivated successfully. Jan 14 13:54:23.991429 systemd-logind[1617]: Session 24 logged out. Waiting for processes to exit. Jan 14 13:54:23.994299 systemd-logind[1617]: Removed session 24. Jan 14 13:54:24.012976 kernel: audit: type=1106 audit(1768398863.978:882): pid=5777 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:24.013027 kernel: audit: type=1104 audit(1768398863.978:883): pid=5777 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:23.978000 audit[5777]: CRED_DISP pid=5777 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 14 13:54:23.984000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.127:22-10.0.0.1:53854 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'