Jan 24 11:47:45.865591 kernel: Linux version 6.12.66-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.1_p20250801 p4) 14.3.1 20250801, GNU ld (Gentoo 2.45 p3) 2.45.0) #1 SMP PREEMPT_DYNAMIC Sat Jan 24 09:07:34 -00 2026 Jan 24 11:47:45.865889 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=7953d3c7acaad6ee79638a10c67ea9f0b3a8597919989b6fbf2f9a1742d4ba63 Jan 24 11:47:45.865963 kernel: BIOS-provided physical RAM map: Jan 24 11:47:45.865975 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Jan 24 11:47:45.865984 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Jan 24 11:47:45.865992 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Jan 24 11:47:45.866002 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable Jan 24 11:47:45.866011 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved Jan 24 11:47:45.866052 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Jan 24 11:47:45.866066 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Jan 24 11:47:45.866115 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Jan 24 11:47:45.866125 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Jan 24 11:47:45.866134 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Jan 24 11:47:45.866143 kernel: NX (Execute Disable) protection: active Jan 24 11:47:45.868539 kernel: APIC: Static calls initialized Jan 24 11:47:45.868677 kernel: SMBIOS 2.8 present. Jan 24 11:47:45.869623 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014 Jan 24 11:47:45.869669 kernel: DMI: Memory slots populated: 1/1 Jan 24 11:47:45.869679 kernel: Hypervisor detected: KVM Jan 24 11:47:45.869725 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Jan 24 11:47:45.869734 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Jan 24 11:47:45.869744 kernel: kvm-clock: using sched offset of 13608453452 cycles Jan 24 11:47:45.869783 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Jan 24 11:47:45.869796 kernel: tsc: Detected 2445.426 MHz processor Jan 24 11:47:45.869919 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jan 24 11:47:45.869961 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jan 24 11:47:45.870047 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Jan 24 11:47:45.870089 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Jan 24 11:47:45.870124 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jan 24 11:47:45.870134 kernel: Using GB pages for direct mapping Jan 24 11:47:45.870146 kernel: ACPI: Early table checksum verification disabled Jan 24 11:47:45.870194 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS ) Jan 24 11:47:45.870205 kernel: ACPI: RSDT 0x000000009CFE241A 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 24 11:47:45.870215 kernel: ACPI: FACP 0x000000009CFE21FA 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Jan 24 11:47:45.870225 kernel: ACPI: DSDT 0x000000009CFE0040 0021BA (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 24 11:47:45.870235 kernel: ACPI: FACS 0x000000009CFE0000 000040 Jan 24 11:47:45.870248 kernel: ACPI: APIC 0x000000009CFE22EE 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 24 11:47:45.872069 kernel: ACPI: HPET 0x000000009CFE237E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 24 11:47:45.872206 kernel: ACPI: MCFG 0x000000009CFE23B6 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 24 11:47:45.872219 kernel: ACPI: WAET 0x000000009CFE23F2 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 24 11:47:45.872270 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21fa-0x9cfe22ed] Jan 24 11:47:45.872282 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21f9] Jan 24 11:47:45.872293 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f] Jan 24 11:47:45.872377 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22ee-0x9cfe237d] Jan 24 11:47:45.872515 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe237e-0x9cfe23b5] Jan 24 11:47:45.872527 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23b6-0x9cfe23f1] Jan 24 11:47:45.872539 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23f2-0x9cfe2419] Jan 24 11:47:45.872550 kernel: No NUMA configuration found Jan 24 11:47:45.872561 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff] Jan 24 11:47:45.872572 kernel: NODE_DATA(0) allocated [mem 0x9cfd4dc0-0x9cfdbfff] Jan 24 11:47:45.872618 kernel: Zone ranges: Jan 24 11:47:45.872629 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jan 24 11:47:45.872641 kernel: DMA32 [mem 0x0000000001000000-0x000000009cfdbfff] Jan 24 11:47:45.872652 kernel: Normal empty Jan 24 11:47:45.872663 kernel: Device empty Jan 24 11:47:45.872673 kernel: Movable zone start for each node Jan 24 11:47:45.872719 kernel: Early memory node ranges Jan 24 11:47:45.872767 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Jan 24 11:47:45.872778 kernel: node 0: [mem 0x0000000000100000-0x000000009cfdbfff] Jan 24 11:47:45.872790 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff] Jan 24 11:47:45.872829 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jan 24 11:47:45.872841 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Jan 24 11:47:45.872870 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Jan 24 11:47:45.872882 kernel: ACPI: PM-Timer IO Port: 0x608 Jan 24 11:47:45.872893 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Jan 24 11:47:45.872941 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Jan 24 11:47:45.872953 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Jan 24 11:47:45.872986 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Jan 24 11:47:45.872999 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Jan 24 11:47:45.873011 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Jan 24 11:47:45.873022 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Jan 24 11:47:45.873033 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jan 24 11:47:45.873078 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Jan 24 11:47:45.873091 kernel: TSC deadline timer available Jan 24 11:47:45.873102 kernel: CPU topo: Max. logical packages: 1 Jan 24 11:47:45.873112 kernel: CPU topo: Max. logical dies: 1 Jan 24 11:47:45.873123 kernel: CPU topo: Max. dies per package: 1 Jan 24 11:47:45.873132 kernel: CPU topo: Max. threads per core: 1 Jan 24 11:47:45.873143 kernel: CPU topo: Num. cores per package: 4 Jan 24 11:47:45.873153 kernel: CPU topo: Num. threads per package: 4 Jan 24 11:47:45.873209 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs Jan 24 11:47:45.873219 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Jan 24 11:47:45.873230 kernel: kvm-guest: KVM setup pv remote TLB flush Jan 24 11:47:45.873240 kernel: kvm-guest: setup PV sched yield Jan 24 11:47:45.873251 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Jan 24 11:47:45.873263 kernel: Booting paravirtualized kernel on KVM Jan 24 11:47:45.873274 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jan 24 11:47:45.873329 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Jan 24 11:47:45.873342 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 Jan 24 11:47:45.873355 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 Jan 24 11:47:45.873367 kernel: pcpu-alloc: [0] 0 1 2 3 Jan 24 11:47:45.873381 kernel: kvm-guest: PV spinlocks enabled Jan 24 11:47:45.873396 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Jan 24 11:47:45.873409 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=7953d3c7acaad6ee79638a10c67ea9f0b3a8597919989b6fbf2f9a1742d4ba63 Jan 24 11:47:45.874423 kernel: random: crng init done Jan 24 11:47:45.874440 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jan 24 11:47:45.874513 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 24 11:47:45.874527 kernel: Fallback order for Node 0: 0 Jan 24 11:47:45.874535 kernel: Built 1 zonelists, mobility grouping on. Total pages: 642938 Jan 24 11:47:45.874543 kernel: Policy zone: DMA32 Jan 24 11:47:45.874551 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 24 11:47:45.874663 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Jan 24 11:47:45.874671 kernel: ftrace: allocating 40128 entries in 157 pages Jan 24 11:47:45.874678 kernel: ftrace: allocated 157 pages with 5 groups Jan 24 11:47:45.874711 kernel: Dynamic Preempt: voluntary Jan 24 11:47:45.874719 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 24 11:47:45.874728 kernel: rcu: RCU event tracing is enabled. Jan 24 11:47:45.874736 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Jan 24 11:47:45.874770 kernel: Trampoline variant of Tasks RCU enabled. Jan 24 11:47:45.874794 kernel: Rude variant of Tasks RCU enabled. Jan 24 11:47:45.874802 kernel: Tracing variant of Tasks RCU enabled. Jan 24 11:47:45.874810 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 24 11:47:45.874817 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Jan 24 11:47:45.874825 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 24 11:47:45.874833 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 24 11:47:45.874861 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Jan 24 11:47:45.874869 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Jan 24 11:47:45.874877 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jan 24 11:47:45.874940 kernel: Console: colour VGA+ 80x25 Jan 24 11:47:45.874967 kernel: printk: legacy console [ttyS0] enabled Jan 24 11:47:45.874974 kernel: ACPI: Core revision 20240827 Jan 24 11:47:45.874982 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Jan 24 11:47:45.874990 kernel: APIC: Switch to symmetric I/O mode setup Jan 24 11:47:45.874998 kernel: x2apic enabled Jan 24 11:47:45.875045 kernel: APIC: Switched APIC routing to: physical x2apic Jan 24 11:47:45.875904 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Jan 24 11:47:45.875928 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Jan 24 11:47:45.875942 kernel: kvm-guest: setup PV IPIs Jan 24 11:47:45.876079 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jan 24 11:47:45.876093 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x233fd7ba1b0, max_idle_ns: 440795295779 ns Jan 24 11:47:45.876104 kernel: Calibrating delay loop (skipped) preset value.. 4890.85 BogoMIPS (lpj=2445426) Jan 24 11:47:45.876115 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Jan 24 11:47:45.876129 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Jan 24 11:47:45.876144 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Jan 24 11:47:45.876155 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jan 24 11:47:45.876208 kernel: Spectre V2 : Mitigation: Retpolines Jan 24 11:47:45.876221 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Jan 24 11:47:45.876233 kernel: Speculative Store Bypass: Vulnerable Jan 24 11:47:45.876245 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Jan 24 11:47:45.876259 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Jan 24 11:47:45.876273 kernel: active return thunk: srso_alias_return_thunk Jan 24 11:47:45.876288 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Jan 24 11:47:45.877610 kernel: Transient Scheduler Attacks: Forcing mitigation on in a VM Jan 24 11:47:45.877626 kernel: Transient Scheduler Attacks: Vulnerable: Clear CPU buffers attempted, no microcode Jan 24 11:47:45.877638 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jan 24 11:47:45.877650 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jan 24 11:47:45.877662 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jan 24 11:47:45.877676 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jan 24 11:47:45.877723 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Jan 24 11:47:45.877766 kernel: Freeing SMP alternatives memory: 32K Jan 24 11:47:45.877774 kernel: pid_max: default: 32768 minimum: 301 Jan 24 11:47:45.877782 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Jan 24 11:47:45.877790 kernel: landlock: Up and running. Jan 24 11:47:45.877798 kernel: SELinux: Initializing. Jan 24 11:47:45.877855 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 24 11:47:45.877863 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 24 11:47:45.877910 kernel: smpboot: CPU0: AMD EPYC 7763 64-Core Processor (family: 0x19, model: 0x1, stepping: 0x1) Jan 24 11:47:45.877919 kernel: Performance Events: PMU not available due to virtualization, using software events only. Jan 24 11:47:45.877927 kernel: signal: max sigframe size: 1776 Jan 24 11:47:45.877935 kernel: rcu: Hierarchical SRCU implementation. Jan 24 11:47:45.877943 kernel: rcu: Max phase no-delay instances is 400. Jan 24 11:47:45.877951 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Jan 24 11:47:45.877959 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Jan 24 11:47:45.877988 kernel: smp: Bringing up secondary CPUs ... Jan 24 11:47:45.877996 kernel: smpboot: x86: Booting SMP configuration: Jan 24 11:47:45.878004 kernel: .... node #0, CPUs: #1 #2 #3 Jan 24 11:47:45.878011 kernel: smp: Brought up 1 node, 4 CPUs Jan 24 11:47:45.878019 kernel: smpboot: Total of 4 processors activated (19563.40 BogoMIPS) Jan 24 11:47:45.878032 kernel: Memory: 2445296K/2571752K available (14336K kernel code, 2445K rwdata, 31644K rodata, 15536K init, 2500K bss, 120520K reserved, 0K cma-reserved) Jan 24 11:47:45.878047 kernel: devtmpfs: initialized Jan 24 11:47:45.878095 kernel: x86/mm: Memory block size: 128MB Jan 24 11:47:45.878107 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 24 11:47:45.878118 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Jan 24 11:47:45.878129 kernel: pinctrl core: initialized pinctrl subsystem Jan 24 11:47:45.878144 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 24 11:47:45.878157 kernel: audit: initializing netlink subsys (disabled) Jan 24 11:47:45.878168 kernel: audit: type=2000 audit(1769255249.876:1): state=initialized audit_enabled=0 res=1 Jan 24 11:47:45.878219 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 24 11:47:45.878233 kernel: thermal_sys: Registered thermal governor 'user_space' Jan 24 11:47:45.878247 kernel: cpuidle: using governor menu Jan 24 11:47:45.878288 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 24 11:47:45.878300 kernel: dca service started, version 1.12.1 Jan 24 11:47:45.878311 kernel: PCI: ECAM [mem 0xb0000000-0xbfffffff] (base 0xb0000000) for domain 0000 [bus 00-ff] Jan 24 11:47:45.878326 kernel: PCI: ECAM [mem 0xb0000000-0xbfffffff] reserved as E820 entry Jan 24 11:47:45.880064 kernel: PCI: Using configuration type 1 for base access Jan 24 11:47:45.880266 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jan 24 11:47:45.880281 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jan 24 11:47:45.880292 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jan 24 11:47:45.882066 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 24 11:47:45.882092 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jan 24 11:47:45.882104 kernel: ACPI: Added _OSI(Module Device) Jan 24 11:47:45.882281 kernel: ACPI: Added _OSI(Processor Device) Jan 24 11:47:45.882294 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 24 11:47:45.882305 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 24 11:47:45.882317 kernel: ACPI: Interpreter enabled Jan 24 11:47:45.882331 kernel: ACPI: PM: (supports S0 S3 S5) Jan 24 11:47:45.882343 kernel: ACPI: Using IOAPIC for interrupt routing Jan 24 11:47:45.882354 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jan 24 11:47:45.882410 kernel: PCI: Using E820 reservations for host bridge windows Jan 24 11:47:45.882424 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Jan 24 11:47:45.882436 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jan 24 11:47:45.883784 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jan 24 11:47:45.884108 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Jan 24 11:47:45.884404 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Jan 24 11:47:45.884531 kernel: PCI host bridge to bus 0000:00 Jan 24 11:47:45.884929 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jan 24 11:47:45.885210 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Jan 24 11:47:45.885527 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jan 24 11:47:45.885775 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Jan 24 11:47:45.885975 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Jan 24 11:47:45.886291 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Jan 24 11:47:45.886597 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jan 24 11:47:45.888201 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Jan 24 11:47:45.888602 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Jan 24 11:47:45.889000 kernel: pci 0000:00:01.0: BAR 0 [mem 0xfd000000-0xfdffffff pref] Jan 24 11:47:45.889341 kernel: pci 0000:00:01.0: BAR 2 [mem 0xfebd0000-0xfebd0fff] Jan 24 11:47:45.889743 kernel: pci 0000:00:01.0: ROM [mem 0xfebc0000-0xfebcffff pref] Jan 24 11:47:45.890034 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jan 24 11:47:45.890382 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint Jan 24 11:47:45.890791 kernel: pci 0000:00:02.0: BAR 0 [io 0xc0c0-0xc0df] Jan 24 11:47:45.891076 kernel: pci 0000:00:02.0: BAR 1 [mem 0xfebd1000-0xfebd1fff] Jan 24 11:47:45.891417 kernel: pci 0000:00:02.0: BAR 4 [mem 0xfe000000-0xfe003fff 64bit pref] Jan 24 11:47:45.891828 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint Jan 24 11:47:45.892111 kernel: pci 0000:00:03.0: BAR 0 [io 0xc000-0xc07f] Jan 24 11:47:45.892393 kernel: pci 0000:00:03.0: BAR 1 [mem 0xfebd2000-0xfebd2fff] Jan 24 11:47:45.892870 kernel: pci 0000:00:03.0: BAR 4 [mem 0xfe004000-0xfe007fff 64bit pref] Jan 24 11:47:45.893169 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Jan 24 11:47:45.893575 kernel: pci 0000:00:04.0: BAR 0 [io 0xc0e0-0xc0ff] Jan 24 11:47:45.893914 kernel: pci 0000:00:04.0: BAR 1 [mem 0xfebd3000-0xfebd3fff] Jan 24 11:47:45.894171 kernel: pci 0000:00:04.0: BAR 4 [mem 0xfe008000-0xfe00bfff 64bit pref] Jan 24 11:47:45.894502 kernel: pci 0000:00:04.0: ROM [mem 0xfeb80000-0xfebbffff pref] Jan 24 11:47:45.894775 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Jan 24 11:47:45.895037 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Jan 24 11:47:45.895317 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Jan 24 11:47:45.895605 kernel: pci 0000:00:1f.2: BAR 4 [io 0xc100-0xc11f] Jan 24 11:47:45.895865 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xfebd4000-0xfebd4fff] Jan 24 11:47:45.896112 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Jan 24 11:47:45.896396 kernel: pci 0000:00:1f.3: BAR 4 [io 0x0700-0x073f] Jan 24 11:47:45.896494 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Jan 24 11:47:45.896504 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Jan 24 11:47:45.896512 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jan 24 11:47:45.896520 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Jan 24 11:47:45.896528 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Jan 24 11:47:45.896536 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Jan 24 11:47:45.896544 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Jan 24 11:47:45.896578 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Jan 24 11:47:45.896586 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Jan 24 11:47:45.896594 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Jan 24 11:47:45.896602 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Jan 24 11:47:45.896610 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Jan 24 11:47:45.896618 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Jan 24 11:47:45.896625 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Jan 24 11:47:45.896655 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Jan 24 11:47:45.896663 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Jan 24 11:47:45.896671 kernel: iommu: Default domain type: Translated Jan 24 11:47:45.896679 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jan 24 11:47:45.896713 kernel: PCI: Using ACPI for IRQ routing Jan 24 11:47:45.896721 kernel: PCI: pci_cache_line_size set to 64 bytes Jan 24 11:47:45.896730 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Jan 24 11:47:45.896759 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff] Jan 24 11:47:45.896979 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Jan 24 11:47:45.897243 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Jan 24 11:47:45.897523 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jan 24 11:47:45.897537 kernel: vgaarb: loaded Jan 24 11:47:45.897546 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Jan 24 11:47:45.897554 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Jan 24 11:47:45.897598 kernel: clocksource: Switched to clocksource kvm-clock Jan 24 11:47:45.897607 kernel: VFS: Disk quotas dquot_6.6.0 Jan 24 11:47:45.897615 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 24 11:47:45.897623 kernel: pnp: PnP ACPI init Jan 24 11:47:45.897927 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Jan 24 11:47:45.897942 kernel: pnp: PnP ACPI: found 6 devices Jan 24 11:47:45.897951 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jan 24 11:47:45.897987 kernel: NET: Registered PF_INET protocol family Jan 24 11:47:45.897995 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jan 24 11:47:45.898003 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jan 24 11:47:45.898011 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 24 11:47:45.898019 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 24 11:47:45.898033 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jan 24 11:47:45.898048 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jan 24 11:47:45.898098 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 24 11:47:45.898110 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 24 11:47:45.898121 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 24 11:47:45.898132 kernel: NET: Registered PF_XDP protocol family Jan 24 11:47:45.898403 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Jan 24 11:47:45.898779 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Jan 24 11:47:45.899094 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Jan 24 11:47:45.899350 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Jan 24 11:47:45.899727 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Jan 24 11:47:45.900005 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Jan 24 11:47:45.900021 kernel: PCI: CLS 0 bytes, default 64 Jan 24 11:47:45.900029 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x233fd7ba1b0, max_idle_ns: 440795295779 ns Jan 24 11:47:45.900038 kernel: Initialise system trusted keyrings Jan 24 11:47:45.900089 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jan 24 11:47:45.900102 kernel: Key type asymmetric registered Jan 24 11:47:45.900113 kernel: Asymmetric key parser 'x509' registered Jan 24 11:47:45.900124 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Jan 24 11:47:45.900135 kernel: io scheduler mq-deadline registered Jan 24 11:47:45.900146 kernel: io scheduler kyber registered Jan 24 11:47:45.900158 kernel: io scheduler bfq registered Jan 24 11:47:45.900234 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jan 24 11:47:45.900249 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Jan 24 11:47:45.900263 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Jan 24 11:47:45.900277 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Jan 24 11:47:45.900286 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 24 11:47:45.900300 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jan 24 11:47:45.900313 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Jan 24 11:47:45.900324 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jan 24 11:47:45.900380 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jan 24 11:47:45.900817 kernel: rtc_cmos 00:04: RTC can wake from S4 Jan 24 11:47:45.900838 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jan 24 11:47:45.901109 kernel: rtc_cmos 00:04: registered as rtc0 Jan 24 11:47:45.901382 kernel: rtc_cmos 00:04: setting system clock to 2026-01-24T11:47:34 UTC (1769255254) Jan 24 11:47:45.901769 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Jan 24 11:47:45.901838 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Jan 24 11:47:45.901851 kernel: NET: Registered PF_INET6 protocol family Jan 24 11:47:45.901864 kernel: Segment Routing with IPv6 Jan 24 11:47:45.901876 kernel: In-situ OAM (IOAM) with IPv6 Jan 24 11:47:45.901888 kernel: NET: Registered PF_PACKET protocol family Jan 24 11:47:45.901900 kernel: Key type dns_resolver registered Jan 24 11:47:45.901912 kernel: IPI shorthand broadcast: enabled Jan 24 11:47:45.901961 kernel: sched_clock: Marking stable (4238027943, 604238517)->(5567752695, -725486235) Jan 24 11:47:45.901974 kernel: registered taskstats version 1 Jan 24 11:47:45.901986 kernel: Loading compiled-in X.509 certificates Jan 24 11:47:45.901998 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.66-flatcar: a97c6138cc1b5c46f82656a7e055bcfc44b38b5c' Jan 24 11:47:45.902010 kernel: Demotion targets for Node 0: null Jan 24 11:47:45.902021 kernel: Key type .fscrypt registered Jan 24 11:47:45.902035 kernel: Key type fscrypt-provisioning registered Jan 24 11:47:45.902090 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 24 11:47:45.902102 kernel: ima: Allocated hash algorithm: sha1 Jan 24 11:47:45.902114 kernel: ima: No architecture policies found Jan 24 11:47:45.902127 kernel: clk: Disabling unused clocks Jan 24 11:47:45.902141 kernel: Freeing unused kernel image (initmem) memory: 15536K Jan 24 11:47:45.902153 kernel: Write protecting the kernel read-only data: 47104k Jan 24 11:47:45.902165 kernel: Freeing unused kernel image (rodata/data gap) memory: 1124K Jan 24 11:47:45.902218 kernel: Run /init as init process Jan 24 11:47:45.902230 kernel: with arguments: Jan 24 11:47:45.902244 kernel: /init Jan 24 11:47:45.902255 kernel: with environment: Jan 24 11:47:45.902267 kernel: HOME=/ Jan 24 11:47:45.902278 kernel: TERM=linux Jan 24 11:47:45.902290 kernel: SCSI subsystem initialized Jan 24 11:47:45.902336 kernel: libata version 3.00 loaded. Jan 24 11:47:45.904595 kernel: ahci 0000:00:1f.2: version 3.0 Jan 24 11:47:45.904639 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Jan 24 11:47:45.904964 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Jan 24 11:47:45.905245 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Jan 24 11:47:45.905576 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Jan 24 11:47:45.906097 kernel: scsi host0: ahci Jan 24 11:47:45.906663 kernel: scsi host1: ahci Jan 24 11:47:45.907066 kernel: scsi host2: ahci Jan 24 11:47:45.907362 kernel: scsi host3: ahci Jan 24 11:47:45.910955 kernel: scsi host4: ahci Jan 24 11:47:45.911377 kernel: scsi host5: ahci Jan 24 11:47:45.911501 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 26 lpm-pol 1 Jan 24 11:47:45.911519 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 26 lpm-pol 1 Jan 24 11:47:45.911533 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 26 lpm-pol 1 Jan 24 11:47:45.911546 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 26 lpm-pol 1 Jan 24 11:47:45.911559 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 26 lpm-pol 1 Jan 24 11:47:45.911572 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 26 lpm-pol 1 Jan 24 11:47:45.911623 kernel: ata4: SATA link down (SStatus 0 SControl 300) Jan 24 11:47:45.911636 kernel: ata1: SATA link down (SStatus 0 SControl 300) Jan 24 11:47:45.911650 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Jan 24 11:47:45.911663 kernel: ata3.00: LPM support broken, forcing max_power Jan 24 11:47:45.911677 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Jan 24 11:47:45.911729 kernel: ata3.00: applying bridge limits Jan 24 11:47:45.911741 kernel: ata5: SATA link down (SStatus 0 SControl 300) Jan 24 11:47:45.911854 kernel: ata6: SATA link down (SStatus 0 SControl 300) Jan 24 11:47:45.911867 kernel: ata2: SATA link down (SStatus 0 SControl 300) Jan 24 11:47:45.911878 kernel: ata3.00: LPM support broken, forcing max_power Jan 24 11:47:45.911890 kernel: ata3.00: configured for UDMA/100 Jan 24 11:47:45.912266 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Jan 24 11:47:45.914758 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Jan 24 11:47:45.915255 kernel: virtio_blk virtio1: [vda] 27000832 512-byte logical blocks (13.8 GB/12.9 GiB) Jan 24 11:47:45.915325 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jan 24 11:47:45.915338 kernel: GPT:16515071 != 27000831 Jan 24 11:47:45.915349 kernel: GPT:Alternate GPT header not at the end of the disk. Jan 24 11:47:45.915360 kernel: GPT:16515071 != 27000831 Jan 24 11:47:45.915373 kernel: GPT: Use GNU Parted to correct GPT errors. Jan 24 11:47:45.915388 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 24 11:47:45.915847 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Jan 24 11:47:45.915869 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Jan 24 11:47:45.916178 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Jan 24 11:47:45.916199 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 24 11:47:45.916211 kernel: device-mapper: uevent: version 1.0.3 Jan 24 11:47:45.916223 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Jan 24 11:47:45.916235 kernel: device-mapper: verity: sha256 using shash "sha256-generic" Jan 24 11:47:45.916299 kernel: raid6: avx2x4 gen() 23157 MB/s Jan 24 11:47:45.916339 kernel: raid6: avx2x2 gen() 21815 MB/s Jan 24 11:47:45.916352 kernel: raid6: avx2x1 gen() 13221 MB/s Jan 24 11:47:45.916363 kernel: raid6: using algorithm avx2x4 gen() 23157 MB/s Jan 24 11:47:45.916435 kernel: raid6: .... xor() 4735 MB/s, rmw enabled Jan 24 11:47:45.916507 kernel: raid6: using avx2x2 recovery algorithm Jan 24 11:47:45.916524 kernel: xor: automatically using best checksumming function avx Jan 24 11:47:45.916573 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 24 11:47:45.916586 kernel: BTRFS: device fsid d3bd77fc-0f38-45e2-bb37-1f1b4d0917b8 devid 1 transid 34 /dev/mapper/usr (253:0) scanned by mount (181) Jan 24 11:47:45.916635 kernel: BTRFS info (device dm-0): first mount of filesystem d3bd77fc-0f38-45e2-bb37-1f1b4d0917b8 Jan 24 11:47:45.916677 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jan 24 11:47:45.916763 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 24 11:47:45.916778 kernel: BTRFS info (device dm-0): enabling free space tree Jan 24 11:47:45.916790 kernel: loop: module loaded Jan 24 11:47:45.916802 kernel: loop0: detected capacity change from 0 to 100552 Jan 24 11:47:45.916813 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jan 24 11:47:45.916825 kernel: clocksource: Long readout interval, skipping watchdog check: cs_nsec: 3615111238 wd_nsec: 3615111152 Jan 24 11:47:45.916870 systemd[1]: Successfully made /usr/ read-only. Jan 24 11:47:45.916921 systemd[1]: systemd 257.9 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jan 24 11:47:45.916937 systemd[1]: Detected virtualization kvm. Jan 24 11:47:45.916951 systemd[1]: Detected architecture x86-64. Jan 24 11:47:45.916962 systemd[1]: Running in initrd. Jan 24 11:47:45.916974 systemd[1]: No hostname configured, using default hostname. Jan 24 11:47:45.916986 systemd[1]: Hostname set to . Jan 24 11:47:45.917043 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Jan 24 11:47:45.917055 systemd[1]: Queued start job for default target initrd.target. Jan 24 11:47:45.917067 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Jan 24 11:47:45.917079 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 24 11:47:45.917091 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 24 11:47:45.917106 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 24 11:47:45.917122 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 24 11:47:45.917177 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 24 11:47:45.917189 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 24 11:47:45.917203 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 24 11:47:45.917218 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 24 11:47:45.917231 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Jan 24 11:47:45.917282 systemd[1]: Reached target paths.target - Path Units. Jan 24 11:47:45.917297 systemd[1]: Reached target slices.target - Slice Units. Jan 24 11:47:45.917312 systemd[1]: Reached target swap.target - Swaps. Jan 24 11:47:45.917326 systemd[1]: Reached target timers.target - Timer Units. Jan 24 11:47:45.917338 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 24 11:47:45.917350 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 24 11:47:45.917362 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Jan 24 11:47:45.917416 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 24 11:47:45.917429 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Jan 24 11:47:45.917441 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 24 11:47:45.917515 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 24 11:47:45.917529 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 24 11:47:45.917541 systemd[1]: Reached target sockets.target - Socket Units. Jan 24 11:47:45.917553 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jan 24 11:47:45.917608 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 24 11:47:45.917620 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 24 11:47:45.917632 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 24 11:47:45.917645 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Jan 24 11:47:45.917660 systemd[1]: Starting systemd-fsck-usr.service... Jan 24 11:47:45.917674 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 24 11:47:45.917751 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 24 11:47:45.917769 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 24 11:47:45.917781 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 24 11:47:45.917794 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 24 11:47:45.917850 systemd[1]: Finished systemd-fsck-usr.service. Jan 24 11:47:45.917865 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 24 11:47:45.917993 systemd-journald[319]: Collecting audit messages is enabled. Jan 24 11:47:45.918062 systemd-journald[319]: Journal started Jan 24 11:47:45.918086 systemd-journald[319]: Runtime Journal (/run/log/journal/bcdcd4c6c711491299b3f38e21d6275e) is 6M, max 48.2M, 42.1M free. Jan 24 11:47:45.933977 systemd[1]: Started systemd-journald.service - Journal Service. Jan 24 11:47:45.984169 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 24 11:47:46.017067 kernel: audit: type=1130 audit(1769255265.934:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:45.934000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.069573 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 24 11:47:46.084575 kernel: Bridge firewalling registered Jan 24 11:47:46.085416 systemd-modules-load[322]: Inserted module 'br_netfilter' Jan 24 11:47:46.090169 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 24 11:47:46.091154 systemd-tmpfiles[330]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Jan 24 11:47:46.316037 kernel: audit: type=1130 audit(1769255266.102:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.102000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.104767 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 24 11:47:46.340000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.367303 kernel: audit: type=1130 audit(1769255266.340:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.367837 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 24 11:47:46.386544 kernel: audit: type=1130 audit(1769255266.372:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.372000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.386894 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 24 11:47:46.395000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.411854 kernel: audit: type=1130 audit(1769255266.395:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.413051 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 24 11:47:46.425800 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 24 11:47:46.448098 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 24 11:47:46.492771 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 24 11:47:46.519541 kernel: audit: type=1130 audit(1769255266.506:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.506000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.519540 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 24 11:47:46.535288 kernel: audit: type=1130 audit(1769255266.522:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.522000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.523361 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 24 11:47:46.555323 kernel: audit: type=1130 audit(1769255266.540:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.540000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:46.557610 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 24 11:47:46.565000 audit: BPF prog-id=6 op=LOAD Jan 24 11:47:46.568038 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 24 11:47:46.582502 kernel: audit: type=1334 audit(1769255266.565:10): prog-id=6 op=LOAD Jan 24 11:47:46.610827 dracut-cmdline[356]: dracut-109 Jan 24 11:47:46.624575 dracut-cmdline[356]: Using kernel command line parameters: SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=7953d3c7acaad6ee79638a10c67ea9f0b3a8597919989b6fbf2f9a1742d4ba63 Jan 24 11:47:46.695533 systemd-resolved[357]: Positive Trust Anchors: Jan 24 11:47:46.695567 systemd-resolved[357]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 24 11:47:46.695574 systemd-resolved[357]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Jan 24 11:47:46.695619 systemd-resolved[357]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 24 11:47:47.424380 systemd-resolved[357]: Defaulting to hostname 'linux'. Jan 24 11:47:47.432948 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 24 11:47:47.460751 kernel: audit: type=1130 audit(1769255267.437:11): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:47.437000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:47.437899 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 24 11:47:47.715526 kernel: Loading iSCSI transport class v2.0-870. Jan 24 11:47:47.797909 kernel: iscsi: registered transport (tcp) Jan 24 11:47:47.871333 kernel: iscsi: registered transport (qla4xxx) Jan 24 11:47:47.873123 kernel: QLogic iSCSI HBA Driver Jan 24 11:47:47.967144 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 24 11:47:48.055559 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 24 11:47:48.054000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:48.057857 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 24 11:47:48.477255 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 24 11:47:48.490000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:48.515430 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 24 11:47:48.541517 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 24 11:47:48.722867 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 24 11:47:48.728000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:48.730000 audit: BPF prog-id=7 op=LOAD Jan 24 11:47:48.731000 audit: BPF prog-id=8 op=LOAD Jan 24 11:47:48.736341 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 24 11:47:48.824238 systemd-udevd[584]: Using default interface naming scheme 'v257'. Jan 24 11:47:48.895317 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 24 11:47:48.895000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:48.898750 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 24 11:47:49.072589 dracut-pre-trigger[630]: rd.md=0: removing MD RAID activation Jan 24 11:47:49.199324 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 24 11:47:49.218000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:49.219991 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 24 11:47:49.224000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:49.228000 audit: BPF prog-id=9 op=LOAD Jan 24 11:47:49.239870 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 24 11:47:49.266631 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 24 11:47:49.392499 systemd-networkd[722]: lo: Link UP Jan 24 11:47:49.392526 systemd-networkd[722]: lo: Gained carrier Jan 24 11:47:49.398844 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 24 11:47:49.403000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:49.403767 systemd[1]: Reached target network.target - Network. Jan 24 11:47:49.484034 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 24 11:47:49.492000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:49.497662 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 24 11:47:50.626811 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Jan 24 11:47:50.667620 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Jan 24 11:47:50.689500 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 24 11:47:50.724827 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Jan 24 11:47:50.744948 kernel: cryptd: max_cpu_qlen set to 1000 Jan 24 11:47:50.745314 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 24 11:47:50.789920 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Jan 24 11:47:50.791067 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 24 11:47:50.791290 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 24 11:47:50.804000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:50.809989 disk-uuid[765]: Primary Header is updated. Jan 24 11:47:50.809989 disk-uuid[765]: Secondary Entries is updated. Jan 24 11:47:50.809989 disk-uuid[765]: Secondary Header is updated. Jan 24 11:47:50.804964 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 24 11:47:50.825263 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 24 11:47:50.856397 kernel: AES CTR mode by8 optimization enabled Jan 24 11:47:51.103222 systemd-networkd[722]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 24 11:47:51.131786 systemd-networkd[722]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 24 11:47:51.358378 systemd-networkd[722]: eth0: Link UP Jan 24 11:47:51.471376 systemd-networkd[722]: eth0: Gained carrier Jan 24 11:47:51.475606 systemd-networkd[722]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 24 11:47:51.583946 systemd-networkd[722]: eth0: DHCPv4 address 10.0.0.67/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jan 24 11:47:51.741658 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 24 11:47:51.916000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:51.921786 kernel: kauditd_printk_skb: 12 callbacks suppressed Jan 24 11:47:51.921867 kernel: audit: type=1130 audit(1769255271.916:24): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:51.931031 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 24 11:47:51.956795 kernel: audit: type=1130 audit(1769255271.939:25): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:51.939000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:51.943025 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 24 11:47:51.962028 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 24 11:47:51.962155 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 24 11:47:51.987071 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 24 11:47:52.057817 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 24 11:47:52.082823 kernel: audit: type=1130 audit(1769255272.064:26): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:52.064000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:52.616329 disk-uuid[767]: Warning: The kernel is still using the old partition table. Jan 24 11:47:52.616329 disk-uuid[767]: The new table will be used at the next reboot or after you Jan 24 11:47:52.616329 disk-uuid[767]: run partprobe(8) or kpartx(8) Jan 24 11:47:52.616329 disk-uuid[767]: The operation has completed successfully. Jan 24 11:47:52.659294 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 24 11:47:52.662823 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 24 11:47:52.679000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:52.679000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:52.701663 kernel: audit: type=1130 audit(1769255272.679:27): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:52.701774 kernel: audit: type=1131 audit(1769255272.679:28): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:52.708009 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 24 11:47:52.987059 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (856) Jan 24 11:47:53.007424 kernel: BTRFS info (device vda6): first mount of filesystem 1b92a19b-e1e6-4749-8204-553c8c72e265 Jan 24 11:47:53.007798 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 24 11:47:53.044698 kernel: BTRFS info (device vda6): turning on async discard Jan 24 11:47:53.045000 kernel: BTRFS info (device vda6): enabling free space tree Jan 24 11:47:53.182074 kernel: BTRFS info (device vda6): last unmount of filesystem 1b92a19b-e1e6-4749-8204-553c8c72e265 Jan 24 11:47:53.192683 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 24 11:47:53.215412 kernel: audit: type=1130 audit(1769255273.198:29): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:53.198000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:53.265662 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 24 11:47:53.432426 systemd-networkd[722]: eth0: Gained IPv6LL Jan 24 11:47:53.687988 kernel: hrtimer: interrupt took 16966375 ns Jan 24 11:47:56.327116 ignition[875]: Ignition 2.24.0 Jan 24 11:47:56.328213 ignition[875]: Stage: fetch-offline Jan 24 11:47:56.330001 ignition[875]: no configs at "/usr/lib/ignition/base.d" Jan 24 11:47:56.330024 ignition[875]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 24 11:47:56.330523 ignition[875]: parsed url from cmdline: "" Jan 24 11:47:56.330532 ignition[875]: no config URL provided Jan 24 11:47:56.330542 ignition[875]: reading system config file "/usr/lib/ignition/user.ign" Jan 24 11:47:56.330561 ignition[875]: no config at "/usr/lib/ignition/user.ign" Jan 24 11:47:56.330720 ignition[875]: op(1): [started] loading QEMU firmware config module Jan 24 11:47:56.330727 ignition[875]: op(1): executing: "modprobe" "qemu_fw_cfg" Jan 24 11:47:56.389322 ignition[875]: op(1): [finished] loading QEMU firmware config module Jan 24 11:47:56.799527 ignition[875]: parsing config with SHA512: 42f3b64f1480283aac83ddac14cf4610092392d04d3efae0fbbf0a1bc86a998f2609c2353457906e395764f4d859a030a9fab09e0c96bdcdeef72ccecacf894c Jan 24 11:47:57.068969 unknown[875]: fetched base config from "system" Jan 24 11:47:57.069278 unknown[875]: fetched user config from "qemu" Jan 24 11:47:57.071784 ignition[875]: fetch-offline: fetch-offline passed Jan 24 11:47:57.071935 ignition[875]: Ignition finished successfully Jan 24 11:47:57.122066 kernel: audit: type=1130 audit(1769255277.092:30): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:57.092000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:57.085165 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 24 11:47:57.095886 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Jan 24 11:47:57.102271 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 24 11:47:57.297682 ignition[885]: Ignition 2.24.0 Jan 24 11:47:57.297768 ignition[885]: Stage: kargs Jan 24 11:47:57.298075 ignition[885]: no configs at "/usr/lib/ignition/base.d" Jan 24 11:47:57.298258 ignition[885]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 24 11:47:57.863905 ignition[885]: kargs: kargs passed Jan 24 11:47:57.864260 ignition[885]: Ignition finished successfully Jan 24 11:47:57.879893 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 24 11:47:57.901899 kernel: audit: type=1130 audit(1769255277.887:31): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:57.887000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:57.890197 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 24 11:47:58.008248 ignition[893]: Ignition 2.24.0 Jan 24 11:47:58.008292 ignition[893]: Stage: disks Jan 24 11:47:58.008916 ignition[893]: no configs at "/usr/lib/ignition/base.d" Jan 24 11:47:58.008935 ignition[893]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 24 11:47:58.010926 ignition[893]: disks: disks passed Jan 24 11:47:58.011003 ignition[893]: Ignition finished successfully Jan 24 11:47:58.030801 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 24 11:47:58.063903 kernel: audit: type=1130 audit(1769255278.035:32): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:58.035000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:58.036896 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 24 11:47:58.069244 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 24 11:47:58.069397 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 24 11:47:58.098902 systemd[1]: Reached target sysinit.target - System Initialization. Jan 24 11:47:58.109343 systemd[1]: Reached target basic.target - Basic System. Jan 24 11:47:58.139662 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 24 11:47:58.328584 systemd-fsck[903]: ROOT: clean, 15/456736 files, 38230/456704 blocks Jan 24 11:47:58.355419 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 24 11:47:58.401654 kernel: audit: type=1130 audit(1769255278.369:33): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:58.369000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:47:58.383664 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 24 11:47:59.383951 kernel: EXT4-fs (vda9): mounted filesystem 04920273-eebf-4ad5-828c-7340043c8075 r/w with ordered data mode. Quota mode: none. Jan 24 11:47:59.398860 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 24 11:47:59.405433 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 24 11:47:59.512291 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 24 11:47:59.540830 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 24 11:47:59.576533 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jan 24 11:47:59.577296 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 24 11:47:59.578163 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 24 11:47:59.669721 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 24 11:47:59.693836 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (912) Jan 24 11:47:59.683512 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 24 11:47:59.707389 kernel: BTRFS info (device vda6): first mount of filesystem 1b92a19b-e1e6-4749-8204-553c8c72e265 Jan 24 11:47:59.707419 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 24 11:47:59.727533 kernel: BTRFS info (device vda6): turning on async discard Jan 24 11:47:59.727574 kernel: BTRFS info (device vda6): enabling free space tree Jan 24 11:47:59.730181 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 24 11:48:01.279147 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 24 11:48:01.300000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:01.303958 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 24 11:48:01.329822 kernel: audit: type=1130 audit(1769255281.300:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:01.337967 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 24 11:48:01.371410 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 24 11:48:01.386428 kernel: BTRFS info (device vda6): last unmount of filesystem 1b92a19b-e1e6-4749-8204-553c8c72e265 Jan 24 11:48:01.474630 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 24 11:48:01.479000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:01.499585 kernel: audit: type=1130 audit(1769255281.479:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:01.530901 ignition[1010]: INFO : Ignition 2.24.0 Jan 24 11:48:01.530901 ignition[1010]: INFO : Stage: mount Jan 24 11:48:01.542208 ignition[1010]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 24 11:48:01.542208 ignition[1010]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 24 11:48:01.542208 ignition[1010]: INFO : mount: mount passed Jan 24 11:48:01.542208 ignition[1010]: INFO : Ignition finished successfully Jan 24 11:48:01.609305 kernel: audit: type=1130 audit(1769255281.569:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:01.569000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:01.556321 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 24 11:48:01.574667 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 24 11:48:01.673022 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 24 11:48:01.752556 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 (254:6) scanned by mount (1021) Jan 24 11:48:01.773062 kernel: BTRFS info (device vda6): first mount of filesystem 1b92a19b-e1e6-4749-8204-553c8c72e265 Jan 24 11:48:01.773211 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 24 11:48:01.820850 kernel: BTRFS info (device vda6): turning on async discard Jan 24 11:48:01.821165 kernel: BTRFS info (device vda6): enabling free space tree Jan 24 11:48:01.827625 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 24 11:48:01.931722 ignition[1038]: INFO : Ignition 2.24.0 Jan 24 11:48:01.931722 ignition[1038]: INFO : Stage: files Jan 24 11:48:01.931722 ignition[1038]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 24 11:48:01.931722 ignition[1038]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 24 11:48:01.969305 ignition[1038]: DEBUG : files: compiled without relabeling support, skipping Jan 24 11:48:01.975650 ignition[1038]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 24 11:48:01.975650 ignition[1038]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 24 11:48:01.993540 ignition[1038]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 24 11:48:02.000008 ignition[1038]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 24 11:48:02.005660 ignition[1038]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 24 11:48:02.000278 unknown[1038]: wrote ssh authorized keys file for user: core Jan 24 11:48:02.025503 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Jan 24 11:48:02.025503 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Jan 24 11:48:02.171815 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jan 24 11:48:02.500173 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Jan 24 11:48:02.500173 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jan 24 11:48:02.500173 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jan 24 11:48:02.560536 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jan 24 11:48:02.582615 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jan 24 11:48:02.582615 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 24 11:48:02.582615 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 24 11:48:02.582615 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 24 11:48:02.582615 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 24 11:48:02.661075 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 24 11:48:02.661075 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 24 11:48:02.661075 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jan 24 11:48:02.661075 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jan 24 11:48:02.661075 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jan 24 11:48:02.661075 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Jan 24 11:48:02.908282 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jan 24 11:48:08.510959 ignition[1038]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jan 24 11:48:08.510959 ignition[1038]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Jan 24 11:48:08.534280 ignition[1038]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 24 11:48:08.534280 ignition[1038]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 24 11:48:08.534280 ignition[1038]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Jan 24 11:48:08.534280 ignition[1038]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Jan 24 11:48:08.534280 ignition[1038]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 24 11:48:08.534280 ignition[1038]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 24 11:48:08.534280 ignition[1038]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Jan 24 11:48:08.534280 ignition[1038]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Jan 24 11:48:08.636315 ignition[1038]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Jan 24 11:48:08.669904 ignition[1038]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Jan 24 11:48:08.669904 ignition[1038]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Jan 24 11:48:08.669904 ignition[1038]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Jan 24 11:48:08.669904 ignition[1038]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Jan 24 11:48:08.669904 ignition[1038]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 24 11:48:08.669904 ignition[1038]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 24 11:48:08.669904 ignition[1038]: INFO : files: files passed Jan 24 11:48:08.669904 ignition[1038]: INFO : Ignition finished successfully Jan 24 11:48:08.731584 kernel: audit: type=1130 audit(1769255288.696:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:08.696000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:08.677343 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 24 11:48:08.723021 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 24 11:48:08.782137 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 24 11:48:08.812000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:08.812000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:08.798608 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 24 11:48:08.860346 kernel: audit: type=1130 audit(1769255288.812:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:08.860387 kernel: audit: type=1131 audit(1769255288.812:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:08.798774 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 24 11:48:08.954895 initrd-setup-root-after-ignition[1069]: grep: /sysroot/oem/oem-release: No such file or directory Jan 24 11:48:08.970915 initrd-setup-root-after-ignition[1075]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 24 11:48:08.983307 initrd-setup-root-after-ignition[1071]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 24 11:48:08.983307 initrd-setup-root-after-ignition[1071]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 24 11:48:09.007554 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 24 11:48:09.036536 kernel: audit: type=1130 audit(1769255289.012:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:09.012000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:09.014052 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 24 11:48:09.031075 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 24 11:48:09.168258 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 24 11:48:09.199919 kernel: audit: type=1130 audit(1769255289.173:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:09.199994 kernel: audit: type=1131 audit(1769255289.173:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:09.173000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:09.173000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:09.168559 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 24 11:48:09.174880 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 24 11:48:09.214548 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 24 11:48:09.227386 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 24 11:48:09.237058 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 24 11:48:09.840691 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 24 11:48:09.878000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:09.889121 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 24 11:48:09.923851 kernel: audit: type=1130 audit(1769255289.878:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:09.980152 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Jan 24 11:48:09.980377 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 24 11:48:09.989316 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 24 11:48:10.006227 systemd[1]: Stopped target timers.target - Timer Units. Jan 24 11:48:10.031874 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 24 11:48:10.081671 kernel: audit: type=1131 audit(1769255290.051:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:10.051000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:10.032189 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 24 11:48:10.090896 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 24 11:48:10.100787 systemd[1]: Stopped target basic.target - Basic System. Jan 24 11:48:10.121000 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 24 11:48:10.125595 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 24 11:48:10.142177 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 24 11:48:10.171241 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Jan 24 11:48:10.178046 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 24 11:48:10.192351 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 24 11:48:10.204318 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 24 11:48:10.208516 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 24 11:48:10.224687 systemd[1]: Stopped target swap.target - Swaps. Jan 24 11:48:10.238000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:10.232033 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 24 11:48:10.268866 kernel: audit: type=1131 audit(1769255290.238:45): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:10.232393 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 24 11:48:10.285729 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 24 11:48:10.298934 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 24 11:48:10.358684 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 24 11:48:10.366599 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 24 11:48:10.379344 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 24 11:48:10.417615 kernel: audit: type=1131 audit(1769255290.398:46): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:10.398000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:10.379842 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 24 11:48:10.418403 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 24 11:48:10.419001 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 24 11:48:10.432000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:10.433188 systemd[1]: Stopped target paths.target - Path Units. Jan 24 11:48:10.444960 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 24 11:48:10.449331 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 24 11:48:10.455845 systemd[1]: Stopped target slices.target - Slice Units. Jan 24 11:48:10.473271 systemd[1]: Stopped target sockets.target - Socket Units. Jan 24 11:48:10.502121 systemd[1]: iscsid.socket: Deactivated successfully. Jan 24 11:48:10.502429 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 24 11:48:10.516116 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 24 11:48:10.516358 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 24 11:48:10.538745 systemd[1]: systemd-journald-audit.socket: Deactivated successfully. Jan 24 11:48:10.538987 systemd[1]: Closed systemd-journald-audit.socket - Journal Audit Socket. Jan 24 11:48:10.567000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:10.552176 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 24 11:48:10.574000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:10.552437 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 24 11:48:10.568973 systemd[1]: ignition-files.service: Deactivated successfully. Jan 24 11:48:10.569266 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 24 11:48:10.602420 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 24 11:48:10.622118 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 24 11:48:10.626219 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 24 11:48:10.642929 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 24 11:48:10.661000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:10.662027 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 24 11:48:10.666234 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 24 11:48:10.678000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:10.678981 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 24 11:48:10.685427 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 24 11:48:10.695000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:10.707760 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 24 11:48:10.711740 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 24 11:48:10.717000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:10.717000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:10.735139 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 24 11:48:10.743787 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 24 11:48:10.744042 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 24 11:48:10.763000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:11.924672 ignition[1095]: INFO : Ignition 2.24.0 Jan 24 11:48:11.924672 ignition[1095]: INFO : Stage: umount Jan 24 11:48:11.924672 ignition[1095]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 24 11:48:11.924672 ignition[1095]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jan 24 11:48:11.973518 ignition[1095]: INFO : umount: umount passed Jan 24 11:48:11.973518 ignition[1095]: INFO : Ignition finished successfully Jan 24 11:48:11.989088 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 24 11:48:11.992000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:11.990737 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 24 11:48:12.012238 systemd[1]: Stopped target network.target - Network. Jan 24 11:48:12.026705 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 24 11:48:12.031000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.026903 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 24 11:48:12.061000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.073000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.080000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup-pre comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.090000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.032744 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 24 11:48:12.032894 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 24 11:48:12.062186 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 24 11:48:12.062342 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 24 11:48:12.074624 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 24 11:48:12.074727 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 24 11:48:12.145000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.081162 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 24 11:48:12.081264 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 24 11:48:12.092313 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 24 11:48:12.174000 audit: BPF prog-id=6 op=UNLOAD Jan 24 11:48:12.101721 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 24 11:48:12.183000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.134967 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 24 11:48:12.135206 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 24 11:48:12.172678 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 24 11:48:12.172977 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 24 11:48:12.216399 systemd[1]: Stopped target network-pre.target - Preparation for Network. Jan 24 11:48:12.221000 audit: BPF prog-id=9 op=UNLOAD Jan 24 11:48:12.225997 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 24 11:48:12.226128 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 24 11:48:12.242333 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 24 11:48:12.242563 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 24 11:48:12.242658 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 24 11:48:12.263000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.269000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.264589 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 24 11:48:12.269000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.264700 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 24 11:48:12.270006 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 24 11:48:12.270091 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 24 11:48:12.270387 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 24 11:48:12.318194 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 24 11:48:12.318560 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 24 11:48:12.342000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.344557 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 24 11:48:12.344657 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 24 11:48:12.357506 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 24 11:48:12.384000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.357680 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 24 11:48:12.399000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.372668 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 24 11:48:12.373009 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 24 11:48:12.386319 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 24 11:48:12.418000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.386567 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 24 11:48:12.406637 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 24 11:48:12.406770 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 24 11:48:12.432725 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 24 11:48:12.444000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.444557 systemd[1]: systemd-network-generator.service: Deactivated successfully. Jan 24 11:48:12.475000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.485000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.497000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.444687 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Jan 24 11:48:12.445055 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 24 11:48:12.445135 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 24 11:48:12.476540 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Jan 24 11:48:12.476651 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 24 11:48:12.486539 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 24 11:48:12.486746 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 24 11:48:12.498655 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 24 11:48:12.498789 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 24 11:48:12.578000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.582421 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 24 11:48:12.594000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.594000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.584992 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 24 11:48:12.685053 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 24 11:48:12.686425 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 24 11:48:12.700000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:12.703151 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 24 11:48:12.715774 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 24 11:48:12.836338 systemd[1]: Switching root. Jan 24 11:48:12.933959 systemd-journald[319]: Received SIGTERM from PID 1 (systemd). Jan 24 11:48:12.934323 systemd-journald[319]: Journal stopped Jan 24 11:48:20.233676 kernel: SELinux: policy capability network_peer_controls=1 Jan 24 11:48:20.234228 kernel: SELinux: policy capability open_perms=1 Jan 24 11:48:20.234369 kernel: SELinux: policy capability extended_socket_class=1 Jan 24 11:48:20.234396 kernel: SELinux: policy capability always_check_network=0 Jan 24 11:48:20.234523 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 24 11:48:20.234580 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 24 11:48:20.235813 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jan 24 11:48:20.236047 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 24 11:48:20.236082 kernel: SELinux: policy capability userspace_initial_context=0 Jan 24 11:48:20.236174 systemd[1]: Successfully loaded SELinux policy in 183.760ms. Jan 24 11:48:20.236308 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 19.871ms. Jan 24 11:48:20.236338 systemd[1]: systemd 257.9 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jan 24 11:48:20.236388 systemd[1]: Detected virtualization kvm. Jan 24 11:48:20.236412 systemd[1]: Detected architecture x86-64. Jan 24 11:48:20.236515 systemd[1]: Detected first boot. Jan 24 11:48:20.236568 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Jan 24 11:48:20.236662 zram_generator::config[1141]: No configuration found. Jan 24 11:48:20.236729 kernel: Guest personality initialized and is inactive Jan 24 11:48:20.236806 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Jan 24 11:48:20.236900 kernel: Initialized host personality Jan 24 11:48:20.236924 kernel: NET: Registered PF_VSOCK protocol family Jan 24 11:48:20.236947 systemd[1]: Populated /etc with preset unit settings. Jan 24 11:48:20.236970 kernel: kauditd_printk_skb: 40 callbacks suppressed Jan 24 11:48:20.237072 kernel: audit: type=1334 audit(1769255298.440:87): prog-id=12 op=LOAD Jan 24 11:48:20.237121 kernel: audit: type=1334 audit(1769255298.441:88): prog-id=3 op=UNLOAD Jan 24 11:48:20.237144 kernel: audit: type=1334 audit(1769255298.441:89): prog-id=13 op=LOAD Jan 24 11:48:20.237164 kernel: audit: type=1334 audit(1769255298.444:90): prog-id=14 op=LOAD Jan 24 11:48:20.237224 kernel: audit: type=1334 audit(1769255298.465:91): prog-id=4 op=UNLOAD Jan 24 11:48:20.237246 kernel: audit: type=1334 audit(1769255298.466:92): prog-id=5 op=UNLOAD Jan 24 11:48:20.237267 kernel: audit: type=1131 audit(1769255298.572:93): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.237332 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jan 24 11:48:20.237355 kernel: audit: type=1334 audit(1769255298.598:94): prog-id=12 op=UNLOAD Jan 24 11:48:20.237402 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jan 24 11:48:20.237422 kernel: audit: type=1130 audit(1769255298.612:95): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.237440 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jan 24 11:48:20.237527 kernel: audit: type=1131 audit(1769255298.612:96): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.237613 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jan 24 11:48:20.237681 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jan 24 11:48:20.237730 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jan 24 11:48:20.237751 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jan 24 11:48:20.237770 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jan 24 11:48:20.237822 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jan 24 11:48:20.237842 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jan 24 11:48:20.237904 systemd[1]: Created slice user.slice - User and Session Slice. Jan 24 11:48:20.237925 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 24 11:48:20.237944 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 24 11:48:20.237963 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jan 24 11:48:20.237983 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jan 24 11:48:20.238072 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jan 24 11:48:20.238124 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 24 11:48:20.238145 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jan 24 11:48:20.238167 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 24 11:48:20.238188 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 24 11:48:20.238210 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jan 24 11:48:20.238230 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jan 24 11:48:20.238243 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jan 24 11:48:20.238293 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jan 24 11:48:20.238332 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 24 11:48:20.238394 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 24 11:48:20.238408 systemd[1]: Reached target remote-veritysetup.target - Remote Verity Protected Volumes. Jan 24 11:48:20.238420 systemd[1]: Reached target slices.target - Slice Units. Jan 24 11:48:20.238433 systemd[1]: Reached target swap.target - Swaps. Jan 24 11:48:20.238530 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jan 24 11:48:20.238617 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jan 24 11:48:20.238649 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Jan 24 11:48:20.238680 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Jan 24 11:48:20.238759 systemd[1]: Listening on systemd-mountfsd.socket - DDI File System Mounter Socket. Jan 24 11:48:20.238800 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 24 11:48:20.238819 systemd[1]: Listening on systemd-nsresourced.socket - Namespace Resource Manager Socket. Jan 24 11:48:20.238837 systemd[1]: Listening on systemd-oomd.socket - Userspace Out-Of-Memory (OOM) Killer Socket. Jan 24 11:48:20.238915 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 24 11:48:20.238935 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 24 11:48:20.238953 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jan 24 11:48:20.238971 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jan 24 11:48:20.238992 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jan 24 11:48:20.239010 systemd[1]: Mounting media.mount - External Media Directory... Jan 24 11:48:20.239050 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:48:20.239098 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jan 24 11:48:20.239116 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jan 24 11:48:20.239159 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jan 24 11:48:20.239178 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jan 24 11:48:20.239195 systemd[1]: Reached target machines.target - Containers. Jan 24 11:48:20.239213 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jan 24 11:48:20.239231 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 24 11:48:20.239293 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 24 11:48:20.239331 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jan 24 11:48:20.239350 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 24 11:48:20.239369 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 24 11:48:20.239388 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 24 11:48:20.239403 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jan 24 11:48:20.239534 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 24 11:48:20.239559 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jan 24 11:48:20.239572 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jan 24 11:48:20.239585 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jan 24 11:48:20.239597 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jan 24 11:48:20.239617 systemd[1]: Stopped systemd-fsck-usr.service. Jan 24 11:48:20.239639 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 24 11:48:20.239714 kernel: ACPI: bus type drm_connector registered Jan 24 11:48:20.239763 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 24 11:48:20.239778 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 24 11:48:20.239791 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 24 11:48:20.239850 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jan 24 11:48:20.239910 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Jan 24 11:48:20.239938 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 24 11:48:20.239968 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:48:20.239981 kernel: fuse: init (API version 7.41) Jan 24 11:48:20.239994 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jan 24 11:48:20.240008 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jan 24 11:48:20.240159 systemd-journald[1227]: Collecting audit messages is enabled. Jan 24 11:48:20.240268 systemd[1]: Mounted media.mount - External Media Directory. Jan 24 11:48:20.240290 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jan 24 11:48:20.240329 systemd-journald[1227]: Journal started Jan 24 11:48:20.240360 systemd-journald[1227]: Runtime Journal (/run/log/journal/bcdcd4c6c711491299b3f38e21d6275e) is 6M, max 48.2M, 42.1M free. Jan 24 11:48:19.099000 audit[1]: EVENT_LISTENER pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Jan 24 11:48:19.844000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:19.877000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:19.895000 audit: BPF prog-id=14 op=UNLOAD Jan 24 11:48:19.895000 audit: BPF prog-id=13 op=UNLOAD Jan 24 11:48:19.918000 audit: BPF prog-id=15 op=LOAD Jan 24 11:48:19.935000 audit: BPF prog-id=16 op=LOAD Jan 24 11:48:19.936000 audit: BPF prog-id=17 op=LOAD Jan 24 11:48:20.209000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Jan 24 11:48:20.209000 audit[1227]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=6 a1=7fff9fae43d0 a2=4000 a3=0 items=0 ppid=1 pid=1227 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:20.209000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Jan 24 11:48:18.395219 systemd[1]: Queued start job for default target multi-user.target. Jan 24 11:48:18.488269 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Jan 24 11:48:18.512801 systemd[1]: systemd-journald.service: Deactivated successfully. Jan 24 11:48:18.573343 systemd[1]: systemd-journald.service: Consumed 2.047s CPU time. Jan 24 11:48:20.247636 systemd[1]: Started systemd-journald.service - Journal Service. Jan 24 11:48:20.256000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.260642 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jan 24 11:48:20.275245 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jan 24 11:48:20.283193 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jan 24 11:48:20.301000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.303828 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 24 11:48:20.315000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.317043 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jan 24 11:48:20.317437 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jan 24 11:48:20.322000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.323926 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 24 11:48:20.322000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.325185 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 24 11:48:20.340940 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 24 11:48:20.341328 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 24 11:48:20.339000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.339000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.344000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.344000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.347780 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 24 11:48:20.348728 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 24 11:48:20.373784 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jan 24 11:48:20.374309 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jan 24 11:48:20.372000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.372000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.378000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.380000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.381751 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 24 11:48:20.382158 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 24 11:48:20.394760 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 24 11:48:20.392000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.393000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.408000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.417333 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 24 11:48:20.433000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.469404 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jan 24 11:48:20.493000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.499185 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Jan 24 11:48:20.507000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-load-credentials comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.574801 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 24 11:48:20.580971 systemd[1]: Listening on systemd-importd.socket - Disk Image Download Service Socket. Jan 24 11:48:20.596800 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jan 24 11:48:20.606541 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jan 24 11:48:20.611398 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jan 24 11:48:20.614608 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 24 11:48:20.628710 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Jan 24 11:48:20.691104 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 24 11:48:20.692098 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 24 11:48:20.700256 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jan 24 11:48:20.723102 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jan 24 11:48:20.733205 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 24 11:48:20.805802 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jan 24 11:48:20.814832 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 24 11:48:20.824669 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 24 11:48:20.840154 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jan 24 11:48:20.880213 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 24 11:48:20.902000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.902905 systemd-journald[1227]: Time spent on flushing to /var/log/journal/bcdcd4c6c711491299b3f38e21d6275e is 100.812ms for 1121 entries. Jan 24 11:48:20.902905 systemd-journald[1227]: System Journal (/var/log/journal/bcdcd4c6c711491299b3f38e21d6275e) is 8M, max 163.5M, 155.5M free. Jan 24 11:48:21.075574 systemd-journald[1227]: Received client request to flush runtime journal. Jan 24 11:48:20.944000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:20.897580 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 24 11:48:20.903026 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jan 24 11:48:20.921189 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jan 24 11:48:20.930414 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jan 24 11:48:20.983583 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jan 24 11:48:21.014012 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Jan 24 11:48:21.103201 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jan 24 11:48:21.105919 kernel: loop1: detected capacity change from 0 to 50784 Jan 24 11:48:21.116000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:21.133415 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 24 11:48:21.137000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:21.172355 systemd-tmpfiles[1262]: ACLs are not supported, ignoring. Jan 24 11:48:21.172389 systemd-tmpfiles[1262]: ACLs are not supported, ignoring. Jan 24 11:48:21.184070 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 24 11:48:21.194000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:21.200096 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jan 24 11:48:21.207296 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jan 24 11:48:21.208746 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Jan 24 11:48:21.214000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:21.232527 kernel: loop2: detected capacity change from 0 to 111560 Jan 24 11:48:21.329085 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jan 24 11:48:21.332788 kernel: loop3: detected capacity change from 0 to 229808 Jan 24 11:48:21.338000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:21.341000 audit: BPF prog-id=18 op=LOAD Jan 24 11:48:21.341000 audit: BPF prog-id=19 op=LOAD Jan 24 11:48:21.341000 audit: BPF prog-id=20 op=LOAD Jan 24 11:48:21.345300 systemd[1]: Starting systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer... Jan 24 11:48:21.373000 audit: BPF prog-id=21 op=LOAD Jan 24 11:48:21.375951 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 24 11:48:21.396000 audit: BPF prog-id=22 op=LOAD Jan 24 11:48:21.396000 audit: BPF prog-id=23 op=LOAD Jan 24 11:48:21.396000 audit: BPF prog-id=24 op=LOAD Jan 24 11:48:21.389695 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 24 11:48:21.398387 systemd[1]: Starting systemd-nsresourced.service - Namespace Resource Manager... Jan 24 11:48:21.404000 audit: BPF prog-id=25 op=LOAD Jan 24 11:48:21.404000 audit: BPF prog-id=26 op=LOAD Jan 24 11:48:21.405000 audit: BPF prog-id=27 op=LOAD Jan 24 11:48:21.409682 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jan 24 11:48:21.811914 kernel: loop4: detected capacity change from 0 to 50784 Jan 24 11:48:21.863231 systemd-tmpfiles[1288]: ACLs are not supported, ignoring. Jan 24 11:48:21.866567 systemd-tmpfiles[1288]: ACLs are not supported, ignoring. Jan 24 11:48:21.893301 kernel: loop5: detected capacity change from 0 to 111560 Jan 24 11:48:21.892000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:21.887022 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 24 11:48:21.921919 systemd-nsresourced[1289]: Not setting up BPF subsystem, as functionality has been disabled at compile time. Jan 24 11:48:21.923527 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jan 24 11:48:21.925523 kernel: loop6: detected capacity change from 0 to 229808 Jan 24 11:48:21.929000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:21.944000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-nsresourced comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:21.938686 systemd[1]: Started systemd-nsresourced.service - Namespace Resource Manager. Jan 24 11:48:21.982961 (sd-merge)[1292]: Using extensions 'containerd-flatcar.raw', 'docker-flatcar.raw', 'kubernetes.raw'. Jan 24 11:48:21.995441 (sd-merge)[1292]: Merged extensions into '/usr'. Jan 24 11:48:22.275963 systemd[1]: Reload requested from client PID 1261 ('systemd-sysext') (unit systemd-sysext.service)... Jan 24 11:48:22.276046 systemd[1]: Reloading... Jan 24 11:48:22.369414 systemd-oomd[1286]: No swap; memory pressure usage will be degraded Jan 24 11:48:22.440552 zram_generator::config[1337]: No configuration found. Jan 24 11:48:22.443176 systemd-resolved[1287]: Positive Trust Anchors: Jan 24 11:48:22.443729 systemd-resolved[1287]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 24 11:48:22.443806 systemd-resolved[1287]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Jan 24 11:48:22.443961 systemd-resolved[1287]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 24 11:48:22.473318 systemd-resolved[1287]: Defaulting to hostname 'linux'. Jan 24 11:48:23.240360 systemd[1]: Reloading finished in 962 ms. Jan 24 11:48:23.301929 systemd[1]: Started systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer. Jan 24 11:48:23.315000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-oomd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:23.316416 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 24 11:48:23.322000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:23.323305 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jan 24 11:48:23.329000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:23.346354 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 24 11:48:23.705426 systemd[1]: Starting ensure-sysext.service... Jan 24 11:48:23.710914 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 24 11:48:23.716000 audit: BPF prog-id=28 op=LOAD Jan 24 11:48:23.719637 kernel: kauditd_printk_skb: 53 callbacks suppressed Jan 24 11:48:23.719711 kernel: audit: type=1334 audit(1769255303.716:148): prog-id=28 op=LOAD Jan 24 11:48:23.724177 kernel: audit: type=1334 audit(1769255303.716:149): prog-id=25 op=UNLOAD Jan 24 11:48:23.716000 audit: BPF prog-id=25 op=UNLOAD Jan 24 11:48:23.716000 audit: BPF prog-id=29 op=LOAD Jan 24 11:48:23.735567 kernel: audit: type=1334 audit(1769255303.716:150): prog-id=29 op=LOAD Jan 24 11:48:23.743000 audit: BPF prog-id=30 op=LOAD Jan 24 11:48:23.783975 kernel: audit: type=1334 audit(1769255303.743:151): prog-id=30 op=LOAD Jan 24 11:48:23.784094 kernel: audit: type=1334 audit(1769255303.743:152): prog-id=26 op=UNLOAD Jan 24 11:48:23.743000 audit: BPF prog-id=26 op=UNLOAD Jan 24 11:48:23.788898 kernel: audit: type=1334 audit(1769255303.743:153): prog-id=27 op=UNLOAD Jan 24 11:48:23.743000 audit: BPF prog-id=27 op=UNLOAD Jan 24 11:48:23.792544 kernel: audit: type=1334 audit(1769255303.745:154): prog-id=31 op=LOAD Jan 24 11:48:23.745000 audit: BPF prog-id=31 op=LOAD Jan 24 11:48:23.796011 kernel: audit: type=1334 audit(1769255303.745:155): prog-id=21 op=UNLOAD Jan 24 11:48:23.745000 audit: BPF prog-id=21 op=UNLOAD Jan 24 11:48:23.799160 kernel: audit: type=1334 audit(1769255303.755:156): prog-id=32 op=LOAD Jan 24 11:48:23.755000 audit: BPF prog-id=32 op=LOAD Jan 24 11:48:23.798277 systemd[1]: Reload requested from client PID 1374 ('systemctl') (unit ensure-sysext.service)... Jan 24 11:48:23.798295 systemd[1]: Reloading... Jan 24 11:48:23.755000 audit: BPF prog-id=15 op=UNLOAD Jan 24 11:48:23.761000 audit: BPF prog-id=33 op=LOAD Jan 24 11:48:23.803570 kernel: audit: type=1334 audit(1769255303.755:157): prog-id=15 op=UNLOAD Jan 24 11:48:23.761000 audit: BPF prog-id=34 op=LOAD Jan 24 11:48:23.761000 audit: BPF prog-id=16 op=UNLOAD Jan 24 11:48:23.761000 audit: BPF prog-id=17 op=UNLOAD Jan 24 11:48:23.767000 audit: BPF prog-id=35 op=LOAD Jan 24 11:48:23.767000 audit: BPF prog-id=18 op=UNLOAD Jan 24 11:48:23.767000 audit: BPF prog-id=36 op=LOAD Jan 24 11:48:23.767000 audit: BPF prog-id=37 op=LOAD Jan 24 11:48:23.767000 audit: BPF prog-id=19 op=UNLOAD Jan 24 11:48:23.767000 audit: BPF prog-id=20 op=UNLOAD Jan 24 11:48:23.767000 audit: BPF prog-id=38 op=LOAD Jan 24 11:48:23.767000 audit: BPF prog-id=22 op=UNLOAD Jan 24 11:48:23.767000 audit: BPF prog-id=39 op=LOAD Jan 24 11:48:23.767000 audit: BPF prog-id=40 op=LOAD Jan 24 11:48:23.767000 audit: BPF prog-id=23 op=UNLOAD Jan 24 11:48:23.767000 audit: BPF prog-id=24 op=UNLOAD Jan 24 11:48:23.825929 systemd-tmpfiles[1375]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Jan 24 11:48:23.826627 systemd-tmpfiles[1375]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Jan 24 11:48:23.827214 systemd-tmpfiles[1375]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jan 24 11:48:23.830979 systemd-tmpfiles[1375]: ACLs are not supported, ignoring. Jan 24 11:48:23.831130 systemd-tmpfiles[1375]: ACLs are not supported, ignoring. Jan 24 11:48:23.844427 systemd-tmpfiles[1375]: Detected autofs mount point /boot during canonicalization of boot. Jan 24 11:48:23.845668 systemd-tmpfiles[1375]: Skipping /boot Jan 24 11:48:23.907610 systemd-tmpfiles[1375]: Detected autofs mount point /boot during canonicalization of boot. Jan 24 11:48:23.907631 systemd-tmpfiles[1375]: Skipping /boot Jan 24 11:48:24.198544 zram_generator::config[1410]: No configuration found. Jan 24 11:48:24.797072 systemd[1]: Reloading finished in 995 ms. Jan 24 11:48:24.866684 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jan 24 11:48:24.873000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:24.879000 audit: BPF prog-id=41 op=LOAD Jan 24 11:48:24.879000 audit: BPF prog-id=32 op=UNLOAD Jan 24 11:48:24.879000 audit: BPF prog-id=42 op=LOAD Jan 24 11:48:24.879000 audit: BPF prog-id=43 op=LOAD Jan 24 11:48:24.879000 audit: BPF prog-id=33 op=UNLOAD Jan 24 11:48:24.879000 audit: BPF prog-id=34 op=UNLOAD Jan 24 11:48:24.881000 audit: BPF prog-id=44 op=LOAD Jan 24 11:48:24.884000 audit: BPF prog-id=35 op=UNLOAD Jan 24 11:48:24.884000 audit: BPF prog-id=45 op=LOAD Jan 24 11:48:24.884000 audit: BPF prog-id=46 op=LOAD Jan 24 11:48:24.884000 audit: BPF prog-id=36 op=UNLOAD Jan 24 11:48:24.884000 audit: BPF prog-id=37 op=UNLOAD Jan 24 11:48:24.898000 audit: BPF prog-id=47 op=LOAD Jan 24 11:48:24.899000 audit: BPF prog-id=28 op=UNLOAD Jan 24 11:48:24.899000 audit: BPF prog-id=48 op=LOAD Jan 24 11:48:24.899000 audit: BPF prog-id=49 op=LOAD Jan 24 11:48:24.899000 audit: BPF prog-id=29 op=UNLOAD Jan 24 11:48:24.899000 audit: BPF prog-id=30 op=UNLOAD Jan 24 11:48:24.909000 audit: BPF prog-id=50 op=LOAD Jan 24 11:48:24.909000 audit: BPF prog-id=31 op=UNLOAD Jan 24 11:48:24.911000 audit: BPF prog-id=51 op=LOAD Jan 24 11:48:24.911000 audit: BPF prog-id=38 op=UNLOAD Jan 24 11:48:24.911000 audit: BPF prog-id=52 op=LOAD Jan 24 11:48:24.913000 audit: BPF prog-id=53 op=LOAD Jan 24 11:48:24.913000 audit: BPF prog-id=39 op=UNLOAD Jan 24 11:48:24.913000 audit: BPF prog-id=40 op=UNLOAD Jan 24 11:48:24.920162 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 24 11:48:24.924000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:25.027657 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 24 11:48:25.043397 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jan 24 11:48:25.100631 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jan 24 11:48:25.127827 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jan 24 11:48:25.134000 audit: BPF prog-id=8 op=UNLOAD Jan 24 11:48:25.134000 audit: BPF prog-id=7 op=UNLOAD Jan 24 11:48:25.139000 audit: BPF prog-id=54 op=LOAD Jan 24 11:48:25.139000 audit: BPF prog-id=55 op=LOAD Jan 24 11:48:25.142849 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 24 11:48:25.151992 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jan 24 11:48:25.172505 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:48:25.172794 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 24 11:48:25.176584 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 24 11:48:25.188791 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 24 11:48:25.203176 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 24 11:48:25.203000 audit[1457]: SYSTEM_BOOT pid=1457 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Jan 24 11:48:25.207774 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 24 11:48:25.208134 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 24 11:48:25.208276 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 24 11:48:25.208406 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:48:25.215316 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 24 11:48:25.216343 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 24 11:48:25.226000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:25.226000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:25.228999 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 24 11:48:25.229559 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 24 11:48:25.240000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:25.240000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:25.242767 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 24 11:48:25.243287 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 24 11:48:25.260000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:25.260000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:25.276346 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 24 11:48:25.276729 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 24 11:48:25.282653 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:48:25.283967 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 24 11:48:25.289659 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 24 11:48:25.297708 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 24 11:48:25.303000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Jan 24 11:48:25.303000 audit[1479]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffc43d65df0 a2=420 a3=0 items=0 ppid=1446 pid=1479 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:25.303000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 24 11:48:25.305984 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 24 11:48:25.312720 augenrules[1479]: No rules Jan 24 11:48:25.313946 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 24 11:48:25.314916 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 24 11:48:25.315567 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 24 11:48:25.316096 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:48:25.320053 systemd[1]: audit-rules.service: Deactivated successfully. Jan 24 11:48:25.320695 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 24 11:48:25.328378 systemd-udevd[1456]: Using default interface naming scheme 'v257'. Jan 24 11:48:25.329848 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jan 24 11:48:25.336517 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 24 11:48:25.336937 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 24 11:48:25.343114 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 24 11:48:25.343654 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 24 11:48:25.360800 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 24 11:48:25.361181 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 24 11:48:25.376155 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jan 24 11:48:25.384562 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jan 24 11:48:25.399386 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:48:25.402409 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 24 11:48:25.410740 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 24 11:48:25.413075 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 24 11:48:25.467739 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 24 11:48:25.474950 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 24 11:48:25.487090 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 24 11:48:25.492337 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 24 11:48:25.492853 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 24 11:48:25.493064 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 24 11:48:25.493242 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 24 11:48:25.493386 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 24 11:48:25.497016 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 24 11:48:25.507570 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 24 11:48:25.508093 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 24 11:48:25.514953 augenrules[1493]: /sbin/augenrules: No change Jan 24 11:48:25.519017 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 24 11:48:25.519666 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 24 11:48:25.528408 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 24 11:48:25.528815 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 24 11:48:25.535714 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 24 11:48:25.536181 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 24 11:48:25.547000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Jan 24 11:48:25.547000 audit[1531]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7fff3ff60e70 a2=420 a3=0 items=0 ppid=1493 pid=1531 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:25.547000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 24 11:48:25.548000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Jan 24 11:48:25.548000 audit[1531]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7fff3ff63300 a2=420 a3=0 items=0 ppid=1493 pid=1531 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:25.548000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 24 11:48:25.550409 augenrules[1531]: No rules Jan 24 11:48:25.557276 systemd[1]: audit-rules.service: Deactivated successfully. Jan 24 11:48:25.557930 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 24 11:48:25.573538 systemd[1]: Finished ensure-sysext.service. Jan 24 11:48:25.595192 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 24 11:48:25.600611 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 24 11:48:25.600686 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 24 11:48:25.604730 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jan 24 11:48:26.589815 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jan 24 11:48:26.600093 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Jan 24 11:48:26.651174 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 24 11:48:26.679949 systemd[1]: Reached target time-set.target - System Time Set. Jan 24 11:48:26.690558 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Jan 24 11:48:26.693803 systemd-networkd[1544]: lo: Link UP Jan 24 11:48:26.694496 systemd-networkd[1544]: lo: Gained carrier Jan 24 11:48:26.695732 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jan 24 11:48:26.700363 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 24 11:48:26.701182 systemd-networkd[1544]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 24 11:48:26.701256 systemd-networkd[1544]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 24 11:48:26.706041 systemd[1]: Reached target network.target - Network. Jan 24 11:48:26.707263 systemd-networkd[1544]: eth0: Link UP Jan 24 11:48:26.708144 systemd-networkd[1544]: eth0: Gained carrier Jan 24 11:48:26.708169 systemd-networkd[1544]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 24 11:48:26.712733 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Jan 24 11:48:26.737522 kernel: mousedev: PS/2 mouse device common for all mice Jan 24 11:48:26.754061 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jan 24 11:48:26.836699 systemd-networkd[1544]: eth0: DHCPv4 address 10.0.0.67/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jan 24 11:48:26.841593 systemd-timesyncd[1545]: Network configuration changed, trying to establish connection. Jan 24 11:48:27.395936 systemd-timesyncd[1545]: Contacted time server 10.0.0.1:123 (10.0.0.1). Jan 24 11:48:27.396356 systemd-timesyncd[1545]: Initial clock synchronization to Sat 2026-01-24 11:48:27.395132 UTC. Jan 24 11:48:27.396930 systemd-resolved[1287]: Clock change detected. Flushing caches. Jan 24 11:48:27.406552 kernel: ACPI: button: Power Button [PWRF] Jan 24 11:48:27.435790 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jan 24 11:48:27.477336 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Jan 24 11:48:27.484404 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Jan 24 11:48:27.519236 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Jan 24 11:48:28.046793 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 24 11:48:28.422791 ldconfig[1448]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jan 24 11:48:28.453560 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jan 24 11:48:28.964970 systemd-networkd[1544]: eth0: Gained IPv6LL Jan 24 11:48:29.082540 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jan 24 11:48:29.099643 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 24 11:48:29.146888 systemd[1]: Reached target network-online.target - Network is Online. Jan 24 11:48:29.156141 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jan 24 11:48:29.273085 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jan 24 11:48:29.278031 systemd[1]: Reached target sysinit.target - System Initialization. Jan 24 11:48:29.293568 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jan 24 11:48:29.331068 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jan 24 11:48:29.422810 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Jan 24 11:48:29.498884 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jan 24 11:48:29.565602 kernel: kvm_amd: TSC scaling supported Jan 24 11:48:29.566246 kernel: kvm_amd: Nested Virtualization enabled Jan 24 11:48:29.566369 kernel: kvm_amd: Nested Paging enabled Jan 24 11:48:29.566399 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Jan 24 11:48:29.566424 kernel: kvm_amd: PMU virtualization is disabled Jan 24 11:48:29.619465 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jan 24 11:48:29.624324 systemd[1]: Started systemd-sysupdate-reboot.timer - Reboot Automatically After System Update. Jan 24 11:48:29.629829 systemd[1]: Started systemd-sysupdate.timer - Automatic System Update. Jan 24 11:48:29.633911 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jan 24 11:48:29.638029 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jan 24 11:48:29.638392 systemd[1]: Reached target paths.target - Path Units. Jan 24 11:48:29.642167 systemd[1]: Reached target timers.target - Timer Units. Jan 24 11:48:29.651363 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jan 24 11:48:29.664499 systemd[1]: Starting docker.socket - Docker Socket for the API... Jan 24 11:48:29.676581 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Jan 24 11:48:29.683114 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Jan 24 11:48:29.691383 systemd[1]: Reached target ssh-access.target - SSH Access Available. Jan 24 11:48:29.822421 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jan 24 11:48:29.828568 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Jan 24 11:48:29.836936 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jan 24 11:48:29.842147 systemd[1]: Reached target sockets.target - Socket Units. Jan 24 11:48:29.845616 systemd[1]: Reached target basic.target - Basic System. Jan 24 11:48:29.849235 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jan 24 11:48:29.849337 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jan 24 11:48:29.851163 systemd[1]: Starting containerd.service - containerd container runtime... Jan 24 11:48:29.864769 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Jan 24 11:48:29.870327 kernel: EDAC MC: Ver: 3.0.0 Jan 24 11:48:29.889450 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jan 24 11:48:29.895333 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jan 24 11:48:29.926769 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jan 24 11:48:29.943106 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jan 24 11:48:29.949205 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jan 24 11:48:29.949516 jq[1597]: false Jan 24 11:48:29.955174 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Jan 24 11:48:29.980955 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:48:29.993532 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jan 24 11:48:30.015917 extend-filesystems[1598]: Found /dev/vda6 Jan 24 11:48:30.020257 extend-filesystems[1598]: Found /dev/vda9 Jan 24 11:48:30.024141 extend-filesystems[1598]: Checking size of /dev/vda9 Jan 24 11:48:30.033822 google_oslogin_nss_cache[1599]: oslogin_cache_refresh[1599]: Refreshing passwd entry cache Jan 24 11:48:30.030768 oslogin_cache_refresh[1599]: Refreshing passwd entry cache Jan 24 11:48:30.044916 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jan 24 11:48:30.050848 extend-filesystems[1598]: Resized partition /dev/vda9 Jan 24 11:48:30.055968 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jan 24 11:48:30.065108 extend-filesystems[1613]: resize2fs 1.47.3 (8-Jul-2025) Jan 24 11:48:30.071187 google_oslogin_nss_cache[1599]: oslogin_cache_refresh[1599]: Failure getting users, quitting Jan 24 11:48:30.071187 google_oslogin_nss_cache[1599]: oslogin_cache_refresh[1599]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Jan 24 11:48:30.071187 google_oslogin_nss_cache[1599]: oslogin_cache_refresh[1599]: Refreshing group entry cache Jan 24 11:48:30.065998 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jan 24 11:48:30.065785 oslogin_cache_refresh[1599]: Failure getting users, quitting Jan 24 11:48:30.065813 oslogin_cache_refresh[1599]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Jan 24 11:48:30.068399 oslogin_cache_refresh[1599]: Refreshing group entry cache Jan 24 11:48:30.087427 kernel: EXT4-fs (vda9): resizing filesystem from 456704 to 1784827 blocks Jan 24 11:48:30.094686 oslogin_cache_refresh[1599]: Failure getting groups, quitting Jan 24 11:48:30.096378 google_oslogin_nss_cache[1599]: oslogin_cache_refresh[1599]: Failure getting groups, quitting Jan 24 11:48:30.096378 google_oslogin_nss_cache[1599]: oslogin_cache_refresh[1599]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Jan 24 11:48:30.094749 oslogin_cache_refresh[1599]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Jan 24 11:48:30.099468 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jan 24 11:48:30.125813 systemd[1]: Starting systemd-logind.service - User Login Management... Jan 24 11:48:30.138665 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jan 24 11:48:30.140860 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jan 24 11:48:30.144202 systemd[1]: Starting update-engine.service - Update Engine... Jan 24 11:48:30.157253 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jan 24 11:48:30.172920 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jan 24 11:48:30.195827 kernel: EXT4-fs (vda9): resized filesystem to 1784827 Jan 24 11:48:30.192379 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jan 24 11:48:30.193002 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jan 24 11:48:30.193636 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Jan 24 11:48:30.194067 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Jan 24 11:48:30.212486 systemd[1]: motdgen.service: Deactivated successfully. Jan 24 11:48:30.241251 jq[1625]: true Jan 24 11:48:30.217559 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jan 24 11:48:30.241150 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jan 24 11:48:30.241677 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jan 24 11:48:30.245872 update_engine[1622]: I20260124 11:48:30.245743 1622 main.cc:92] Flatcar Update Engine starting Jan 24 11:48:30.251862 extend-filesystems[1613]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Jan 24 11:48:30.251862 extend-filesystems[1613]: old_desc_blocks = 1, new_desc_blocks = 1 Jan 24 11:48:30.251862 extend-filesystems[1613]: The filesystem on /dev/vda9 is now 1784827 (4k) blocks long. Jan 24 11:48:30.290047 extend-filesystems[1598]: Resized filesystem in /dev/vda9 Jan 24 11:48:30.296538 systemd[1]: extend-filesystems.service: Deactivated successfully. Jan 24 11:48:30.297784 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jan 24 11:48:30.307098 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jan 24 11:48:30.490397 jq[1636]: true Jan 24 11:48:30.542015 systemd[1]: coreos-metadata.service: Deactivated successfully. Jan 24 11:48:30.555079 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Jan 24 11:48:31.263608 tar[1633]: linux-amd64/LICENSE Jan 24 11:48:31.263608 tar[1633]: linux-amd64/helm Jan 24 11:48:31.272914 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jan 24 11:48:31.371203 bash[1680]: Updated "/home/core/.ssh/authorized_keys" Jan 24 11:48:31.371900 dbus-daemon[1595]: [system] SELinux support is enabled Jan 24 11:48:31.372711 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jan 24 11:48:31.433904 update_engine[1622]: I20260124 11:48:31.431555 1622 update_check_scheduler.cc:74] Next update check in 11m42s Jan 24 11:48:31.439422 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jan 24 11:48:31.450117 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Jan 24 11:48:31.450763 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jan 24 11:48:31.450810 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jan 24 11:48:31.456039 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jan 24 11:48:31.456074 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jan 24 11:48:31.463444 systemd[1]: Started update-engine.service - Update Engine. Jan 24 11:48:31.478106 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jan 24 11:48:31.920399 sshd_keygen[1646]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jan 24 11:48:31.946059 systemd-logind[1621]: Watching system buttons on /dev/input/event2 (Power Button) Jan 24 11:48:31.946134 systemd-logind[1621]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jan 24 11:48:31.948684 systemd-logind[1621]: New seat seat0. Jan 24 11:48:31.951540 systemd[1]: Started systemd-logind.service - User Login Management. Jan 24 11:48:32.027250 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jan 24 11:48:32.180873 systemd[1]: Starting issuegen.service - Generate /run/issue... Jan 24 11:48:32.368403 systemd[1]: issuegen.service: Deactivated successfully. Jan 24 11:48:32.369004 systemd[1]: Finished issuegen.service - Generate /run/issue. Jan 24 11:48:32.391621 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jan 24 11:48:32.852338 locksmithd[1682]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jan 24 11:48:32.861961 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jan 24 11:48:32.879764 systemd[1]: Started getty@tty1.service - Getty on tty1. Jan 24 11:48:32.890955 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jan 24 11:48:32.919393 systemd[1]: Reached target getty.target - Login Prompts. Jan 24 11:48:33.877541 containerd[1647]: time="2026-01-24T11:48:33Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Jan 24 11:48:33.886525 containerd[1647]: time="2026-01-24T11:48:33.886480297Z" level=info msg="starting containerd" revision=fcd43222d6b07379a4be9786bda52438f0dd16a1 version=v2.1.5 Jan 24 11:48:34.253811 containerd[1647]: time="2026-01-24T11:48:34.252392050Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="219.78µs" Jan 24 11:48:34.253811 containerd[1647]: time="2026-01-24T11:48:34.252489151Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Jan 24 11:48:34.253811 containerd[1647]: time="2026-01-24T11:48:34.253106694Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Jan 24 11:48:34.253811 containerd[1647]: time="2026-01-24T11:48:34.253159994Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Jan 24 11:48:34.255254 containerd[1647]: time="2026-01-24T11:48:34.254438951Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Jan 24 11:48:34.255254 containerd[1647]: time="2026-01-24T11:48:34.254489606Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jan 24 11:48:34.255254 containerd[1647]: time="2026-01-24T11:48:34.254631190Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jan 24 11:48:34.255254 containerd[1647]: time="2026-01-24T11:48:34.254651067Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jan 24 11:48:34.255254 containerd[1647]: time="2026-01-24T11:48:34.255237242Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jan 24 11:48:34.256177 containerd[1647]: time="2026-01-24T11:48:34.255322952Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jan 24 11:48:34.256177 containerd[1647]: time="2026-01-24T11:48:34.255348169Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jan 24 11:48:34.256177 containerd[1647]: time="2026-01-24T11:48:34.255360021Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Jan 24 11:48:34.259412 containerd[1647]: time="2026-01-24T11:48:34.259253231Z" level=info msg="skip loading plugin" error="EROFS unsupported, please `modprobe erofs`: skip plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Jan 24 11:48:34.259412 containerd[1647]: time="2026-01-24T11:48:34.259376882Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Jan 24 11:48:34.260515 containerd[1647]: time="2026-01-24T11:48:34.259854894Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Jan 24 11:48:34.260515 containerd[1647]: time="2026-01-24T11:48:34.260432102Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jan 24 11:48:34.260515 containerd[1647]: time="2026-01-24T11:48:34.260501361Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jan 24 11:48:34.260515 containerd[1647]: time="2026-01-24T11:48:34.260517110Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Jan 24 11:48:34.261006 containerd[1647]: time="2026-01-24T11:48:34.260917769Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Jan 24 11:48:34.262780 containerd[1647]: time="2026-01-24T11:48:34.262703062Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Jan 24 11:48:34.263175 containerd[1647]: time="2026-01-24T11:48:34.263150226Z" level=info msg="metadata content store policy set" policy=shared Jan 24 11:48:34.291843 containerd[1647]: time="2026-01-24T11:48:34.288613965Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Jan 24 11:48:34.291843 containerd[1647]: time="2026-01-24T11:48:34.289222351Z" level=info msg="loading plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Jan 24 11:48:34.291843 containerd[1647]: time="2026-01-24T11:48:34.289491594Z" level=info msg="skip loading plugin" error="could not find mkfs.erofs: exec: \"mkfs.erofs\": executable file not found in $PATH: skip plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Jan 24 11:48:34.291843 containerd[1647]: time="2026-01-24T11:48:34.289525307Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Jan 24 11:48:34.291843 containerd[1647]: time="2026-01-24T11:48:34.289546156Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Jan 24 11:48:34.291843 containerd[1647]: time="2026-01-24T11:48:34.289563078Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Jan 24 11:48:34.291843 containerd[1647]: time="2026-01-24T11:48:34.289581612Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Jan 24 11:48:34.291843 containerd[1647]: time="2026-01-24T11:48:34.289594566Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Jan 24 11:48:34.291843 containerd[1647]: time="2026-01-24T11:48:34.289610155Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Jan 24 11:48:34.291843 containerd[1647]: time="2026-01-24T11:48:34.289627889Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Jan 24 11:48:34.291843 containerd[1647]: time="2026-01-24T11:48:34.289696417Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Jan 24 11:48:34.291843 containerd[1647]: time="2026-01-24T11:48:34.289717646Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Jan 24 11:48:34.291843 containerd[1647]: time="2026-01-24T11:48:34.289770104Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Jan 24 11:48:34.291843 containerd[1647]: time="2026-01-24T11:48:34.289793768Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.290114687Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.290145665Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.290170031Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.290463859Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.290491571Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.290506218Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.290525624Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.290544610Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.290588011Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.290609220Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.290622937Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.290709648Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.290914361Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.290936151Z" level=info msg="Start snapshots syncer" Jan 24 11:48:34.292481 containerd[1647]: time="2026-01-24T11:48:34.291060473Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Jan 24 11:48:34.296110 containerd[1647]: time="2026-01-24T11:48:34.295914536Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"cgroupWritable\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"\",\"binDirs\":[\"/opt/cni/bin\"],\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogLineSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Jan 24 11:48:34.296110 containerd[1647]: time="2026-01-24T11:48:34.296068362Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Jan 24 11:48:34.296700 containerd[1647]: time="2026-01-24T11:48:34.296249180Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Jan 24 11:48:34.296700 containerd[1647]: time="2026-01-24T11:48:34.296570079Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Jan 24 11:48:34.296700 containerd[1647]: time="2026-01-24T11:48:34.296601317Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Jan 24 11:48:34.296700 containerd[1647]: time="2026-01-24T11:48:34.296618940Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Jan 24 11:48:34.296700 containerd[1647]: time="2026-01-24T11:48:34.296635711Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Jan 24 11:48:34.296700 containerd[1647]: time="2026-01-24T11:48:34.296656741Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Jan 24 11:48:34.296700 containerd[1647]: time="2026-01-24T11:48:34.296672971Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Jan 24 11:48:34.296700 containerd[1647]: time="2026-01-24T11:48:34.296687298Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Jan 24 11:48:34.296700 containerd[1647]: time="2026-01-24T11:48:34.296701485Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Jan 24 11:48:34.297032 containerd[1647]: time="2026-01-24T11:48:34.296719027Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Jan 24 11:48:34.297032 containerd[1647]: time="2026-01-24T11:48:34.296860712Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jan 24 11:48:34.297032 containerd[1647]: time="2026-01-24T11:48:34.296889706Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jan 24 11:48:34.297032 containerd[1647]: time="2026-01-24T11:48:34.296903121Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jan 24 11:48:34.297032 containerd[1647]: time="2026-01-24T11:48:34.296918630Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jan 24 11:48:34.297032 containerd[1647]: time="2026-01-24T11:48:34.296931234Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Jan 24 11:48:34.320453 containerd[1647]: time="2026-01-24T11:48:34.320379620Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Jan 24 11:48:34.320590 containerd[1647]: time="2026-01-24T11:48:34.320465230Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Jan 24 11:48:34.320820 containerd[1647]: time="2026-01-24T11:48:34.320725616Z" level=info msg="runtime interface created" Jan 24 11:48:34.320870 containerd[1647]: time="2026-01-24T11:48:34.320822617Z" level=info msg="created NRI interface" Jan 24 11:48:34.320870 containerd[1647]: time="2026-01-24T11:48:34.320846402Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Jan 24 11:48:34.320920 containerd[1647]: time="2026-01-24T11:48:34.320878892Z" level=info msg="Connect containerd service" Jan 24 11:48:34.321016 containerd[1647]: time="2026-01-24T11:48:34.320958892Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jan 24 11:48:34.330839 containerd[1647]: time="2026-01-24T11:48:34.330717580Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 24 11:48:35.160771 tar[1633]: linux-amd64/README.md Jan 24 11:48:35.568695 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jan 24 11:48:35.894543 containerd[1647]: time="2026-01-24T11:48:35.887182283Z" level=info msg="Start subscribing containerd event" Jan 24 11:48:36.069592 containerd[1647]: time="2026-01-24T11:48:35.898526226Z" level=info msg="Start recovering state" Jan 24 11:48:36.069592 containerd[1647]: time="2026-01-24T11:48:35.895550190Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jan 24 11:48:36.069592 containerd[1647]: time="2026-01-24T11:48:35.966097768Z" level=info msg=serving... address=/run/containerd/containerd.sock Jan 24 11:48:36.069592 containerd[1647]: time="2026-01-24T11:48:35.966339106Z" level=info msg="Start event monitor" Jan 24 11:48:36.069592 containerd[1647]: time="2026-01-24T11:48:35.966800981Z" level=info msg="Start cni network conf syncer for default" Jan 24 11:48:36.069592 containerd[1647]: time="2026-01-24T11:48:35.966927467Z" level=info msg="Start streaming server" Jan 24 11:48:36.069592 containerd[1647]: time="2026-01-24T11:48:35.966946793Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Jan 24 11:48:36.069592 containerd[1647]: time="2026-01-24T11:48:35.966956351Z" level=info msg="runtime interface starting up..." Jan 24 11:48:36.069592 containerd[1647]: time="2026-01-24T11:48:35.966965167Z" level=info msg="starting plugins..." Jan 24 11:48:36.069592 containerd[1647]: time="2026-01-24T11:48:35.971381322Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Jan 24 11:48:36.145601 systemd[1]: Started containerd.service - containerd container runtime. Jan 24 11:48:36.151117 containerd[1647]: time="2026-01-24T11:48:36.150400687Z" level=info msg="containerd successfully booted in 2.276401s" Jan 24 11:48:37.851645 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jan 24 11:48:37.863801 systemd[1]: Started sshd@0-10.0.0.67:22-10.0.0.1:46092.service - OpenSSH per-connection server daemon (10.0.0.1:46092). Jan 24 11:48:39.539172 sshd[1731]: Accepted publickey for core from 10.0.0.1 port 46092 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:48:39.547964 sshd-session[1731]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:48:39.587802 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jan 24 11:48:39.590109 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jan 24 11:48:39.615637 systemd-logind[1621]: New session 1 of user core. Jan 24 11:48:39.663039 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jan 24 11:48:39.676750 systemd[1]: Starting user@500.service - User Manager for UID 500... Jan 24 11:48:39.726840 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:48:39.728454 systemd[1]: Reached target multi-user.target - Multi-User System. Jan 24 11:48:39.735860 (kubelet)[1747]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:48:39.739355 (systemd)[1743]: pam_unix(systemd-user:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:48:39.749864 systemd-logind[1621]: New session 2 of user core. Jan 24 11:48:40.921852 systemd[1743]: Queued start job for default target default.target. Jan 24 11:48:40.936101 systemd[1743]: Created slice app.slice - User Application Slice. Jan 24 11:48:40.936156 systemd[1743]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of User's Temporary Directories. Jan 24 11:48:40.936184 systemd[1743]: Reached target paths.target - Paths. Jan 24 11:48:40.937046 systemd[1743]: Reached target timers.target - Timers. Jan 24 11:48:40.943242 systemd[1743]: Starting dbus.socket - D-Bus User Message Bus Socket... Jan 24 11:48:40.947183 systemd[1743]: Starting systemd-tmpfiles-setup.service - Create User Files and Directories... Jan 24 11:48:41.337591 systemd[1743]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jan 24 11:48:41.337867 systemd[1743]: Reached target sockets.target - Sockets. Jan 24 11:48:41.360409 systemd[1743]: Finished systemd-tmpfiles-setup.service - Create User Files and Directories. Jan 24 11:48:41.360652 systemd[1743]: Reached target basic.target - Basic System. Jan 24 11:48:41.362109 systemd[1]: Started user@500.service - User Manager for UID 500. Jan 24 11:48:41.364140 systemd[1743]: Reached target default.target - Main User Target. Jan 24 11:48:41.364704 systemd[1743]: Startup finished in 1.590s. Jan 24 11:48:41.381611 systemd[1]: Started session-1.scope - Session 1 of User core. Jan 24 11:48:41.382989 systemd[1]: Startup finished in 12.376s (kernel) + 30.569s (initrd) + 27.532s (userspace) = 1min 10.478s. Jan 24 11:48:41.457256 systemd[1]: Started sshd@1-10.0.0.67:22-10.0.0.1:46108.service - OpenSSH per-connection server daemon (10.0.0.1:46108). Jan 24 11:48:41.742247 sshd[1766]: Accepted publickey for core from 10.0.0.1 port 46108 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:48:41.745888 sshd-session[1766]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:48:41.791904 systemd-logind[1621]: New session 3 of user core. Jan 24 11:48:41.832570 systemd[1]: Started session-3.scope - Session 3 of User core. Jan 24 11:48:41.882731 sshd[1770]: Connection closed by 10.0.0.1 port 46108 Jan 24 11:48:41.879352 sshd-session[1766]: pam_unix(sshd:session): session closed for user core Jan 24 11:48:41.894202 systemd[1]: sshd@1-10.0.0.67:22-10.0.0.1:46108.service: Deactivated successfully. Jan 24 11:48:41.897499 systemd[1]: session-3.scope: Deactivated successfully. Jan 24 11:48:41.902602 systemd-logind[1621]: Session 3 logged out. Waiting for processes to exit. Jan 24 11:48:41.912030 systemd[1]: Started sshd@2-10.0.0.67:22-10.0.0.1:46116.service - OpenSSH per-connection server daemon (10.0.0.1:46116). Jan 24 11:48:41.913258 systemd-logind[1621]: Removed session 3. Jan 24 11:48:42.036621 sshd[1776]: Accepted publickey for core from 10.0.0.1 port 46116 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:48:42.044688 sshd-session[1776]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:48:42.068720 systemd-logind[1621]: New session 4 of user core. Jan 24 11:48:42.087586 systemd[1]: Started session-4.scope - Session 4 of User core. Jan 24 11:48:42.111959 sshd[1781]: Connection closed by 10.0.0.1 port 46116 Jan 24 11:48:42.112425 sshd-session[1776]: pam_unix(sshd:session): session closed for user core Jan 24 11:48:42.125057 systemd[1]: sshd@2-10.0.0.67:22-10.0.0.1:46116.service: Deactivated successfully. Jan 24 11:48:42.128126 systemd[1]: session-4.scope: Deactivated successfully. Jan 24 11:48:42.132858 systemd-logind[1621]: Session 4 logged out. Waiting for processes to exit. Jan 24 11:48:42.136131 systemd-logind[1621]: Removed session 4. Jan 24 11:48:42.138233 systemd[1]: Started sshd@3-10.0.0.67:22-10.0.0.1:46132.service - OpenSSH per-connection server daemon (10.0.0.1:46132). Jan 24 11:48:42.235859 sshd[1787]: Accepted publickey for core from 10.0.0.1 port 46132 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:48:42.239433 sshd-session[1787]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:48:42.257034 systemd-logind[1621]: New session 5 of user core. Jan 24 11:48:42.264129 systemd[1]: Started session-5.scope - Session 5 of User core. Jan 24 11:48:42.293099 sshd[1791]: Connection closed by 10.0.0.1 port 46132 Jan 24 11:48:42.294653 sshd-session[1787]: pam_unix(sshd:session): session closed for user core Jan 24 11:48:42.315042 systemd[1]: sshd@3-10.0.0.67:22-10.0.0.1:46132.service: Deactivated successfully. Jan 24 11:48:42.319069 systemd[1]: session-5.scope: Deactivated successfully. Jan 24 11:48:42.323038 systemd-logind[1621]: Session 5 logged out. Waiting for processes to exit. Jan 24 11:48:42.327875 systemd[1]: Started sshd@4-10.0.0.67:22-10.0.0.1:51242.service - OpenSSH per-connection server daemon (10.0.0.1:51242). Jan 24 11:48:42.329414 systemd-logind[1621]: Removed session 5. Jan 24 11:48:42.389930 kubelet[1747]: E0124 11:48:42.389756 1747 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:48:42.395527 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:48:42.395804 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:48:42.396806 systemd[1]: kubelet.service: Consumed 6.631s CPU time, 266.1M memory peak. Jan 24 11:48:42.420691 sshd[1798]: Accepted publickey for core from 10.0.0.1 port 51242 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:48:42.427874 sshd-session[1798]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:48:42.441563 systemd-logind[1621]: New session 6 of user core. Jan 24 11:48:42.451690 systemd[1]: Started session-6.scope - Session 6 of User core. Jan 24 11:48:42.494176 sudo[1804]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jan 24 11:48:42.494690 sudo[1804]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 24 11:48:42.512681 sudo[1804]: pam_unix(sudo:session): session closed for user root Jan 24 11:48:42.516063 sshd[1803]: Connection closed by 10.0.0.1 port 51242 Jan 24 11:48:42.517563 sshd-session[1798]: pam_unix(sshd:session): session closed for user core Jan 24 11:48:42.536987 systemd[1]: sshd@4-10.0.0.67:22-10.0.0.1:51242.service: Deactivated successfully. Jan 24 11:48:42.539764 systemd[1]: session-6.scope: Deactivated successfully. Jan 24 11:48:42.546557 systemd-logind[1621]: Session 6 logged out. Waiting for processes to exit. Jan 24 11:48:42.550970 systemd[1]: Started sshd@5-10.0.0.67:22-10.0.0.1:51256.service - OpenSSH per-connection server daemon (10.0.0.1:51256). Jan 24 11:48:42.558860 systemd-logind[1621]: Removed session 6. Jan 24 11:48:42.632010 sshd[1811]: Accepted publickey for core from 10.0.0.1 port 51256 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:48:42.634329 sshd-session[1811]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:48:42.648747 systemd-logind[1621]: New session 7 of user core. Jan 24 11:48:42.667047 systemd[1]: Started session-7.scope - Session 7 of User core. Jan 24 11:48:42.711482 sudo[1818]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jan 24 11:48:42.712080 sudo[1818]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 24 11:48:42.756378 sudo[1818]: pam_unix(sudo:session): session closed for user root Jan 24 11:48:42.778246 sudo[1817]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Jan 24 11:48:42.778996 sudo[1817]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 24 11:48:42.798392 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 24 11:48:42.909000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Jan 24 11:48:42.913250 kernel: kauditd_printk_skb: 64 callbacks suppressed Jan 24 11:48:42.913478 kernel: audit: type=1305 audit(1769255322.909:216): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Jan 24 11:48:42.915048 augenrules[1842]: No rules Jan 24 11:48:42.919678 systemd[1]: audit-rules.service: Deactivated successfully. Jan 24 11:48:42.909000 audit[1842]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7fff1a847df0 a2=420 a3=0 items=0 ppid=1823 pid=1842 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:42.928608 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 24 11:48:42.943228 kernel: audit: type=1300 audit(1769255322.909:216): arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7fff1a847df0 a2=420 a3=0 items=0 ppid=1823 pid=1842 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:42.943358 kernel: audit: type=1327 audit(1769255322.909:216): proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 24 11:48:42.909000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 24 11:48:42.941341 sudo[1817]: pam_unix(sudo:session): session closed for user root Jan 24 11:48:42.930000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:42.948748 sshd[1816]: Connection closed by 10.0.0.1 port 51256 Jan 24 11:48:42.949605 sshd-session[1811]: pam_unix(sshd:session): session closed for user core Jan 24 11:48:42.956117 kernel: audit: type=1130 audit(1769255322.930:217): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:42.930000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:42.975100 kernel: audit: type=1131 audit(1769255322.930:218): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:42.977087 kernel: audit: type=1106 audit(1769255322.940:219): pid=1817 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:48:42.977161 kernel: audit: type=1104 audit(1769255322.940:220): pid=1817 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:48:42.940000 audit[1817]: USER_END pid=1817 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:48:42.940000 audit[1817]: CRED_DISP pid=1817 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:48:42.946000 audit[1811]: USER_END pid=1811 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:48:42.987901 kernel: audit: type=1106 audit(1769255322.946:221): pid=1811 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:48:42.946000 audit[1811]: CRED_DISP pid=1811 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:48:43.036837 kernel: audit: type=1104 audit(1769255322.946:222): pid=1811 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:48:43.068000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@5-10.0.0.67:22-10.0.0.1:51256 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:43.070506 systemd[1]: sshd@5-10.0.0.67:22-10.0.0.1:51256.service: Deactivated successfully. Jan 24 11:48:43.106380 kernel: audit: type=1131 audit(1769255323.068:223): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@5-10.0.0.67:22-10.0.0.1:51256 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:43.140956 systemd[1]: session-7.scope: Deactivated successfully. Jan 24 11:48:43.225765 systemd-logind[1621]: Session 7 logged out. Waiting for processes to exit. Jan 24 11:48:43.252000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-10.0.0.67:22-10.0.0.1:51262 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:43.254716 systemd[1]: Started sshd@6-10.0.0.67:22-10.0.0.1:51262.service - OpenSSH per-connection server daemon (10.0.0.1:51262). Jan 24 11:48:43.284105 systemd-logind[1621]: Removed session 7. Jan 24 11:48:43.470761 sshd[1851]: Accepted publickey for core from 10.0.0.1 port 51262 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:48:43.469000 audit[1851]: USER_ACCT pid=1851 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:48:43.473000 audit[1851]: CRED_ACQ pid=1851 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:48:43.473000 audit[1851]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd400faca0 a2=3 a3=0 items=0 ppid=1 pid=1851 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=8 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:43.473000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:48:43.477727 sshd-session[1851]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:48:43.527123 systemd-logind[1621]: New session 8 of user core. Jan 24 11:48:43.541194 systemd[1]: Started session-8.scope - Session 8 of User core. Jan 24 11:48:43.560000 audit[1851]: USER_START pid=1851 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:48:43.567000 audit[1855]: CRED_ACQ pid=1855 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:48:43.593000 audit[1856]: USER_ACCT pid=1856 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_unix,pam_faillock acct="core" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:48:43.623000 audit[1856]: CRED_REFR pid=1856 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:48:43.606199 sudo[1856]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jan 24 11:48:43.624897 sudo[1856]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 24 11:48:43.627000 audit[1856]: USER_START pid=1856 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:48:47.714756 systemd[1]: Starting docker.service - Docker Application Container Engine... Jan 24 11:48:47.771459 (dockerd)[1878]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jan 24 11:48:50.930675 dockerd[1878]: time="2026-01-24T11:48:50.928692493Z" level=info msg="Starting up" Jan 24 11:48:50.934636 dockerd[1878]: time="2026-01-24T11:48:50.934562121Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Jan 24 11:48:51.018920 dockerd[1878]: time="2026-01-24T11:48:51.018555815Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Jan 24 11:48:51.153585 dockerd[1878]: time="2026-01-24T11:48:51.153248786Z" level=info msg="Loading containers: start." Jan 24 11:48:51.181354 kernel: Initializing XFRM netlink socket Jan 24 11:48:51.760000 audit[1932]: NETFILTER_CFG table=nat:2 family=2 entries=2 op=nft_register_chain pid=1932 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.764447 kernel: kauditd_printk_skb: 11 callbacks suppressed Jan 24 11:48:51.764674 kernel: audit: type=1325 audit(1769255331.760:233): table=nat:2 family=2 entries=2 op=nft_register_chain pid=1932 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.760000 audit[1932]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffd29bcbc70 a2=0 a3=0 items=0 ppid=1878 pid=1932 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.784360 kernel: audit: type=1300 audit(1769255331.760:233): arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffd29bcbc70 a2=0 a3=0 items=0 ppid=1878 pid=1932 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.760000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Jan 24 11:48:51.785319 kernel: audit: type=1327 audit(1769255331.760:233): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Jan 24 11:48:51.765000 audit[1934]: NETFILTER_CFG table=filter:3 family=2 entries=2 op=nft_register_chain pid=1934 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.796906 kernel: audit: type=1325 audit(1769255331.765:234): table=filter:3 family=2 entries=2 op=nft_register_chain pid=1934 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.796988 kernel: audit: type=1300 audit(1769255331.765:234): arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffe25cbf4a0 a2=0 a3=0 items=0 ppid=1878 pid=1934 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.765000 audit[1934]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffe25cbf4a0 a2=0 a3=0 items=0 ppid=1878 pid=1934 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.825226 kernel: audit: type=1327 audit(1769255331.765:234): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Jan 24 11:48:51.765000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Jan 24 11:48:51.770000 audit[1936]: NETFILTER_CFG table=filter:4 family=2 entries=1 op=nft_register_chain pid=1936 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.836986 kernel: audit: type=1325 audit(1769255331.770:235): table=filter:4 family=2 entries=1 op=nft_register_chain pid=1936 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.837188 kernel: audit: type=1300 audit(1769255331.770:235): arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe8109e1d0 a2=0 a3=0 items=0 ppid=1878 pid=1936 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.770000 audit[1936]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe8109e1d0 a2=0 a3=0 items=0 ppid=1878 pid=1936 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.849997 kernel: audit: type=1327 audit(1769255331.770:235): proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Jan 24 11:48:51.770000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Jan 24 11:48:51.855482 kernel: audit: type=1325 audit(1769255331.775:236): table=filter:5 family=2 entries=1 op=nft_register_chain pid=1938 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.775000 audit[1938]: NETFILTER_CFG table=filter:5 family=2 entries=1 op=nft_register_chain pid=1938 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.775000 audit[1938]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd738d7ec0 a2=0 a3=0 items=0 ppid=1878 pid=1938 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.775000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D425249444745 Jan 24 11:48:51.779000 audit[1940]: NETFILTER_CFG table=filter:6 family=2 entries=1 op=nft_register_chain pid=1940 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.779000 audit[1940]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffd0e995150 a2=0 a3=0 items=0 ppid=1878 pid=1940 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.779000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D4354 Jan 24 11:48:51.783000 audit[1942]: NETFILTER_CFG table=filter:7 family=2 entries=1 op=nft_register_chain pid=1942 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.783000 audit[1942]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffccb3bbc00 a2=0 a3=0 items=0 ppid=1878 pid=1942 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.783000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 24 11:48:51.788000 audit[1944]: NETFILTER_CFG table=filter:8 family=2 entries=1 op=nft_register_chain pid=1944 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.788000 audit[1944]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffee8558840 a2=0 a3=0 items=0 ppid=1878 pid=1944 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.788000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 24 11:48:51.790000 audit[1946]: NETFILTER_CFG table=nat:9 family=2 entries=2 op=nft_register_chain pid=1946 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.790000 audit[1946]: SYSCALL arch=c000003e syscall=46 success=yes exit=384 a0=3 a1=7ffc777ef4b0 a2=0 a3=0 items=0 ppid=1878 pid=1946 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.790000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Jan 24 11:48:51.933000 audit[1949]: NETFILTER_CFG table=nat:10 family=2 entries=2 op=nft_register_chain pid=1949 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.933000 audit[1949]: SYSCALL arch=c000003e syscall=46 success=yes exit=472 a0=3 a1=7ffca5926460 a2=0 a3=0 items=0 ppid=1878 pid=1949 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.933000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003132372E302E302E302F38 Jan 24 11:48:51.938000 audit[1951]: NETFILTER_CFG table=filter:11 family=2 entries=2 op=nft_register_chain pid=1951 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.938000 audit[1951]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffc96608080 a2=0 a3=0 items=0 ppid=1878 pid=1951 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.938000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D464F5257415244 Jan 24 11:48:51.943000 audit[1953]: NETFILTER_CFG table=filter:12 family=2 entries=1 op=nft_register_rule pid=1953 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.943000 audit[1953]: SYSCALL arch=c000003e syscall=46 success=yes exit=236 a0=3 a1=7fff9a547080 a2=0 a3=0 items=0 ppid=1878 pid=1953 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.943000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D425249444745 Jan 24 11:48:51.951000 audit[1955]: NETFILTER_CFG table=filter:13 family=2 entries=1 op=nft_register_rule pid=1955 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.951000 audit[1955]: SYSCALL arch=c000003e syscall=46 success=yes exit=248 a0=3 a1=7ffc7cf32190 a2=0 a3=0 items=0 ppid=1878 pid=1955 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.951000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 24 11:48:51.956000 audit[1957]: NETFILTER_CFG table=filter:14 family=2 entries=1 op=nft_register_rule pid=1957 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:51.956000 audit[1957]: SYSCALL arch=c000003e syscall=46 success=yes exit=232 a0=3 a1=7ffefb5287e0 a2=0 a3=0 items=0 ppid=1878 pid=1957 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:51.956000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D4354 Jan 24 11:48:52.470000 audit[1987]: NETFILTER_CFG table=nat:15 family=10 entries=2 op=nft_register_chain pid=1987 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.470000 audit[1987]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7fff86b8eb80 a2=0 a3=0 items=0 ppid=1878 pid=1987 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.470000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Jan 24 11:48:52.480000 audit[1989]: NETFILTER_CFG table=filter:16 family=10 entries=2 op=nft_register_chain pid=1989 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.480000 audit[1989]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffd5bee0f20 a2=0 a3=0 items=0 ppid=1878 pid=1989 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.480000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Jan 24 11:48:52.493000 audit[1991]: NETFILTER_CFG table=filter:17 family=10 entries=1 op=nft_register_chain pid=1991 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.493000 audit[1991]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe692478c0 a2=0 a3=0 items=0 ppid=1878 pid=1991 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.493000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Jan 24 11:48:52.500000 audit[1993]: NETFILTER_CFG table=filter:18 family=10 entries=1 op=nft_register_chain pid=1993 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.500000 audit[1993]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff298cba60 a2=0 a3=0 items=0 ppid=1878 pid=1993 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.500000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D425249444745 Jan 24 11:48:52.526000 audit[1995]: NETFILTER_CFG table=filter:19 family=10 entries=1 op=nft_register_chain pid=1995 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.526000 audit[1995]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffdf47f36f0 a2=0 a3=0 items=0 ppid=1878 pid=1995 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.526000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D4354 Jan 24 11:48:52.531000 audit[1997]: NETFILTER_CFG table=filter:20 family=10 entries=1 op=nft_register_chain pid=1997 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.531000 audit[1997]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffc32716c70 a2=0 a3=0 items=0 ppid=1878 pid=1997 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.531000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 24 11:48:52.540000 audit[1999]: NETFILTER_CFG table=filter:21 family=10 entries=1 op=nft_register_chain pid=1999 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.540000 audit[1999]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffdde2026b0 a2=0 a3=0 items=0 ppid=1878 pid=1999 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.540000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 24 11:48:52.549000 audit[2001]: NETFILTER_CFG table=nat:22 family=10 entries=2 op=nft_register_chain pid=2001 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.549000 audit[2001]: SYSCALL arch=c000003e syscall=46 success=yes exit=384 a0=3 a1=7ffdfa327f20 a2=0 a3=0 items=0 ppid=1878 pid=2001 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.549000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Jan 24 11:48:52.562000 audit[2003]: NETFILTER_CFG table=nat:23 family=10 entries=2 op=nft_register_chain pid=2003 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.562000 audit[2003]: SYSCALL arch=c000003e syscall=46 success=yes exit=484 a0=3 a1=7ffe02d20210 a2=0 a3=0 items=0 ppid=1878 pid=2003 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.562000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003A3A312F313238 Jan 24 11:48:52.567000 audit[2005]: NETFILTER_CFG table=filter:24 family=10 entries=2 op=nft_register_chain pid=2005 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.567000 audit[2005]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffe3043ef70 a2=0 a3=0 items=0 ppid=1878 pid=2005 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.567000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D464F5257415244 Jan 24 11:48:52.573000 audit[2007]: NETFILTER_CFG table=filter:25 family=10 entries=1 op=nft_register_rule pid=2007 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.573000 audit[2007]: SYSCALL arch=c000003e syscall=46 success=yes exit=236 a0=3 a1=7ffece2a1f70 a2=0 a3=0 items=0 ppid=1878 pid=2007 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.573000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D425249444745 Jan 24 11:48:52.576444 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jan 24 11:48:52.584622 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:48:52.587000 audit[2010]: NETFILTER_CFG table=filter:26 family=10 entries=1 op=nft_register_rule pid=2010 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.587000 audit[2010]: SYSCALL arch=c000003e syscall=46 success=yes exit=248 a0=3 a1=7ffe5f246930 a2=0 a3=0 items=0 ppid=1878 pid=2010 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.587000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 24 11:48:52.593000 audit[2012]: NETFILTER_CFG table=filter:27 family=10 entries=1 op=nft_register_rule pid=2012 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.593000 audit[2012]: SYSCALL arch=c000003e syscall=46 success=yes exit=232 a0=3 a1=7ffc3cad4eb0 a2=0 a3=0 items=0 ppid=1878 pid=2012 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.593000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D4354 Jan 24 11:48:52.632000 audit[2019]: NETFILTER_CFG table=filter:28 family=2 entries=1 op=nft_register_chain pid=2019 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:52.632000 audit[2019]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffcc9b42c20 a2=0 a3=0 items=0 ppid=1878 pid=2019 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.632000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Jan 24 11:48:52.656000 audit[2021]: NETFILTER_CFG table=filter:29 family=2 entries=1 op=nft_register_rule pid=2021 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:52.656000 audit[2021]: SYSCALL arch=c000003e syscall=46 success=yes exit=212 a0=3 a1=7ffdbc450740 a2=0 a3=0 items=0 ppid=1878 pid=2021 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.656000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E Jan 24 11:48:52.663000 audit[2023]: NETFILTER_CFG table=filter:30 family=2 entries=1 op=nft_register_rule pid=2023 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:52.663000 audit[2023]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7fffbb2e9080 a2=0 a3=0 items=0 ppid=1878 pid=2023 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.663000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Jan 24 11:48:52.668000 audit[2025]: NETFILTER_CFG table=filter:31 family=10 entries=1 op=nft_register_chain pid=2025 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.668000 audit[2025]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffc44cba770 a2=0 a3=0 items=0 ppid=1878 pid=2025 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.668000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Jan 24 11:48:52.678000 audit[2027]: NETFILTER_CFG table=filter:32 family=10 entries=1 op=nft_register_rule pid=2027 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.678000 audit[2027]: SYSCALL arch=c000003e syscall=46 success=yes exit=212 a0=3 a1=7ffec610a520 a2=0 a3=0 items=0 ppid=1878 pid=2027 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.678000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E Jan 24 11:48:52.687000 audit[2029]: NETFILTER_CFG table=filter:33 family=10 entries=1 op=nft_register_rule pid=2029 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:48:52.687000 audit[2029]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7fff47f71460 a2=0 a3=0 items=0 ppid=1878 pid=2029 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.687000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Jan 24 11:48:52.776000 audit[2033]: NETFILTER_CFG table=nat:34 family=2 entries=2 op=nft_register_chain pid=2033 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:52.776000 audit[2033]: SYSCALL arch=c000003e syscall=46 success=yes exit=520 a0=3 a1=7ffc1c444830 a2=0 a3=0 items=0 ppid=1878 pid=2033 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.776000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900504F5354524F5554494E47002D73003137322E31372E302E302F31360000002D6F00646F636B657230002D6A004D415351554552414445 Jan 24 11:48:52.790000 audit[2035]: NETFILTER_CFG table=nat:35 family=2 entries=1 op=nft_register_rule pid=2035 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:52.790000 audit[2035]: SYSCALL arch=c000003e syscall=46 success=yes exit=288 a0=3 a1=7fff03c1e1b0 a2=0 a3=0 items=0 ppid=1878 pid=2035 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.790000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900444F434B4552002D6900646F636B657230002D6A0052455455524E Jan 24 11:48:52.821000 audit[2043]: NETFILTER_CFG table=filter:36 family=2 entries=1 op=nft_register_rule pid=2043 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:52.821000 audit[2043]: SYSCALL arch=c000003e syscall=46 success=yes exit=300 a0=3 a1=7ffce5848940 a2=0 a3=0 items=0 ppid=1878 pid=2043 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.821000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D464F5257415244002D6900646F636B657230002D6A00414343455054 Jan 24 11:48:52.853000 audit[2049]: NETFILTER_CFG table=filter:37 family=2 entries=1 op=nft_register_rule pid=2049 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:52.853000 audit[2049]: SYSCALL arch=c000003e syscall=46 success=yes exit=376 a0=3 a1=7ffe98bcc8c0 a2=0 a3=0 items=0 ppid=1878 pid=2049 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.853000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45520000002D6900646F636B657230002D6F00646F636B657230002D6A0044524F50 Jan 24 11:48:52.864000 audit[2051]: NETFILTER_CFG table=filter:38 family=2 entries=1 op=nft_register_rule pid=2051 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:52.864000 audit[2051]: SYSCALL arch=c000003e syscall=46 success=yes exit=512 a0=3 a1=7ffe3c0b7670 a2=0 a3=0 items=0 ppid=1878 pid=2051 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.864000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D4354002D6F00646F636B657230002D6D00636F6E6E747261636B002D2D637473746174650052454C415445442C45535441424C4953484544002D6A00414343455054 Jan 24 11:48:52.879000 audit[2053]: NETFILTER_CFG table=filter:39 family=2 entries=1 op=nft_register_rule pid=2053 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:52.879000 audit[2053]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffc1f0279a0 a2=0 a3=0 items=0 ppid=1878 pid=2053 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.879000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D425249444745002D6F00646F636B657230002D6A00444F434B4552 Jan 24 11:48:52.889000 audit[2055]: NETFILTER_CFG table=filter:40 family=2 entries=1 op=nft_register_rule pid=2055 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:52.889000 audit[2055]: SYSCALL arch=c000003e syscall=46 success=yes exit=428 a0=3 a1=7ffdcd448440 a2=0 a3=0 items=0 ppid=1878 pid=2055 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.889000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D49534F4C4154494F4E2D53544147452D31002D6900646F636B6572300000002D6F00646F636B657230002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 24 11:48:52.909000 audit[2057]: NETFILTER_CFG table=filter:41 family=2 entries=1 op=nft_register_rule pid=2057 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:48:52.909000 audit[2057]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffc9bc7e0a0 a2=0 a3=0 items=0 ppid=1878 pid=2057 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:48:52.909000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4900444F434B45522D49534F4C4154494F4E2D53544147452D32002D6F00646F636B657230002D6A0044524F50 Jan 24 11:48:52.917760 systemd-networkd[1544]: docker0: Link UP Jan 24 11:48:52.947898 dockerd[1878]: time="2026-01-24T11:48:52.947628957Z" level=info msg="Loading containers: done." Jan 24 11:48:53.046215 dockerd[1878]: time="2026-01-24T11:48:53.045644390Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jan 24 11:48:53.046215 dockerd[1878]: time="2026-01-24T11:48:53.046173578Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Jan 24 11:48:53.046580 dockerd[1878]: time="2026-01-24T11:48:53.046543679Z" level=info msg="Initializing buildkit" Jan 24 11:48:53.142771 dockerd[1878]: time="2026-01-24T11:48:53.141916820Z" level=info msg="Completed buildkit initialization" Jan 24 11:48:53.152687 dockerd[1878]: time="2026-01-24T11:48:53.152590226Z" level=info msg="Daemon has completed initialization" Jan 24 11:48:53.152932 dockerd[1878]: time="2026-01-24T11:48:53.152659867Z" level=info msg="API listen on /run/docker.sock" Jan 24 11:48:53.153247 systemd[1]: Started docker.service - Docker Application Container Engine. Jan 24 11:48:53.152000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=docker comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:53.714080 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:48:53.713000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:48:53.734926 (kubelet)[2100]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:48:53.917794 kubelet[2100]: E0124 11:48:53.917616 2100 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:48:53.925767 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:48:53.926134 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:48:53.925000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:48:53.927084 systemd[1]: kubelet.service: Consumed 842ms CPU time, 109.8M memory peak. Jan 24 11:48:57.125491 containerd[1647]: time="2026-01-24T11:48:57.123212312Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.7\"" Jan 24 11:48:58.068817 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1604643460.mount: Deactivated successfully. Jan 24 11:49:00.297567 containerd[1647]: time="2026-01-24T11:49:00.296070973Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:00.311214 containerd[1647]: time="2026-01-24T11:49:00.309950288Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.7: active requests=0, bytes read=29139078" Jan 24 11:49:00.336799 containerd[1647]: time="2026-01-24T11:49:00.331319203Z" level=info msg="ImageCreate event name:\"sha256:021d1ceeffb11df7a9fb9adfa0ad0a30dcd13cb3d630022066f184cdcb93731b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:00.367008 containerd[1647]: time="2026-01-24T11:49:00.365967864Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:9585226cb85d1dc0f0ef5f7a75f04e4bc91ddd82de249533bd293aa3cf958dab\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:00.393653 containerd[1647]: time="2026-01-24T11:49:00.392099411Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.7\" with image id \"sha256:021d1ceeffb11df7a9fb9adfa0ad0a30dcd13cb3d630022066f184cdcb93731b\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.7\", repo digest \"registry.k8s.io/kube-apiserver@sha256:9585226cb85d1dc0f0ef5f7a75f04e4bc91ddd82de249533bd293aa3cf958dab\", size \"30111311\" in 3.268574367s" Jan 24 11:49:00.393653 containerd[1647]: time="2026-01-24T11:49:00.392551494Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.7\" returns image reference \"sha256:021d1ceeffb11df7a9fb9adfa0ad0a30dcd13cb3d630022066f184cdcb93731b\"" Jan 24 11:49:00.399402 containerd[1647]: time="2026-01-24T11:49:00.399342418Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.7\"" Jan 24 11:49:04.091927 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jan 24 11:49:04.168542 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:49:06.182605 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:49:06.183000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:49:06.186792 kernel: kauditd_printk_skb: 113 callbacks suppressed Jan 24 11:49:06.226955 kernel: audit: type=1130 audit(1769255346.183:276): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:49:06.590130 (kubelet)[2186]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:49:08.853683 kubelet[2186]: E0124 11:49:08.853523 2186 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:49:08.871340 kernel: audit: type=1131 audit(1769255348.860:277): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:49:08.860000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:49:08.859137 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:49:08.859457 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:49:08.870835 systemd[1]: kubelet.service: Consumed 3.024s CPU time, 110.5M memory peak. Jan 24 11:49:10.245785 containerd[1647]: time="2026-01-24T11:49:10.244519376Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:10.248061 containerd[1647]: time="2026-01-24T11:49:10.247513049Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.7: active requests=0, bytes read=26008626" Jan 24 11:49:10.252884 containerd[1647]: time="2026-01-24T11:49:10.252774550Z" level=info msg="ImageCreate event name:\"sha256:29c7cab9d8e681d047281fd3711baf13c28f66923480fb11c8f22ddb7ca742d1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:10.263232 containerd[1647]: time="2026-01-24T11:49:10.262939967Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:f69d77ca0626b5a4b7b432c18de0952941181db7341c80eb89731f46d1d0c230\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:10.265862 containerd[1647]: time="2026-01-24T11:49:10.265743637Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.7\" with image id \"sha256:29c7cab9d8e681d047281fd3711baf13c28f66923480fb11c8f22ddb7ca742d1\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.7\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:f69d77ca0626b5a4b7b432c18de0952941181db7341c80eb89731f46d1d0c230\", size \"27673815\" in 9.86622814s" Jan 24 11:49:10.266172 containerd[1647]: time="2026-01-24T11:49:10.266085497Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.7\" returns image reference \"sha256:29c7cab9d8e681d047281fd3711baf13c28f66923480fb11c8f22ddb7ca742d1\"" Jan 24 11:49:10.285399 containerd[1647]: time="2026-01-24T11:49:10.284524462Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.7\"" Jan 24 11:49:14.851112 containerd[1647]: time="2026-01-24T11:49:14.850569681Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:14.853906 containerd[1647]: time="2026-01-24T11:49:14.852660406Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.7: active requests=0, bytes read=20149965" Jan 24 11:49:14.855418 containerd[1647]: time="2026-01-24T11:49:14.855319609Z" level=info msg="ImageCreate event name:\"sha256:f457f6fcd712acb5b9beef873f6f4a4869182f9eb52ea6e24824fd4ac4eed393\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:14.861021 containerd[1647]: time="2026-01-24T11:49:14.860172193Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:21bda321d8b4d48eb059fbc1593203d55d8b3bc7acd0584e04e55504796d78d0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:14.861292 containerd[1647]: time="2026-01-24T11:49:14.861108792Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.7\" with image id \"sha256:f457f6fcd712acb5b9beef873f6f4a4869182f9eb52ea6e24824fd4ac4eed393\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.7\", repo digest \"registry.k8s.io/kube-scheduler@sha256:21bda321d8b4d48eb059fbc1593203d55d8b3bc7acd0584e04e55504796d78d0\", size \"21815154\" in 4.576453838s" Jan 24 11:49:14.861292 containerd[1647]: time="2026-01-24T11:49:14.861165106Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.7\" returns image reference \"sha256:f457f6fcd712acb5b9beef873f6f4a4869182f9eb52ea6e24824fd4ac4eed393\"" Jan 24 11:49:14.863696 containerd[1647]: time="2026-01-24T11:49:14.863410374Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.7\"" Jan 24 11:49:16.935334 update_engine[1622]: I20260124 11:49:16.934182 1622 update_attempter.cc:509] Updating boot flags... Jan 24 11:49:19.078675 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Jan 24 11:49:19.082885 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:49:19.195084 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2343525298.mount: Deactivated successfully. Jan 24 11:49:20.038757 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:49:20.037000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:49:20.049786 kernel: audit: type=1130 audit(1769255360.037:278): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:49:20.056860 (kubelet)[2233]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:49:20.546378 kubelet[2233]: E0124 11:49:20.546175 2233 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:49:20.553407 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:49:20.553884 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:49:20.554000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:49:20.556479 systemd[1]: kubelet.service: Consumed 1.339s CPU time, 108.9M memory peak. Jan 24 11:49:20.566348 kernel: audit: type=1131 audit(1769255360.554:279): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:49:22.044954 containerd[1647]: time="2026-01-24T11:49:22.044041423Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:22.048054 containerd[1647]: time="2026-01-24T11:49:22.046844286Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.7: active requests=0, bytes read=31926374" Jan 24 11:49:22.054367 containerd[1647]: time="2026-01-24T11:49:22.054146810Z" level=info msg="ImageCreate event name:\"sha256:0929027b17fc30cb9de279f3bdba4e130b991a1dab7978a7db2e5feb2091853c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:22.058829 containerd[1647]: time="2026-01-24T11:49:22.058669115Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:ec25702b19026e9c0d339bc1c3bd231435a59f28b5fccb21e1b1078a357380f5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:22.059910 containerd[1647]: time="2026-01-24T11:49:22.059681963Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.7\" with image id \"sha256:0929027b17fc30cb9de279f3bdba4e130b991a1dab7978a7db2e5feb2091853c\", repo tag \"registry.k8s.io/kube-proxy:v1.33.7\", repo digest \"registry.k8s.io/kube-proxy@sha256:ec25702b19026e9c0d339bc1c3bd231435a59f28b5fccb21e1b1078a357380f5\", size \"31929115\" in 7.196224362s" Jan 24 11:49:22.059910 containerd[1647]: time="2026-01-24T11:49:22.059835999Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.7\" returns image reference \"sha256:0929027b17fc30cb9de279f3bdba4e130b991a1dab7978a7db2e5feb2091853c\"" Jan 24 11:49:22.061939 containerd[1647]: time="2026-01-24T11:49:22.061892065Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Jan 24 11:49:23.511821 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2144988823.mount: Deactivated successfully. Jan 24 11:49:28.376678 containerd[1647]: time="2026-01-24T11:49:28.376008955Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:28.376678 containerd[1647]: time="2026-01-24T11:49:28.377338075Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20808236" Jan 24 11:49:28.380806 containerd[1647]: time="2026-01-24T11:49:28.380719483Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:28.391659 containerd[1647]: time="2026-01-24T11:49:28.391145409Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:28.393761 containerd[1647]: time="2026-01-24T11:49:28.393659137Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 6.330654075s" Jan 24 11:49:28.393761 containerd[1647]: time="2026-01-24T11:49:28.393726262Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Jan 24 11:49:28.401761 containerd[1647]: time="2026-01-24T11:49:28.395798886Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Jan 24 11:49:29.069795 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1708293456.mount: Deactivated successfully. Jan 24 11:49:29.088413 containerd[1647]: time="2026-01-24T11:49:29.087074514Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 24 11:49:29.091534 containerd[1647]: time="2026-01-24T11:49:29.091331862Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Jan 24 11:49:29.097558 containerd[1647]: time="2026-01-24T11:49:29.097093581Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 24 11:49:29.104522 containerd[1647]: time="2026-01-24T11:49:29.102562503Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 24 11:49:29.104522 containerd[1647]: time="2026-01-24T11:49:29.104168931Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 708.335452ms" Jan 24 11:49:29.104522 containerd[1647]: time="2026-01-24T11:49:29.104245614Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Jan 24 11:49:29.108222 containerd[1647]: time="2026-01-24T11:49:29.107904059Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Jan 24 11:49:30.576760 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Jan 24 11:49:30.585150 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:49:31.236000 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1283283197.mount: Deactivated successfully. Jan 24 11:49:31.968987 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:49:31.967000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:49:31.987462 kernel: audit: type=1130 audit(1769255371.967:280): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:49:32.084039 (kubelet)[2319]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:49:32.319632 kubelet[2319]: E0124 11:49:32.318070 2319 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:49:32.332127 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:49:32.335971 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:49:32.335000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:49:32.347542 kernel: audit: type=1131 audit(1769255372.335:281): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:49:32.336870 systemd[1]: kubelet.service: Consumed 695ms CPU time, 110.6M memory peak. Jan 24 11:49:42.577220 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Jan 24 11:49:42.583619 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:49:43.653223 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:49:43.652000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:49:43.667764 kernel: audit: type=1130 audit(1769255383.652:282): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:49:43.690096 (kubelet)[2375]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 24 11:49:43.922834 kubelet[2375]: E0124 11:49:43.920069 2375 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 24 11:49:43.927180 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 24 11:49:43.927611 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 24 11:49:43.928634 systemd[1]: kubelet.service: Consumed 756ms CPU time, 110.6M memory peak. Jan 24 11:49:43.927000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:49:43.938543 kernel: audit: type=1131 audit(1769255383.927:283): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:49:45.348872 containerd[1647]: time="2026-01-24T11:49:45.347918165Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:45.433050 containerd[1647]: time="2026-01-24T11:49:45.375555018Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58785593" Jan 24 11:49:45.533984 containerd[1647]: time="2026-01-24T11:49:45.526732878Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:46.088895 containerd[1647]: time="2026-01-24T11:49:46.086760401Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:49:46.095066 containerd[1647]: time="2026-01-24T11:49:46.094848725Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 16.986887951s" Jan 24 11:49:46.095066 containerd[1647]: time="2026-01-24T11:49:46.095018542Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Jan 24 11:49:52.893176 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:49:52.893626 systemd[1]: kubelet.service: Consumed 756ms CPU time, 110.6M memory peak. Jan 24 11:49:52.891000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:49:52.902474 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:49:52.921366 kernel: audit: type=1130 audit(1769255392.891:284): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:49:52.922499 kernel: audit: type=1131 audit(1769255392.891:285): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:49:52.891000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:49:52.990704 systemd[1]: Reload requested from client PID 2418 ('systemctl') (unit session-8.scope)... Jan 24 11:49:52.990752 systemd[1]: Reloading... Jan 24 11:49:53.186106 zram_generator::config[2464]: No configuration found. Jan 24 11:49:53.641136 systemd[1]: Reloading finished in 649 ms. Jan 24 11:49:53.671000 audit: BPF prog-id=61 op=LOAD Jan 24 11:49:53.671000 audit: BPF prog-id=56 op=UNLOAD Jan 24 11:49:53.683819 kernel: audit: type=1334 audit(1769255393.671:286): prog-id=61 op=LOAD Jan 24 11:49:53.684002 kernel: audit: type=1334 audit(1769255393.671:287): prog-id=56 op=UNLOAD Jan 24 11:49:53.684055 kernel: audit: type=1334 audit(1769255393.673:288): prog-id=62 op=LOAD Jan 24 11:49:53.673000 audit: BPF prog-id=62 op=LOAD Jan 24 11:49:53.690759 kernel: audit: type=1334 audit(1769255393.673:289): prog-id=47 op=UNLOAD Jan 24 11:49:53.673000 audit: BPF prog-id=47 op=UNLOAD Jan 24 11:49:53.673000 audit: BPF prog-id=63 op=LOAD Jan 24 11:49:53.700236 kernel: audit: type=1334 audit(1769255393.673:290): prog-id=63 op=LOAD Jan 24 11:49:53.673000 audit: BPF prog-id=64 op=LOAD Jan 24 11:49:53.702550 kernel: audit: type=1334 audit(1769255393.673:291): prog-id=64 op=LOAD Jan 24 11:49:53.673000 audit: BPF prog-id=48 op=UNLOAD Jan 24 11:49:53.709510 kernel: audit: type=1334 audit(1769255393.673:292): prog-id=48 op=UNLOAD Jan 24 11:49:53.709577 kernel: audit: type=1334 audit(1769255393.673:293): prog-id=49 op=UNLOAD Jan 24 11:49:53.673000 audit: BPF prog-id=49 op=UNLOAD Jan 24 11:49:53.674000 audit: BPF prog-id=65 op=LOAD Jan 24 11:49:53.674000 audit: BPF prog-id=51 op=UNLOAD Jan 24 11:49:53.674000 audit: BPF prog-id=66 op=LOAD Jan 24 11:49:53.674000 audit: BPF prog-id=67 op=LOAD Jan 24 11:49:53.674000 audit: BPF prog-id=52 op=UNLOAD Jan 24 11:49:53.674000 audit: BPF prog-id=53 op=UNLOAD Jan 24 11:49:53.680000 audit: BPF prog-id=68 op=LOAD Jan 24 11:49:53.680000 audit: BPF prog-id=57 op=UNLOAD Jan 24 11:49:53.727000 audit: BPF prog-id=69 op=LOAD Jan 24 11:49:53.727000 audit: BPF prog-id=44 op=UNLOAD Jan 24 11:49:53.727000 audit: BPF prog-id=70 op=LOAD Jan 24 11:49:53.727000 audit: BPF prog-id=71 op=LOAD Jan 24 11:49:53.727000 audit: BPF prog-id=45 op=UNLOAD Jan 24 11:49:53.727000 audit: BPF prog-id=46 op=UNLOAD Jan 24 11:49:53.730000 audit: BPF prog-id=72 op=LOAD Jan 24 11:49:53.730000 audit: BPF prog-id=50 op=UNLOAD Jan 24 11:49:53.731000 audit: BPF prog-id=73 op=LOAD Jan 24 11:49:53.731000 audit: BPF prog-id=74 op=LOAD Jan 24 11:49:53.731000 audit: BPF prog-id=54 op=UNLOAD Jan 24 11:49:53.731000 audit: BPF prog-id=55 op=UNLOAD Jan 24 11:49:53.741000 audit: BPF prog-id=75 op=LOAD Jan 24 11:49:53.741000 audit: BPF prog-id=58 op=UNLOAD Jan 24 11:49:53.741000 audit: BPF prog-id=76 op=LOAD Jan 24 11:49:53.743000 audit: BPF prog-id=77 op=LOAD Jan 24 11:49:53.743000 audit: BPF prog-id=59 op=UNLOAD Jan 24 11:49:53.743000 audit: BPF prog-id=60 op=UNLOAD Jan 24 11:49:53.747000 audit: BPF prog-id=78 op=LOAD Jan 24 11:49:53.747000 audit: BPF prog-id=41 op=UNLOAD Jan 24 11:49:53.747000 audit: BPF prog-id=79 op=LOAD Jan 24 11:49:53.748000 audit: BPF prog-id=80 op=LOAD Jan 24 11:49:53.748000 audit: BPF prog-id=42 op=UNLOAD Jan 24 11:49:53.748000 audit: BPF prog-id=43 op=UNLOAD Jan 24 11:49:53.796744 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jan 24 11:49:53.797000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 24 11:49:53.798518 systemd[1]: kubelet.service: Failed with result 'signal'. Jan 24 11:49:53.799095 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:49:53.799202 systemd[1]: kubelet.service: Consumed 238ms CPU time, 98.5M memory peak. Jan 24 11:49:53.809007 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:49:54.194245 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:49:54.193000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:49:54.215627 (kubelet)[2512]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 24 11:49:54.329568 kubelet[2512]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 24 11:49:54.329568 kubelet[2512]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jan 24 11:49:54.329568 kubelet[2512]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 24 11:49:54.329568 kubelet[2512]: I0124 11:49:54.329341 2512 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 24 11:49:55.285055 kubelet[2512]: I0124 11:49:55.283664 2512 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Jan 24 11:49:55.285055 kubelet[2512]: I0124 11:49:55.283731 2512 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 24 11:49:55.285055 kubelet[2512]: I0124 11:49:55.284075 2512 server.go:956] "Client rotation is on, will bootstrap in background" Jan 24 11:49:55.465779 kubelet[2512]: E0124 11:49:55.464972 2512 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.67:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.67:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jan 24 11:49:55.471182 kubelet[2512]: I0124 11:49:55.468660 2512 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 24 11:49:55.500512 kubelet[2512]: I0124 11:49:55.498714 2512 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 24 11:49:55.517856 kubelet[2512]: I0124 11:49:55.514662 2512 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 24 11:49:55.518044 kubelet[2512]: I0124 11:49:55.517975 2512 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 24 11:49:55.518551 kubelet[2512]: I0124 11:49:55.518012 2512 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 24 11:49:55.518743 kubelet[2512]: I0124 11:49:55.518594 2512 topology_manager.go:138] "Creating topology manager with none policy" Jan 24 11:49:55.518743 kubelet[2512]: I0124 11:49:55.518615 2512 container_manager_linux.go:303] "Creating device plugin manager" Jan 24 11:49:55.520166 kubelet[2512]: I0124 11:49:55.519856 2512 state_mem.go:36] "Initialized new in-memory state store" Jan 24 11:49:55.528998 kubelet[2512]: I0124 11:49:55.528871 2512 kubelet.go:480] "Attempting to sync node with API server" Jan 24 11:49:55.528998 kubelet[2512]: I0124 11:49:55.528933 2512 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 24 11:49:55.530619 kubelet[2512]: I0124 11:49:55.530093 2512 kubelet.go:386] "Adding apiserver pod source" Jan 24 11:49:55.537372 kubelet[2512]: I0124 11:49:55.536346 2512 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 24 11:49:55.555138 kubelet[2512]: I0124 11:49:55.555049 2512 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.1.5" apiVersion="v1" Jan 24 11:49:55.561525 kubelet[2512]: I0124 11:49:55.557655 2512 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Jan 24 11:49:55.561525 kubelet[2512]: E0124 11:49:55.561011 2512 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.67:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.67:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 24 11:49:55.561525 kubelet[2512]: E0124 11:49:55.561056 2512 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.67:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.67:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 24 11:49:55.563784 kubelet[2512]: W0124 11:49:55.562892 2512 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jan 24 11:49:55.572893 kubelet[2512]: I0124 11:49:55.572822 2512 watchdog_linux.go:99] "Systemd watchdog is not enabled" Jan 24 11:49:55.572982 kubelet[2512]: I0124 11:49:55.572969 2512 server.go:1289] "Started kubelet" Jan 24 11:49:55.574110 kubelet[2512]: I0124 11:49:55.573535 2512 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Jan 24 11:49:55.582508 kubelet[2512]: I0124 11:49:55.582183 2512 server.go:317] "Adding debug handlers to kubelet server" Jan 24 11:49:55.584402 kubelet[2512]: I0124 11:49:55.584314 2512 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 24 11:49:55.588551 kubelet[2512]: I0124 11:49:55.587068 2512 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 24 11:49:55.598947 kubelet[2512]: I0124 11:49:55.598897 2512 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 24 11:49:55.616548 kubelet[2512]: I0124 11:49:55.616174 2512 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 24 11:49:55.617781 kubelet[2512]: E0124 11:49:55.615955 2512 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.67:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.67:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.188da8688a107b59 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-01-24 11:49:55.572874073 +0000 UTC m=+1.341655382,LastTimestamp:2026-01-24 11:49:55.572874073 +0000 UTC m=+1.341655382,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jan 24 11:49:55.624985 kubelet[2512]: I0124 11:49:55.624959 2512 volume_manager.go:297] "Starting Kubelet Volume Manager" Jan 24 11:49:55.625597 kubelet[2512]: I0124 11:49:55.625574 2512 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Jan 24 11:49:55.625923 kubelet[2512]: I0124 11:49:55.625904 2512 reconciler.go:26] "Reconciler: start to sync state" Jan 24 11:49:55.627509 kubelet[2512]: I0124 11:49:55.627242 2512 factory.go:223] Registration of the systemd container factory successfully Jan 24 11:49:55.627724 kubelet[2512]: I0124 11:49:55.627691 2512 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 24 11:49:55.629501 kubelet[2512]: E0124 11:49:55.627764 2512 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.67:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.67:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 24 11:49:55.630211 kubelet[2512]: E0124 11:49:55.629718 2512 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:49:55.630524 kubelet[2512]: E0124 11:49:55.630210 2512 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 24 11:49:55.629000 audit[2529]: NETFILTER_CFG table=mangle:42 family=2 entries=2 op=nft_register_chain pid=2529 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:49:55.629000 audit[2529]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffd8e00dca0 a2=0 a3=0 items=0 ppid=2512 pid=2529 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:55.629000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Jan 24 11:49:55.634708 kubelet[2512]: E0124 11:49:55.634643 2512 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.67:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.67:6443: connect: connection refused" interval="200ms" Jan 24 11:49:55.636000 audit[2532]: NETFILTER_CFG table=filter:43 family=2 entries=1 op=nft_register_chain pid=2532 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:49:55.636000 audit[2532]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc2b3f1d60 a2=0 a3=0 items=0 ppid=2512 pid=2532 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:55.636000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 Jan 24 11:49:55.641247 kubelet[2512]: I0124 11:49:55.640846 2512 factory.go:223] Registration of the containerd container factory successfully Jan 24 11:49:55.649000 audit[2534]: NETFILTER_CFG table=filter:44 family=2 entries=2 op=nft_register_chain pid=2534 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:49:55.649000 audit[2534]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffe1f0bf9d0 a2=0 a3=0 items=0 ppid=2512 pid=2534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:55.649000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 24 11:49:55.661000 audit[2537]: NETFILTER_CFG table=filter:45 family=2 entries=2 op=nft_register_chain pid=2537 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:49:55.661000 audit[2537]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffda62d8f40 a2=0 a3=0 items=0 ppid=2512 pid=2537 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:55.661000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 24 11:49:55.677480 kubelet[2512]: I0124 11:49:55.677099 2512 cpu_manager.go:221] "Starting CPU manager" policy="none" Jan 24 11:49:55.677480 kubelet[2512]: I0124 11:49:55.677225 2512 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jan 24 11:49:55.677480 kubelet[2512]: I0124 11:49:55.677369 2512 state_mem.go:36] "Initialized new in-memory state store" Jan 24 11:49:55.686856 kubelet[2512]: I0124 11:49:55.686823 2512 policy_none.go:49] "None policy: Start" Jan 24 11:49:55.686968 kubelet[2512]: I0124 11:49:55.686884 2512 memory_manager.go:186] "Starting memorymanager" policy="None" Jan 24 11:49:55.686968 kubelet[2512]: I0124 11:49:55.686932 2512 state_mem.go:35] "Initializing new in-memory state store" Jan 24 11:49:55.694000 audit[2542]: NETFILTER_CFG table=filter:46 family=2 entries=1 op=nft_register_rule pid=2542 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:49:55.694000 audit[2542]: SYSCALL arch=c000003e syscall=46 success=yes exit=924 a0=3 a1=7fff8849b900 a2=0 a3=0 items=0 ppid=2512 pid=2542 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:55.694000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D41004B5542452D4649524557414C4C002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E7400626C6F636B20696E636F6D696E67206C6F63616C6E657420636F6E6E656374696F6E73002D2D647374003132372E302E302E302F38 Jan 24 11:49:55.697972 kubelet[2512]: I0124 11:49:55.696705 2512 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Jan 24 11:49:55.700000 audit[2544]: NETFILTER_CFG table=mangle:47 family=10 entries=2 op=nft_register_chain pid=2544 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:49:55.700000 audit[2544]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffc1d5798a0 a2=0 a3=0 items=0 ppid=2512 pid=2544 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:55.700000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Jan 24 11:49:55.701000 audit[2545]: NETFILTER_CFG table=mangle:48 family=2 entries=1 op=nft_register_chain pid=2545 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:49:55.701000 audit[2545]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffcedff9bb0 a2=0 a3=0 items=0 ppid=2512 pid=2545 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:55.701000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Jan 24 11:49:55.704863 kubelet[2512]: I0124 11:49:55.703991 2512 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Jan 24 11:49:55.704863 kubelet[2512]: I0124 11:49:55.704043 2512 status_manager.go:230] "Starting to sync pod status with apiserver" Jan 24 11:49:55.704863 kubelet[2512]: I0124 11:49:55.704071 2512 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jan 24 11:49:55.704863 kubelet[2512]: I0124 11:49:55.704081 2512 kubelet.go:2436] "Starting kubelet main sync loop" Jan 24 11:49:55.704863 kubelet[2512]: E0124 11:49:55.704135 2512 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 24 11:49:55.703000 audit[2546]: NETFILTER_CFG table=nat:49 family=2 entries=1 op=nft_register_chain pid=2546 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:49:55.703000 audit[2546]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc1df2fb40 a2=0 a3=0 items=0 ppid=2512 pid=2546 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:55.703000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Jan 24 11:49:55.708514 kubelet[2512]: E0124 11:49:55.708098 2512 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.67:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.67:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 24 11:49:55.708000 audit[2548]: NETFILTER_CFG table=filter:50 family=2 entries=1 op=nft_register_chain pid=2548 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:49:55.708000 audit[2548]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe78ec9d70 a2=0 a3=0 items=0 ppid=2512 pid=2548 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:55.708000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Jan 24 11:49:55.712000 audit[2547]: NETFILTER_CFG table=mangle:51 family=10 entries=1 op=nft_register_chain pid=2547 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:49:55.712000 audit[2547]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffdffb26ab0 a2=0 a3=0 items=0 ppid=2512 pid=2547 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:55.712000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Jan 24 11:49:55.714000 audit[2549]: NETFILTER_CFG table=nat:52 family=10 entries=1 op=nft_register_chain pid=2549 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:49:55.714000 audit[2549]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe26d80480 a2=0 a3=0 items=0 ppid=2512 pid=2549 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:55.714000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Jan 24 11:49:55.717000 audit[2550]: NETFILTER_CFG table=filter:53 family=10 entries=1 op=nft_register_chain pid=2550 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:49:55.717000 audit[2550]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe333ce0b0 a2=0 a3=0 items=0 ppid=2512 pid=2550 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:55.717000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Jan 24 11:49:55.731570 kubelet[2512]: E0124 11:49:55.730455 2512 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:49:55.736634 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jan 24 11:49:55.792852 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jan 24 11:49:55.801378 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jan 24 11:49:55.804725 kubelet[2512]: E0124 11:49:55.804542 2512 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jan 24 11:49:55.831614 kubelet[2512]: E0124 11:49:55.831570 2512 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 24 11:49:55.833000 kubelet[2512]: E0124 11:49:55.832019 2512 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Jan 24 11:49:55.833525 kubelet[2512]: I0124 11:49:55.833389 2512 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 24 11:49:55.833568 kubelet[2512]: I0124 11:49:55.833504 2512 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 24 11:49:55.834235 kubelet[2512]: I0124 11:49:55.834136 2512 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 24 11:49:55.835346 kubelet[2512]: E0124 11:49:55.835176 2512 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.67:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.67:6443: connect: connection refused" interval="400ms" Jan 24 11:49:55.839513 kubelet[2512]: E0124 11:49:55.839247 2512 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jan 24 11:49:55.839513 kubelet[2512]: E0124 11:49:55.839385 2512 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jan 24 11:49:56.109410 kubelet[2512]: I0124 11:49:56.108165 2512 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6e6cfcfb327385445a9bb0d2bc2fd5d4-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6e6cfcfb327385445a9bb0d2bc2fd5d4\") " pod="kube-system/kube-scheduler-localhost" Jan 24 11:49:56.121727 kubelet[2512]: I0124 11:49:56.120365 2512 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:49:56.121727 kubelet[2512]: E0124 11:49:56.120754 2512 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.67:6443/api/v1/nodes\": dial tcp 10.0.0.67:6443: connect: connection refused" node="localhost" Jan 24 11:49:56.172995 systemd[1]: Created slice kubepods-burstable-pod6e6cfcfb327385445a9bb0d2bc2fd5d4.slice - libcontainer container kubepods-burstable-pod6e6cfcfb327385445a9bb0d2bc2fd5d4.slice. Jan 24 11:49:56.206087 kubelet[2512]: E0124 11:49:56.204850 2512 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:49:56.225742 systemd[1]: Created slice kubepods-burstable-podb2417606a328cd170dda8398a7770fe4.slice - libcontainer container kubepods-burstable-podb2417606a328cd170dda8398a7770fe4.slice. Jan 24 11:49:56.236506 kubelet[2512]: E0124 11:49:56.236220 2512 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.67:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.67:6443: connect: connection refused" interval="800ms" Jan 24 11:49:56.248564 kubelet[2512]: E0124 11:49:56.246876 2512 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:49:56.260765 systemd[1]: Created slice kubepods-burstable-pod66e26b992bcd7ea6fb75e339cf7a3f7d.slice - libcontainer container kubepods-burstable-pod66e26b992bcd7ea6fb75e339cf7a3f7d.slice. Jan 24 11:49:56.269850 kubelet[2512]: E0124 11:49:56.267771 2512 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:49:56.325726 kubelet[2512]: I0124 11:49:56.324686 2512 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:49:56.325726 kubelet[2512]: I0124 11:49:56.324742 2512 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b2417606a328cd170dda8398a7770fe4-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"b2417606a328cd170dda8398a7770fe4\") " pod="kube-system/kube-apiserver-localhost" Jan 24 11:49:56.325726 kubelet[2512]: I0124 11:49:56.324771 2512 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b2417606a328cd170dda8398a7770fe4-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"b2417606a328cd170dda8398a7770fe4\") " pod="kube-system/kube-apiserver-localhost" Jan 24 11:49:56.325726 kubelet[2512]: I0124 11:49:56.324803 2512 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:49:56.325726 kubelet[2512]: I0124 11:49:56.324827 2512 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:49:56.326079 kubelet[2512]: I0124 11:49:56.324852 2512 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b2417606a328cd170dda8398a7770fe4-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"b2417606a328cd170dda8398a7770fe4\") " pod="kube-system/kube-apiserver-localhost" Jan 24 11:49:56.326079 kubelet[2512]: I0124 11:49:56.324870 2512 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:49:56.326079 kubelet[2512]: I0124 11:49:56.324893 2512 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:49:56.333595 kubelet[2512]: I0124 11:49:56.332715 2512 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:49:56.333595 kubelet[2512]: E0124 11:49:56.333151 2512 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.67:6443/api/v1/nodes\": dial tcp 10.0.0.67:6443: connect: connection refused" node="localhost" Jan 24 11:49:56.438085 kubelet[2512]: E0124 11:49:56.434730 2512 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.67:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.67:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 24 11:49:56.504522 kubelet[2512]: E0124 11:49:56.498887 2512 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.67:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.67:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 24 11:49:56.521885 kubelet[2512]: E0124 11:49:56.521241 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:49:56.574509 kubelet[2512]: E0124 11:49:56.573663 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:49:56.670711 kubelet[2512]: E0124 11:49:56.577606 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:49:56.671771 containerd[1647]: time="2026-01-24T11:49:56.626796554Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6e6cfcfb327385445a9bb0d2bc2fd5d4,Namespace:kube-system,Attempt:0,}" Jan 24 11:49:56.671771 containerd[1647]: time="2026-01-24T11:49:56.668208151Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:66e26b992bcd7ea6fb75e339cf7a3f7d,Namespace:kube-system,Attempt:0,}" Jan 24 11:49:56.673180 kubelet[2512]: E0124 11:49:56.671209 2512 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.67:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.67:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 24 11:49:56.716161 containerd[1647]: time="2026-01-24T11:49:56.709771744Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:b2417606a328cd170dda8398a7770fe4,Namespace:kube-system,Attempt:0,}" Jan 24 11:49:56.894504 kubelet[2512]: I0124 11:49:56.893720 2512 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:49:57.072928 kubelet[2512]: E0124 11:49:56.895723 2512 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.67:6443/api/v1/nodes\": dial tcp 10.0.0.67:6443: connect: connection refused" node="localhost" Jan 24 11:49:57.092212 kubelet[2512]: E0124 11:49:57.085924 2512 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.67:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.67:6443: connect: connection refused" interval="1.6s" Jan 24 11:49:57.188248 kubelet[2512]: E0124 11:49:57.187629 2512 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.67:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.67:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 24 11:49:57.643466 kubelet[2512]: E0124 11:49:57.642947 2512 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.0.0.67:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.67:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jan 24 11:49:58.095177 kubelet[2512]: I0124 11:49:58.093230 2512 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:49:58.095177 kubelet[2512]: E0124 11:49:58.095018 2512 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.67:6443/api/v1/nodes\": dial tcp 10.0.0.67:6443: connect: connection refused" node="localhost" Jan 24 11:49:58.184414 containerd[1647]: time="2026-01-24T11:49:58.180689398Z" level=info msg="connecting to shim 4a95108dd84d5620bbce07e08082acff25e837447724f0b2cb1ea1fc3e7ddde6" address="unix:///run/containerd/s/e778a2d245ed037f51735cce3d2b99c0d59843448210aad0488703e6956a3587" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:49:58.191982 containerd[1647]: time="2026-01-24T11:49:58.187622474Z" level=info msg="connecting to shim 4aadaee71f33bed3272dbde5ceac3626eec87be92bd1991e07c0cead4cfe13e1" address="unix:///run/containerd/s/e43e666e62d58a48a3976da58f95e68f633d4b4f6e2d2106515a017bf8f39467" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:49:58.194223 containerd[1647]: time="2026-01-24T11:49:58.194142854Z" level=info msg="connecting to shim 04a2d0581942060efb504e6b881240a2eb48d2748a75d386f525d7b4a3d239d3" address="unix:///run/containerd/s/b10f8ef1c7667849f1dc396b7ff875f13fc04875b472f1efaf696ed9b8913a98" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:49:58.497066 systemd[1]: Started cri-containerd-04a2d0581942060efb504e6b881240a2eb48d2748a75d386f525d7b4a3d239d3.scope - libcontainer container 04a2d0581942060efb504e6b881240a2eb48d2748a75d386f525d7b4a3d239d3. Jan 24 11:49:58.552972 systemd[1]: Started cri-containerd-4a95108dd84d5620bbce07e08082acff25e837447724f0b2cb1ea1fc3e7ddde6.scope - libcontainer container 4a95108dd84d5620bbce07e08082acff25e837447724f0b2cb1ea1fc3e7ddde6. Jan 24 11:49:58.578950 systemd[1]: Started cri-containerd-4aadaee71f33bed3272dbde5ceac3626eec87be92bd1991e07c0cead4cfe13e1.scope - libcontainer container 4aadaee71f33bed3272dbde5ceac3626eec87be92bd1991e07c0cead4cfe13e1. Jan 24 11:49:58.634000 audit: BPF prog-id=81 op=LOAD Jan 24 11:49:58.638710 kernel: kauditd_printk_skb: 70 callbacks suppressed Jan 24 11:49:58.639895 kernel: audit: type=1334 audit(1769255398.634:340): prog-id=81 op=LOAD Jan 24 11:49:58.638000 audit: BPF prog-id=82 op=LOAD Jan 24 11:49:58.647148 kernel: audit: type=1334 audit(1769255398.638:341): prog-id=82 op=LOAD Jan 24 11:49:58.638000 audit[2610]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a6238 a2=98 a3=0 items=0 ppid=2571 pid=2610 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.638000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461393531303864643834643536323062626365303765303830383261 Jan 24 11:49:58.679118 kernel: audit: type=1300 audit(1769255398.638:341): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a6238 a2=98 a3=0 items=0 ppid=2571 pid=2610 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.679243 kernel: audit: type=1327 audit(1769255398.638:341): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461393531303864643834643536323062626365303765303830383261 Jan 24 11:49:58.679374 kernel: audit: type=1334 audit(1769255398.638:342): prog-id=82 op=UNLOAD Jan 24 11:49:58.638000 audit: BPF prog-id=82 op=UNLOAD Jan 24 11:49:58.695196 kernel: audit: type=1300 audit(1769255398.638:342): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2571 pid=2610 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.638000 audit[2610]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2571 pid=2610 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.695618 kubelet[2512]: E0124 11:49:58.688881 2512 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.67:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.67:6443: connect: connection refused" interval="3.2s" Jan 24 11:49:58.638000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461393531303864643834643536323062626365303765303830383261 Jan 24 11:49:58.639000 audit: BPF prog-id=83 op=LOAD Jan 24 11:49:58.742454 kernel: audit: type=1327 audit(1769255398.638:342): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461393531303864643834643536323062626365303765303830383261 Jan 24 11:49:58.742534 kernel: audit: type=1334 audit(1769255398.639:343): prog-id=83 op=LOAD Jan 24 11:49:58.742569 kernel: audit: type=1334 audit(1769255398.639:344): prog-id=84 op=LOAD Jan 24 11:49:58.639000 audit: BPF prog-id=84 op=LOAD Jan 24 11:49:58.759578 kernel: audit: type=1300 audit(1769255398.639:344): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a6488 a2=98 a3=0 items=0 ppid=2571 pid=2610 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.639000 audit[2610]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a6488 a2=98 a3=0 items=0 ppid=2571 pid=2610 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.759810 kubelet[2512]: E0124 11:49:58.746562 2512 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.0.0.67:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.67:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 24 11:49:58.639000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461393531303864643834643536323062626365303765303830383261 Jan 24 11:49:58.640000 audit: BPF prog-id=85 op=LOAD Jan 24 11:49:58.640000 audit[2610]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a6218 a2=98 a3=0 items=0 ppid=2571 pid=2610 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.640000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461393531303864643834643536323062626365303765303830383261 Jan 24 11:49:58.640000 audit: BPF prog-id=85 op=UNLOAD Jan 24 11:49:58.640000 audit[2610]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2571 pid=2610 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.640000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461393531303864643834643536323062626365303765303830383261 Jan 24 11:49:58.640000 audit: BPF prog-id=84 op=UNLOAD Jan 24 11:49:58.640000 audit[2610]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2571 pid=2610 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.640000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461393531303864643834643536323062626365303765303830383261 Jan 24 11:49:58.640000 audit: BPF prog-id=86 op=LOAD Jan 24 11:49:58.640000 audit[2614]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001b0238 a2=98 a3=0 items=0 ppid=2580 pid=2614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.640000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034613264303538313934323036306566623530346536623838313234 Jan 24 11:49:58.641000 audit: BPF prog-id=86 op=UNLOAD Jan 24 11:49:58.641000 audit[2614]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2580 pid=2614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.641000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034613264303538313934323036306566623530346536623838313234 Jan 24 11:49:58.642000 audit: BPF prog-id=87 op=LOAD Jan 24 11:49:58.642000 audit[2614]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001b0488 a2=98 a3=0 items=0 ppid=2580 pid=2614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.642000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034613264303538313934323036306566623530346536623838313234 Jan 24 11:49:58.643000 audit: BPF prog-id=88 op=LOAD Jan 24 11:49:58.640000 audit: BPF prog-id=89 op=LOAD Jan 24 11:49:58.640000 audit[2610]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a66e8 a2=98 a3=0 items=0 ppid=2571 pid=2610 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.640000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461393531303864643834643536323062626365303765303830383261 Jan 24 11:49:58.643000 audit[2614]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001b0218 a2=98 a3=0 items=0 ppid=2580 pid=2614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.643000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034613264303538313934323036306566623530346536623838313234 Jan 24 11:49:58.644000 audit: BPF prog-id=88 op=UNLOAD Jan 24 11:49:58.644000 audit[2614]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2580 pid=2614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.644000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034613264303538313934323036306566623530346536623838313234 Jan 24 11:49:58.645000 audit: BPF prog-id=87 op=UNLOAD Jan 24 11:49:58.645000 audit[2614]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2580 pid=2614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.645000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034613264303538313934323036306566623530346536623838313234 Jan 24 11:49:58.645000 audit: BPF prog-id=90 op=LOAD Jan 24 11:49:58.645000 audit[2614]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001b06e8 a2=98 a3=0 items=0 ppid=2580 pid=2614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.645000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3034613264303538313934323036306566623530346536623838313234 Jan 24 11:49:58.649000 audit: BPF prog-id=91 op=LOAD Jan 24 11:49:58.650000 audit: BPF prog-id=92 op=LOAD Jan 24 11:49:58.650000 audit[2608]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0000fc238 a2=98 a3=0 items=0 ppid=2576 pid=2608 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.650000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461616461656537316633336265643332373264626465356365616333 Jan 24 11:49:58.651000 audit: BPF prog-id=92 op=UNLOAD Jan 24 11:49:58.651000 audit[2608]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2576 pid=2608 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.651000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461616461656537316633336265643332373264626465356365616333 Jan 24 11:49:58.651000 audit: BPF prog-id=93 op=LOAD Jan 24 11:49:58.651000 audit[2608]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0000fc488 a2=98 a3=0 items=0 ppid=2576 pid=2608 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.651000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461616461656537316633336265643332373264626465356365616333 Jan 24 11:49:58.758000 audit: BPF prog-id=94 op=LOAD Jan 24 11:49:58.758000 audit[2608]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0000fc218 a2=98 a3=0 items=0 ppid=2576 pid=2608 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.758000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461616461656537316633336265643332373264626465356365616333 Jan 24 11:49:58.759000 audit: BPF prog-id=94 op=UNLOAD Jan 24 11:49:58.759000 audit[2608]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2576 pid=2608 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.759000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461616461656537316633336265643332373264626465356365616333 Jan 24 11:49:58.759000 audit: BPF prog-id=93 op=UNLOAD Jan 24 11:49:58.759000 audit[2608]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2576 pid=2608 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.759000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461616461656537316633336265643332373264626465356365616333 Jan 24 11:49:58.759000 audit: BPF prog-id=95 op=LOAD Jan 24 11:49:58.759000 audit[2608]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0000fc6e8 a2=98 a3=0 items=0 ppid=2576 pid=2608 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:58.759000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3461616461656537316633336265643332373264626465356365616333 Jan 24 11:49:58.774678 kubelet[2512]: E0124 11:49:58.774577 2512 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.0.0.67:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.67:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 24 11:49:58.871644 containerd[1647]: time="2026-01-24T11:49:58.871511320Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:66e26b992bcd7ea6fb75e339cf7a3f7d,Namespace:kube-system,Attempt:0,} returns sandbox id \"4a95108dd84d5620bbce07e08082acff25e837447724f0b2cb1ea1fc3e7ddde6\"" Jan 24 11:49:58.873560 kubelet[2512]: E0124 11:49:58.873420 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:49:58.884572 containerd[1647]: time="2026-01-24T11:49:58.882131836Z" level=info msg="CreateContainer within sandbox \"4a95108dd84d5620bbce07e08082acff25e837447724f0b2cb1ea1fc3e7ddde6\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jan 24 11:49:58.898803 containerd[1647]: time="2026-01-24T11:49:58.898743052Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:b2417606a328cd170dda8398a7770fe4,Namespace:kube-system,Attempt:0,} returns sandbox id \"4aadaee71f33bed3272dbde5ceac3626eec87be92bd1991e07c0cead4cfe13e1\"" Jan 24 11:49:58.921954 kubelet[2512]: E0124 11:49:58.921790 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:49:58.935671 containerd[1647]: time="2026-01-24T11:49:58.935505717Z" level=info msg="CreateContainer within sandbox \"4aadaee71f33bed3272dbde5ceac3626eec87be92bd1991e07c0cead4cfe13e1\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jan 24 11:49:58.938713 containerd[1647]: time="2026-01-24T11:49:58.938617583Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6e6cfcfb327385445a9bb0d2bc2fd5d4,Namespace:kube-system,Attempt:0,} returns sandbox id \"04a2d0581942060efb504e6b881240a2eb48d2748a75d386f525d7b4a3d239d3\"" Jan 24 11:49:58.940365 kubelet[2512]: E0124 11:49:58.940211 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:49:58.954728 containerd[1647]: time="2026-01-24T11:49:58.954627779Z" level=info msg="CreateContainer within sandbox \"04a2d0581942060efb504e6b881240a2eb48d2748a75d386f525d7b4a3d239d3\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jan 24 11:49:58.960824 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1002481117.mount: Deactivated successfully. Jan 24 11:49:58.964014 containerd[1647]: time="2026-01-24T11:49:58.963821429Z" level=info msg="Container b0c4c98e002c741713e61e61888663dbd775dc4bae23fe7d0b86e98503190f63: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:49:58.995051 containerd[1647]: time="2026-01-24T11:49:58.993859427Z" level=info msg="CreateContainer within sandbox \"4a95108dd84d5620bbce07e08082acff25e837447724f0b2cb1ea1fc3e7ddde6\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"b0c4c98e002c741713e61e61888663dbd775dc4bae23fe7d0b86e98503190f63\"" Jan 24 11:49:59.002658 containerd[1647]: time="2026-01-24T11:49:58.999232200Z" level=info msg="StartContainer for \"b0c4c98e002c741713e61e61888663dbd775dc4bae23fe7d0b86e98503190f63\"" Jan 24 11:49:59.010207 containerd[1647]: time="2026-01-24T11:49:59.005714797Z" level=info msg="connecting to shim b0c4c98e002c741713e61e61888663dbd775dc4bae23fe7d0b86e98503190f63" address="unix:///run/containerd/s/e778a2d245ed037f51735cce3d2b99c0d59843448210aad0488703e6956a3587" protocol=ttrpc version=3 Jan 24 11:49:59.017987 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1182080381.mount: Deactivated successfully. Jan 24 11:49:59.031895 containerd[1647]: time="2026-01-24T11:49:59.031461769Z" level=info msg="Container 3925ca274f194ad512a9b7ef8bcd146731e5bb29e3b62574d3ae0af8b50f3c06: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:49:59.039649 containerd[1647]: time="2026-01-24T11:49:59.038495003Z" level=info msg="Container 51df99e807c575426a0b1b0edb58bf4d25dbd42f2bb9cd6e380438339a94e0ef: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:49:59.059963 containerd[1647]: time="2026-01-24T11:49:59.059880486Z" level=info msg="CreateContainer within sandbox \"4aadaee71f33bed3272dbde5ceac3626eec87be92bd1991e07c0cead4cfe13e1\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"3925ca274f194ad512a9b7ef8bcd146731e5bb29e3b62574d3ae0af8b50f3c06\"" Jan 24 11:49:59.061170 containerd[1647]: time="2026-01-24T11:49:59.061142641Z" level=info msg="StartContainer for \"3925ca274f194ad512a9b7ef8bcd146731e5bb29e3b62574d3ae0af8b50f3c06\"" Jan 24 11:49:59.075409 containerd[1647]: time="2026-01-24T11:49:59.075179775Z" level=info msg="CreateContainer within sandbox \"04a2d0581942060efb504e6b881240a2eb48d2748a75d386f525d7b4a3d239d3\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"51df99e807c575426a0b1b0edb58bf4d25dbd42f2bb9cd6e380438339a94e0ef\"" Jan 24 11:49:59.078329 containerd[1647]: time="2026-01-24T11:49:59.077183582Z" level=info msg="connecting to shim 3925ca274f194ad512a9b7ef8bcd146731e5bb29e3b62574d3ae0af8b50f3c06" address="unix:///run/containerd/s/e43e666e62d58a48a3976da58f95e68f633d4b4f6e2d2106515a017bf8f39467" protocol=ttrpc version=3 Jan 24 11:49:59.080883 containerd[1647]: time="2026-01-24T11:49:59.078482566Z" level=info msg="StartContainer for \"51df99e807c575426a0b1b0edb58bf4d25dbd42f2bb9cd6e380438339a94e0ef\"" Jan 24 11:49:59.080883 containerd[1647]: time="2026-01-24T11:49:59.079802699Z" level=info msg="connecting to shim 51df99e807c575426a0b1b0edb58bf4d25dbd42f2bb9cd6e380438339a94e0ef" address="unix:///run/containerd/s/b10f8ef1c7667849f1dc396b7ff875f13fc04875b472f1efaf696ed9b8913a98" protocol=ttrpc version=3 Jan 24 11:49:59.080363 systemd[1]: Started cri-containerd-b0c4c98e002c741713e61e61888663dbd775dc4bae23fe7d0b86e98503190f63.scope - libcontainer container b0c4c98e002c741713e61e61888663dbd775dc4bae23fe7d0b86e98503190f63. Jan 24 11:49:59.248193 systemd[1]: Started cri-containerd-51df99e807c575426a0b1b0edb58bf4d25dbd42f2bb9cd6e380438339a94e0ef.scope - libcontainer container 51df99e807c575426a0b1b0edb58bf4d25dbd42f2bb9cd6e380438339a94e0ef. Jan 24 11:49:59.272000 audit: BPF prog-id=96 op=LOAD Jan 24 11:49:59.273000 audit: BPF prog-id=97 op=LOAD Jan 24 11:49:59.273000 audit[2688]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=2571 pid=2688 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:59.273000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6230633463393865303032633734313731336536316536313838383636 Jan 24 11:49:59.274000 audit: BPF prog-id=97 op=UNLOAD Jan 24 11:49:59.274000 audit[2688]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2571 pid=2688 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:59.274000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6230633463393865303032633734313731336536316536313838383636 Jan 24 11:49:59.274000 audit: BPF prog-id=98 op=LOAD Jan 24 11:49:59.274000 audit[2688]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=2571 pid=2688 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:59.274000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6230633463393865303032633734313731336536316536313838383636 Jan 24 11:49:59.275000 audit: BPF prog-id=99 op=LOAD Jan 24 11:49:59.275000 audit[2688]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=2571 pid=2688 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:59.275000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6230633463393865303032633734313731336536316536313838383636 Jan 24 11:49:59.275000 audit: BPF prog-id=99 op=UNLOAD Jan 24 11:49:59.275000 audit[2688]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2571 pid=2688 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:59.275000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6230633463393865303032633734313731336536316536313838383636 Jan 24 11:49:59.275000 audit: BPF prog-id=98 op=UNLOAD Jan 24 11:49:59.275000 audit[2688]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2571 pid=2688 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:59.275000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6230633463393865303032633734313731336536316536313838383636 Jan 24 11:49:59.275000 audit: BPF prog-id=100 op=LOAD Jan 24 11:49:59.275000 audit[2688]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=2571 pid=2688 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:59.275000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6230633463393865303032633734313731336536316536313838383636 Jan 24 11:49:59.307000 audit: BPF prog-id=101 op=LOAD Jan 24 11:49:59.308000 audit: BPF prog-id=102 op=LOAD Jan 24 11:49:59.308000 audit[2700]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=2580 pid=2700 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:59.308000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3531646639396538303763353735343236613062316230656462353862 Jan 24 11:49:59.308000 audit: BPF prog-id=102 op=UNLOAD Jan 24 11:49:59.308000 audit[2700]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2580 pid=2700 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:59.308000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3531646639396538303763353735343236613062316230656462353862 Jan 24 11:49:59.310000 audit: BPF prog-id=103 op=LOAD Jan 24 11:49:59.310000 audit[2700]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=2580 pid=2700 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:59.310000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3531646639396538303763353735343236613062316230656462353862 Jan 24 11:49:59.310000 audit: BPF prog-id=104 op=LOAD Jan 24 11:49:59.310000 audit[2700]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=2580 pid=2700 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:59.310000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3531646639396538303763353735343236613062316230656462353862 Jan 24 11:49:59.310000 audit: BPF prog-id=104 op=UNLOAD Jan 24 11:49:59.310000 audit[2700]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2580 pid=2700 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:59.310000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3531646639396538303763353735343236613062316230656462353862 Jan 24 11:49:59.310000 audit: BPF prog-id=103 op=UNLOAD Jan 24 11:49:59.310000 audit[2700]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2580 pid=2700 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:59.310000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3531646639396538303763353735343236613062316230656462353862 Jan 24 11:49:59.310000 audit: BPF prog-id=105 op=LOAD Jan 24 11:49:59.310000 audit[2700]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=2580 pid=2700 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:49:59.310000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3531646639396538303763353735343236613062316230656462353862 Jan 24 11:49:59.313049 kubelet[2512]: E0124 11:49:59.311418 2512 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.0.0.67:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.67:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 24 11:49:59.350640 systemd[1]: Started cri-containerd-3925ca274f194ad512a9b7ef8bcd146731e5bb29e3b62574d3ae0af8b50f3c06.scope - libcontainer container 3925ca274f194ad512a9b7ef8bcd146731e5bb29e3b62574d3ae0af8b50f3c06. Jan 24 11:50:00.932709 kubelet[2512]: E0124 11:50:00.930722 2512 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.67:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.67:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.188da8688a107b59 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-01-24 11:49:55.572874073 +0000 UTC m=+1.341655382,LastTimestamp:2026-01-24 11:49:55.572874073 +0000 UTC m=+1.341655382,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jan 24 11:50:00.944015 kubelet[2512]: E0124 11:50:00.943926 2512 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.0.0.67:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.67:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 24 11:50:00.954546 kubelet[2512]: I0124 11:50:00.953774 2512 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:50:00.954546 kubelet[2512]: E0124 11:50:00.954397 2512 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.0.0.67:6443/api/v1/nodes\": dial tcp 10.0.0.67:6443: connect: connection refused" node="localhost" Jan 24 11:50:00.962000 audit: BPF prog-id=106 op=LOAD Jan 24 11:50:00.964000 audit: BPF prog-id=107 op=LOAD Jan 24 11:50:00.964000 audit[2701]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=2576 pid=2701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:00.964000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3339323563613237346631393461643531326139623765663862636431 Jan 24 11:50:00.964000 audit: BPF prog-id=107 op=UNLOAD Jan 24 11:50:00.964000 audit[2701]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2576 pid=2701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:00.964000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3339323563613237346631393461643531326139623765663862636431 Jan 24 11:50:00.965000 audit: BPF prog-id=108 op=LOAD Jan 24 11:50:00.965000 audit[2701]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=2576 pid=2701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:00.965000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3339323563613237346631393461643531326139623765663862636431 Jan 24 11:50:00.966000 audit: BPF prog-id=109 op=LOAD Jan 24 11:50:00.966000 audit[2701]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=2576 pid=2701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:00.966000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3339323563613237346631393461643531326139623765663862636431 Jan 24 11:50:00.966000 audit: BPF prog-id=109 op=UNLOAD Jan 24 11:50:00.966000 audit[2701]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2576 pid=2701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:00.966000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3339323563613237346631393461643531326139623765663862636431 Jan 24 11:50:00.966000 audit: BPF prog-id=108 op=UNLOAD Jan 24 11:50:00.966000 audit[2701]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2576 pid=2701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:00.966000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3339323563613237346631393461643531326139623765663862636431 Jan 24 11:50:00.967000 audit: BPF prog-id=110 op=LOAD Jan 24 11:50:00.967000 audit[2701]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=2576 pid=2701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:00.967000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3339323563613237346631393461643531326139623765663862636431 Jan 24 11:50:00.999365 containerd[1647]: time="2026-01-24T11:50:00.999161458Z" level=info msg="StartContainer for \"b0c4c98e002c741713e61e61888663dbd775dc4bae23fe7d0b86e98503190f63\" returns successfully" Jan 24 11:50:01.035132 containerd[1647]: time="2026-01-24T11:50:01.034851652Z" level=info msg="StartContainer for \"51df99e807c575426a0b1b0edb58bf4d25dbd42f2bb9cd6e380438339a94e0ef\" returns successfully" Jan 24 11:50:01.094977 containerd[1647]: time="2026-01-24T11:50:01.094782181Z" level=info msg="StartContainer for \"3925ca274f194ad512a9b7ef8bcd146731e5bb29e3b62574d3ae0af8b50f3c06\" returns successfully" Jan 24 11:50:02.272968 kubelet[2512]: E0124 11:50:02.272926 2512 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:50:02.278251 kubelet[2512]: E0124 11:50:02.275931 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:02.279568 kubelet[2512]: E0124 11:50:02.279540 2512 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:50:02.279919 kubelet[2512]: E0124 11:50:02.279898 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:02.304816 kubelet[2512]: E0124 11:50:02.302093 2512 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:50:02.320489 kubelet[2512]: E0124 11:50:02.310007 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:03.356837 kubelet[2512]: E0124 11:50:03.355494 2512 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:50:03.356837 kubelet[2512]: E0124 11:50:03.356433 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:03.356837 kubelet[2512]: E0124 11:50:03.357077 2512 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:50:03.356837 kubelet[2512]: E0124 11:50:03.357819 2512 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:50:03.356837 kubelet[2512]: E0124 11:50:03.358529 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:03.356837 kubelet[2512]: E0124 11:50:03.358586 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:04.336861 kubelet[2512]: I0124 11:50:04.329137 2512 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:50:04.983722 kubelet[2512]: E0124 11:50:04.982747 2512 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:50:04.983722 kubelet[2512]: E0124 11:50:04.983004 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:04.989849 kubelet[2512]: E0124 11:50:04.984918 2512 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:50:04.989849 kubelet[2512]: E0124 11:50:04.985046 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:05.869245 kubelet[2512]: E0124 11:50:05.867633 2512 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jan 24 11:50:06.475326 kubelet[2512]: E0124 11:50:06.474834 2512 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:50:06.477704 kubelet[2512]: E0124 11:50:06.477595 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:08.659889 kubelet[2512]: E0124 11:50:08.645478 2512 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:50:08.659889 kubelet[2512]: E0124 11:50:08.669671 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:11.252522 kubelet[2512]: E0124 11:50:11.248382 2512 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jan 24 11:50:11.329687 kubelet[2512]: E0124 11:50:11.262635 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:11.663863 kubelet[2512]: E0124 11:50:11.663820 2512 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Jan 24 11:50:11.793690 kubelet[2512]: I0124 11:50:11.793118 2512 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Jan 24 11:50:11.815611 kubelet[2512]: E0124 11:50:11.814988 2512 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.188da8688a107b59 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-01-24 11:49:55.572874073 +0000 UTC m=+1.341655382,LastTimestamp:2026-01-24 11:49:55.572874073 +0000 UTC m=+1.341655382,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jan 24 11:50:11.833676 kubelet[2512]: I0124 11:50:11.833590 2512 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Jan 24 11:50:11.865544 kubelet[2512]: I0124 11:50:11.865384 2512 apiserver.go:52] "Watching apiserver" Jan 24 11:50:11.915720 kubelet[2512]: E0124 11:50:11.915339 2512 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{localhost.188da8688d7a62d5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2026-01-24 11:49:55.630146261 +0000 UTC m=+1.398927571,LastTimestamp:2026-01-24 11:49:55.630146261 +0000 UTC m=+1.398927571,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jan 24 11:50:11.926721 kubelet[2512]: I0124 11:50:11.926585 2512 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Jan 24 11:50:11.949825 kubelet[2512]: E0124 11:50:11.949694 2512 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Jan 24 11:50:11.949965 kubelet[2512]: I0124 11:50:11.949952 2512 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Jan 24 11:50:11.957335 kubelet[2512]: E0124 11:50:11.955611 2512 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Jan 24 11:50:11.957335 kubelet[2512]: I0124 11:50:11.955893 2512 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Jan 24 11:50:12.043003 kubelet[2512]: E0124 11:50:12.037794 2512 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Jan 24 11:50:12.051083 kubelet[2512]: I0124 11:50:12.050387 2512 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Jan 24 11:50:12.053740 kubelet[2512]: E0124 11:50:12.053611 2512 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Jan 24 11:50:12.054096 kubelet[2512]: E0124 11:50:12.053970 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:14.981136 kubelet[2512]: I0124 11:50:14.974056 2512 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Jan 24 11:50:15.029960 kubelet[2512]: E0124 11:50:15.029159 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:15.127219 kubelet[2512]: E0124 11:50:15.127036 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:18.268244 systemd[1]: Reload requested from client PID 2804 ('systemctl') (unit session-8.scope)... Jan 24 11:50:18.268369 systemd[1]: Reloading... Jan 24 11:50:18.512231 kubelet[2512]: I0124 11:50:18.512128 2512 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Jan 24 11:50:18.580370 kubelet[2512]: E0124 11:50:18.580207 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:18.621474 kubelet[2512]: I0124 11:50:18.621247 2512 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=3.6212020259999997 podStartE2EDuration="3.621202026s" podCreationTimestamp="2026-01-24 11:50:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 11:50:15.831131466 +0000 UTC m=+21.599912805" watchObservedRunningTime="2026-01-24 11:50:18.621202026 +0000 UTC m=+24.389983335" Jan 24 11:50:18.665362 zram_generator::config[2847]: No configuration found. Jan 24 11:50:19.173061 kubelet[2512]: I0124 11:50:19.172903 2512 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Jan 24 11:50:19.215382 kubelet[2512]: E0124 11:50:19.203811 2512 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Jan 24 11:50:19.215382 kubelet[2512]: E0124 11:50:19.204039 2512 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:19.705168 systemd[1]: Reloading finished in 1435 ms. Jan 24 11:50:19.775189 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:50:19.811664 systemd[1]: kubelet.service: Deactivated successfully. Jan 24 11:50:19.812162 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:50:19.810000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:50:19.813555 systemd[1]: kubelet.service: Consumed 4.381s CPU time, 132.6M memory peak. Jan 24 11:50:19.814910 kernel: kauditd_printk_skb: 122 callbacks suppressed Jan 24 11:50:19.814954 kernel: audit: type=1131 audit(1769255419.810:388): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:50:19.818363 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 24 11:50:19.824534 kernel: audit: type=1334 audit(1769255419.822:389): prog-id=111 op=LOAD Jan 24 11:50:19.822000 audit: BPF prog-id=111 op=LOAD Jan 24 11:50:19.831231 kernel: audit: type=1334 audit(1769255419.823:390): prog-id=72 op=UNLOAD Jan 24 11:50:19.823000 audit: BPF prog-id=72 op=UNLOAD Jan 24 11:50:19.823000 audit: BPF prog-id=112 op=LOAD Jan 24 11:50:19.839049 kernel: audit: type=1334 audit(1769255419.823:391): prog-id=112 op=LOAD Jan 24 11:50:19.839110 kernel: audit: type=1334 audit(1769255419.823:392): prog-id=113 op=LOAD Jan 24 11:50:19.823000 audit: BPF prog-id=113 op=LOAD Jan 24 11:50:19.823000 audit: BPF prog-id=73 op=UNLOAD Jan 24 11:50:19.847574 kernel: audit: type=1334 audit(1769255419.823:393): prog-id=73 op=UNLOAD Jan 24 11:50:19.847659 kernel: audit: type=1334 audit(1769255419.823:394): prog-id=74 op=UNLOAD Jan 24 11:50:19.847691 kernel: audit: type=1334 audit(1769255419.825:395): prog-id=114 op=LOAD Jan 24 11:50:19.847733 kernel: audit: type=1334 audit(1769255419.828:396): prog-id=69 op=UNLOAD Jan 24 11:50:19.847762 kernel: audit: type=1334 audit(1769255419.828:397): prog-id=115 op=LOAD Jan 24 11:50:19.823000 audit: BPF prog-id=74 op=UNLOAD Jan 24 11:50:19.825000 audit: BPF prog-id=114 op=LOAD Jan 24 11:50:19.828000 audit: BPF prog-id=69 op=UNLOAD Jan 24 11:50:19.828000 audit: BPF prog-id=115 op=LOAD Jan 24 11:50:19.828000 audit: BPF prog-id=116 op=LOAD Jan 24 11:50:19.828000 audit: BPF prog-id=70 op=UNLOAD Jan 24 11:50:19.828000 audit: BPF prog-id=71 op=UNLOAD Jan 24 11:50:19.829000 audit: BPF prog-id=117 op=LOAD Jan 24 11:50:19.829000 audit: BPF prog-id=61 op=UNLOAD Jan 24 11:50:19.838000 audit: BPF prog-id=118 op=LOAD Jan 24 11:50:19.838000 audit: BPF prog-id=68 op=UNLOAD Jan 24 11:50:19.839000 audit: BPF prog-id=119 op=LOAD Jan 24 11:50:19.839000 audit: BPF prog-id=65 op=UNLOAD Jan 24 11:50:19.839000 audit: BPF prog-id=120 op=LOAD Jan 24 11:50:19.839000 audit: BPF prog-id=121 op=LOAD Jan 24 11:50:19.839000 audit: BPF prog-id=66 op=UNLOAD Jan 24 11:50:19.839000 audit: BPF prog-id=67 op=UNLOAD Jan 24 11:50:19.839000 audit: BPF prog-id=122 op=LOAD Jan 24 11:50:19.840000 audit: BPF prog-id=62 op=UNLOAD Jan 24 11:50:19.840000 audit: BPF prog-id=123 op=LOAD Jan 24 11:50:19.840000 audit: BPF prog-id=124 op=LOAD Jan 24 11:50:19.840000 audit: BPF prog-id=63 op=UNLOAD Jan 24 11:50:19.840000 audit: BPF prog-id=64 op=UNLOAD Jan 24 11:50:19.846000 audit: BPF prog-id=125 op=LOAD Jan 24 11:50:19.848000 audit: BPF prog-id=75 op=UNLOAD Jan 24 11:50:19.848000 audit: BPF prog-id=126 op=LOAD Jan 24 11:50:19.848000 audit: BPF prog-id=127 op=LOAD Jan 24 11:50:19.848000 audit: BPF prog-id=76 op=UNLOAD Jan 24 11:50:19.848000 audit: BPF prog-id=77 op=UNLOAD Jan 24 11:50:19.854000 audit: BPF prog-id=128 op=LOAD Jan 24 11:50:19.854000 audit: BPF prog-id=78 op=UNLOAD Jan 24 11:50:19.854000 audit: BPF prog-id=129 op=LOAD Jan 24 11:50:19.854000 audit: BPF prog-id=130 op=LOAD Jan 24 11:50:19.854000 audit: BPF prog-id=79 op=UNLOAD Jan 24 11:50:19.854000 audit: BPF prog-id=80 op=UNLOAD Jan 24 11:50:20.417561 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 24 11:50:20.416000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:50:20.434933 (kubelet)[2895]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 24 11:50:20.779174 kubelet[2895]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 24 11:50:20.779174 kubelet[2895]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jan 24 11:50:20.779174 kubelet[2895]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 24 11:50:20.779174 kubelet[2895]: I0124 11:50:20.774125 2895 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 24 11:50:20.824721 kubelet[2895]: I0124 11:50:20.824122 2895 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Jan 24 11:50:20.824721 kubelet[2895]: I0124 11:50:20.824163 2895 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 24 11:50:20.824721 kubelet[2895]: I0124 11:50:20.824830 2895 server.go:956] "Client rotation is on, will bootstrap in background" Jan 24 11:50:20.829517 kubelet[2895]: I0124 11:50:20.828504 2895 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Jan 24 11:50:20.838512 kubelet[2895]: I0124 11:50:20.834075 2895 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 24 11:50:20.847979 kubelet[2895]: I0124 11:50:20.847251 2895 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 24 11:50:20.864544 kubelet[2895]: I0124 11:50:20.863646 2895 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 24 11:50:20.864544 kubelet[2895]: I0124 11:50:20.864040 2895 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 24 11:50:20.870629 kubelet[2895]: I0124 11:50:20.864086 2895 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 24 11:50:20.870629 kubelet[2895]: I0124 11:50:20.866736 2895 topology_manager.go:138] "Creating topology manager with none policy" Jan 24 11:50:20.870629 kubelet[2895]: I0124 11:50:20.866754 2895 container_manager_linux.go:303] "Creating device plugin manager" Jan 24 11:50:20.870629 kubelet[2895]: I0124 11:50:20.866831 2895 state_mem.go:36] "Initialized new in-memory state store" Jan 24 11:50:20.870629 kubelet[2895]: I0124 11:50:20.867076 2895 kubelet.go:480] "Attempting to sync node with API server" Jan 24 11:50:20.871094 kubelet[2895]: I0124 11:50:20.867093 2895 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 24 11:50:20.871094 kubelet[2895]: I0124 11:50:20.867123 2895 kubelet.go:386] "Adding apiserver pod source" Jan 24 11:50:20.871094 kubelet[2895]: I0124 11:50:20.867147 2895 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 24 11:50:20.871094 kubelet[2895]: I0124 11:50:20.870718 2895 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.1.5" apiVersion="v1" Jan 24 11:50:20.872492 kubelet[2895]: I0124 11:50:20.871524 2895 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Jan 24 11:50:20.889850 kubelet[2895]: I0124 11:50:20.889713 2895 watchdog_linux.go:99] "Systemd watchdog is not enabled" Jan 24 11:50:20.889850 kubelet[2895]: I0124 11:50:20.889804 2895 server.go:1289] "Started kubelet" Jan 24 11:50:20.893220 kubelet[2895]: I0124 11:50:20.892361 2895 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 24 11:50:20.898347 kubelet[2895]: I0124 11:50:20.895812 2895 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Jan 24 11:50:20.903997 kubelet[2895]: I0124 11:50:20.899806 2895 volume_manager.go:297] "Starting Kubelet Volume Manager" Jan 24 11:50:20.915171 kubelet[2895]: I0124 11:50:20.904216 2895 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 24 11:50:20.915171 kubelet[2895]: I0124 11:50:20.906243 2895 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 24 11:50:20.915171 kubelet[2895]: I0124 11:50:20.906672 2895 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Jan 24 11:50:20.915171 kubelet[2895]: I0124 11:50:20.914817 2895 reconciler.go:26] "Reconciler: start to sync state" Jan 24 11:50:20.918366 kubelet[2895]: I0124 11:50:20.918339 2895 factory.go:223] Registration of the containerd container factory successfully Jan 24 11:50:20.918366 kubelet[2895]: I0124 11:50:20.918362 2895 factory.go:223] Registration of the systemd container factory successfully Jan 24 11:50:20.921879 kubelet[2895]: E0124 11:50:20.919239 2895 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 24 11:50:20.921879 kubelet[2895]: I0124 11:50:20.921542 2895 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 24 11:50:20.940109 kubelet[2895]: I0124 11:50:20.938803 2895 server.go:317] "Adding debug handlers to kubelet server" Jan 24 11:50:20.952003 kubelet[2895]: I0124 11:50:20.951621 2895 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 24 11:50:21.066158 kubelet[2895]: I0124 11:50:21.062813 2895 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Jan 24 11:50:21.126523 kubelet[2895]: I0124 11:50:21.124228 2895 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Jan 24 11:50:21.126523 kubelet[2895]: I0124 11:50:21.124387 2895 status_manager.go:230] "Starting to sync pod status with apiserver" Jan 24 11:50:21.131527 kubelet[2895]: I0124 11:50:21.128755 2895 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jan 24 11:50:21.131527 kubelet[2895]: I0124 11:50:21.128795 2895 kubelet.go:2436] "Starting kubelet main sync loop" Jan 24 11:50:21.131527 kubelet[2895]: E0124 11:50:21.128878 2895 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 24 11:50:21.235911 kubelet[2895]: E0124 11:50:21.233877 2895 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jan 24 11:50:21.426013 kubelet[2895]: I0124 11:50:21.425975 2895 cpu_manager.go:221] "Starting CPU manager" policy="none" Jan 24 11:50:21.427319 kubelet[2895]: I0124 11:50:21.426457 2895 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jan 24 11:50:21.427319 kubelet[2895]: I0124 11:50:21.426540 2895 state_mem.go:36] "Initialized new in-memory state store" Jan 24 11:50:21.427319 kubelet[2895]: I0124 11:50:21.426844 2895 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jan 24 11:50:21.427319 kubelet[2895]: I0124 11:50:21.426862 2895 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jan 24 11:50:21.427319 kubelet[2895]: I0124 11:50:21.426888 2895 policy_none.go:49] "None policy: Start" Jan 24 11:50:21.427319 kubelet[2895]: I0124 11:50:21.426898 2895 memory_manager.go:186] "Starting memorymanager" policy="None" Jan 24 11:50:21.427319 kubelet[2895]: I0124 11:50:21.426911 2895 state_mem.go:35] "Initializing new in-memory state store" Jan 24 11:50:21.427319 kubelet[2895]: I0124 11:50:21.427023 2895 state_mem.go:75] "Updated machine memory state" Jan 24 11:50:21.436586 kubelet[2895]: E0124 11:50:21.436346 2895 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jan 24 11:50:21.477562 kubelet[2895]: E0124 11:50:21.476803 2895 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Jan 24 11:50:21.477562 kubelet[2895]: I0124 11:50:21.477095 2895 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 24 11:50:21.477562 kubelet[2895]: I0124 11:50:21.477113 2895 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 24 11:50:21.493226 kubelet[2895]: I0124 11:50:21.492167 2895 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 24 11:50:21.493226 kubelet[2895]: E0124 11:50:21.492757 2895 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jan 24 11:50:21.702653 kubelet[2895]: I0124 11:50:21.699917 2895 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jan 24 11:50:21.753493 kubelet[2895]: I0124 11:50:21.753362 2895 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Jan 24 11:50:21.754382 kubelet[2895]: I0124 11:50:21.753825 2895 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Jan 24 11:50:21.841826 kubelet[2895]: I0124 11:50:21.839762 2895 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Jan 24 11:50:21.841826 kubelet[2895]: I0124 11:50:21.841079 2895 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Jan 24 11:50:21.841826 kubelet[2895]: I0124 11:50:21.841810 2895 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Jan 24 11:50:21.883827 kubelet[2895]: I0124 11:50:21.881587 2895 apiserver.go:52] "Watching apiserver" Jan 24 11:50:21.887562 kubelet[2895]: E0124 11:50:21.884442 2895 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Jan 24 11:50:21.887562 kubelet[2895]: E0124 11:50:21.886173 2895 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Jan 24 11:50:21.908670 kubelet[2895]: I0124 11:50:21.908549 2895 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Jan 24 11:50:21.946134 kubelet[2895]: I0124 11:50:21.945123 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b2417606a328cd170dda8398a7770fe4-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"b2417606a328cd170dda8398a7770fe4\") " pod="kube-system/kube-apiserver-localhost" Jan 24 11:50:21.946134 kubelet[2895]: I0124 11:50:21.945795 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b2417606a328cd170dda8398a7770fe4-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"b2417606a328cd170dda8398a7770fe4\") " pod="kube-system/kube-apiserver-localhost" Jan 24 11:50:21.946134 kubelet[2895]: I0124 11:50:21.946049 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:50:21.946134 kubelet[2895]: I0124 11:50:21.946251 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:50:21.946134 kubelet[2895]: I0124 11:50:21.946338 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b2417606a328cd170dda8398a7770fe4-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"b2417606a328cd170dda8398a7770fe4\") " pod="kube-system/kube-apiserver-localhost" Jan 24 11:50:21.950220 kubelet[2895]: I0124 11:50:21.946458 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:50:21.950220 kubelet[2895]: I0124 11:50:21.946483 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:50:21.950220 kubelet[2895]: I0124 11:50:21.946505 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/66e26b992bcd7ea6fb75e339cf7a3f7d-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"66e26b992bcd7ea6fb75e339cf7a3f7d\") " pod="kube-system/kube-controller-manager-localhost" Jan 24 11:50:21.950220 kubelet[2895]: I0124 11:50:21.946528 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6e6cfcfb327385445a9bb0d2bc2fd5d4-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6e6cfcfb327385445a9bb0d2bc2fd5d4\") " pod="kube-system/kube-scheduler-localhost" Jan 24 11:50:22.193236 kubelet[2895]: E0124 11:50:22.192666 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:22.199960 kubelet[2895]: E0124 11:50:22.194594 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:22.199960 kubelet[2895]: E0124 11:50:22.196779 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:22.364513 kubelet[2895]: I0124 11:50:22.363644 2895 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=4.363599083 podStartE2EDuration="4.363599083s" podCreationTimestamp="2026-01-24 11:50:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 11:50:22.350382705 +0000 UTC m=+1.872175431" watchObservedRunningTime="2026-01-24 11:50:22.363599083 +0000 UTC m=+1.885391789" Jan 24 11:50:22.390060 kubelet[2895]: I0124 11:50:22.389988 2895 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.389970004 podStartE2EDuration="1.389970004s" podCreationTimestamp="2026-01-24 11:50:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 11:50:22.389723314 +0000 UTC m=+1.911516040" watchObservedRunningTime="2026-01-24 11:50:22.389970004 +0000 UTC m=+1.911762710" Jan 24 11:50:22.545030 kubelet[2895]: I0124 11:50:22.539829 2895 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jan 24 11:50:22.545030 kubelet[2895]: I0124 11:50:22.541478 2895 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jan 24 11:50:22.545988 containerd[1647]: time="2026-01-24T11:50:22.540849453Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jan 24 11:50:23.340513 kubelet[2895]: E0124 11:50:23.339720 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:23.358158 kubelet[2895]: E0124 11:50:23.341378 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:23.358158 kubelet[2895]: E0124 11:50:23.341911 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:23.691465 systemd[1]: Created slice kubepods-besteffort-podd37a6796_3091_4e57_8516_b1afa67b5233.slice - libcontainer container kubepods-besteffort-podd37a6796_3091_4e57_8516_b1afa67b5233.slice. Jan 24 11:50:23.829607 kubelet[2895]: I0124 11:50:23.829517 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/d37a6796-3091-4e57-8516-b1afa67b5233-kube-proxy\") pod \"kube-proxy-fvhxl\" (UID: \"d37a6796-3091-4e57-8516-b1afa67b5233\") " pod="kube-system/kube-proxy-fvhxl" Jan 24 11:50:23.829607 kubelet[2895]: I0124 11:50:23.829611 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d37a6796-3091-4e57-8516-b1afa67b5233-lib-modules\") pod \"kube-proxy-fvhxl\" (UID: \"d37a6796-3091-4e57-8516-b1afa67b5233\") " pod="kube-system/kube-proxy-fvhxl" Jan 24 11:50:23.829856 kubelet[2895]: I0124 11:50:23.829646 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd8d8\" (UniqueName: \"kubernetes.io/projected/d37a6796-3091-4e57-8516-b1afa67b5233-kube-api-access-nd8d8\") pod \"kube-proxy-fvhxl\" (UID: \"d37a6796-3091-4e57-8516-b1afa67b5233\") " pod="kube-system/kube-proxy-fvhxl" Jan 24 11:50:23.829856 kubelet[2895]: I0124 11:50:23.829692 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d37a6796-3091-4e57-8516-b1afa67b5233-xtables-lock\") pod \"kube-proxy-fvhxl\" (UID: \"d37a6796-3091-4e57-8516-b1afa67b5233\") " pod="kube-system/kube-proxy-fvhxl" Jan 24 11:50:23.975197 systemd[1]: Created slice kubepods-besteffort-pod0c411ea2_67ac_46d0_ad61_7975c1c76616.slice - libcontainer container kubepods-besteffort-pod0c411ea2_67ac_46d0_ad61_7975c1c76616.slice. Jan 24 11:50:24.011044 kubelet[2895]: E0124 11:50:24.010949 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:24.012811 containerd[1647]: time="2026-01-24T11:50:24.012701048Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-fvhxl,Uid:d37a6796-3091-4e57-8516-b1afa67b5233,Namespace:kube-system,Attempt:0,}" Jan 24 11:50:24.030632 kubelet[2895]: I0124 11:50:24.030597 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dtrf\" (UniqueName: \"kubernetes.io/projected/0c411ea2-67ac-46d0-ad61-7975c1c76616-kube-api-access-8dtrf\") pod \"tigera-operator-7dcd859c48-9rh4s\" (UID: \"0c411ea2-67ac-46d0-ad61-7975c1c76616\") " pod="tigera-operator/tigera-operator-7dcd859c48-9rh4s" Jan 24 11:50:24.031059 kubelet[2895]: I0124 11:50:24.031002 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/0c411ea2-67ac-46d0-ad61-7975c1c76616-var-lib-calico\") pod \"tigera-operator-7dcd859c48-9rh4s\" (UID: \"0c411ea2-67ac-46d0-ad61-7975c1c76616\") " pod="tigera-operator/tigera-operator-7dcd859c48-9rh4s" Jan 24 11:50:24.109176 containerd[1647]: time="2026-01-24T11:50:24.109098899Z" level=info msg="connecting to shim 1b73dd4dd60ae4efffc812cb34490e22fb4483fbc524952fde69a2db0365e0d1" address="unix:///run/containerd/s/75f14aa241c47ffae2f81b58d97f45f685610af46d060bd88d767b58ea8ed788" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:50:24.195758 systemd[1]: Started cri-containerd-1b73dd4dd60ae4efffc812cb34490e22fb4483fbc524952fde69a2db0365e0d1.scope - libcontainer container 1b73dd4dd60ae4efffc812cb34490e22fb4483fbc524952fde69a2db0365e0d1. Jan 24 11:50:24.226000 audit: BPF prog-id=131 op=LOAD Jan 24 11:50:24.227000 audit: BPF prog-id=132 op=LOAD Jan 24 11:50:24.227000 audit[2969]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=2959 pid=2969 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.227000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3162373364643464643630616534656666666338313263623334343930 Jan 24 11:50:24.227000 audit: BPF prog-id=132 op=UNLOAD Jan 24 11:50:24.227000 audit[2969]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2959 pid=2969 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.227000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3162373364643464643630616534656666666338313263623334343930 Jan 24 11:50:24.227000 audit: BPF prog-id=133 op=LOAD Jan 24 11:50:24.227000 audit[2969]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=2959 pid=2969 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.227000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3162373364643464643630616534656666666338313263623334343930 Jan 24 11:50:24.227000 audit: BPF prog-id=134 op=LOAD Jan 24 11:50:24.227000 audit[2969]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=2959 pid=2969 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.227000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3162373364643464643630616534656666666338313263623334343930 Jan 24 11:50:24.228000 audit: BPF prog-id=134 op=UNLOAD Jan 24 11:50:24.228000 audit[2969]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2959 pid=2969 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.228000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3162373364643464643630616534656666666338313263623334343930 Jan 24 11:50:24.228000 audit: BPF prog-id=133 op=UNLOAD Jan 24 11:50:24.228000 audit[2969]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2959 pid=2969 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.228000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3162373364643464643630616534656666666338313263623334343930 Jan 24 11:50:24.228000 audit: BPF prog-id=135 op=LOAD Jan 24 11:50:24.228000 audit[2969]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=2959 pid=2969 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.228000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3162373364643464643630616534656666666338313263623334343930 Jan 24 11:50:24.266727 containerd[1647]: time="2026-01-24T11:50:24.265961610Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-fvhxl,Uid:d37a6796-3091-4e57-8516-b1afa67b5233,Namespace:kube-system,Attempt:0,} returns sandbox id \"1b73dd4dd60ae4efffc812cb34490e22fb4483fbc524952fde69a2db0365e0d1\"" Jan 24 11:50:24.269352 kubelet[2895]: E0124 11:50:24.269179 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:24.292805 containerd[1647]: time="2026-01-24T11:50:24.288047367Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-9rh4s,Uid:0c411ea2-67ac-46d0-ad61-7975c1c76616,Namespace:tigera-operator,Attempt:0,}" Jan 24 11:50:24.324146 containerd[1647]: time="2026-01-24T11:50:24.323153115Z" level=info msg="CreateContainer within sandbox \"1b73dd4dd60ae4efffc812cb34490e22fb4483fbc524952fde69a2db0365e0d1\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jan 24 11:50:24.338352 kubelet[2895]: E0124 11:50:24.337928 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:24.339129 kubelet[2895]: E0124 11:50:24.339099 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:24.372551 containerd[1647]: time="2026-01-24T11:50:24.372443292Z" level=info msg="Container 38302867f034ec0fd420a8dff0d0fbfa3963ce6f6d596f8b36cdaaa4a435438e: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:50:24.382966 containerd[1647]: time="2026-01-24T11:50:24.382871253Z" level=info msg="connecting to shim b6e34f1466bd97b7aca805874a33df28311926530eda9465360444b60d41df9d" address="unix:///run/containerd/s/e55f705d53dda65ce312b92ba2107ed2765c020c727089568e0ecae6d373ed99" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:50:24.395065 containerd[1647]: time="2026-01-24T11:50:24.394785967Z" level=info msg="CreateContainer within sandbox \"1b73dd4dd60ae4efffc812cb34490e22fb4483fbc524952fde69a2db0365e0d1\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"38302867f034ec0fd420a8dff0d0fbfa3963ce6f6d596f8b36cdaaa4a435438e\"" Jan 24 11:50:24.401212 containerd[1647]: time="2026-01-24T11:50:24.399172299Z" level=info msg="StartContainer for \"38302867f034ec0fd420a8dff0d0fbfa3963ce6f6d596f8b36cdaaa4a435438e\"" Jan 24 11:50:24.403852 containerd[1647]: time="2026-01-24T11:50:24.403748643Z" level=info msg="connecting to shim 38302867f034ec0fd420a8dff0d0fbfa3963ce6f6d596f8b36cdaaa4a435438e" address="unix:///run/containerd/s/75f14aa241c47ffae2f81b58d97f45f685610af46d060bd88d767b58ea8ed788" protocol=ttrpc version=3 Jan 24 11:50:24.458141 systemd[1]: Started cri-containerd-38302867f034ec0fd420a8dff0d0fbfa3963ce6f6d596f8b36cdaaa4a435438e.scope - libcontainer container 38302867f034ec0fd420a8dff0d0fbfa3963ce6f6d596f8b36cdaaa4a435438e. Jan 24 11:50:24.499619 systemd[1]: Started cri-containerd-b6e34f1466bd97b7aca805874a33df28311926530eda9465360444b60d41df9d.scope - libcontainer container b6e34f1466bd97b7aca805874a33df28311926530eda9465360444b60d41df9d. Jan 24 11:50:24.736000 audit: BPF prog-id=136 op=LOAD Jan 24 11:50:24.738000 audit: BPF prog-id=137 op=LOAD Jan 24 11:50:24.738000 audit[3023]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=3006 pid=3023 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.738000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6236653334663134363662643937623761636138303538373461333364 Jan 24 11:50:24.739000 audit: BPF prog-id=137 op=UNLOAD Jan 24 11:50:24.739000 audit[3023]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3006 pid=3023 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.739000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6236653334663134363662643937623761636138303538373461333364 Jan 24 11:50:24.739000 audit: BPF prog-id=138 op=LOAD Jan 24 11:50:24.739000 audit[3023]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=3006 pid=3023 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.739000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6236653334663134363662643937623761636138303538373461333364 Jan 24 11:50:24.739000 audit: BPF prog-id=139 op=LOAD Jan 24 11:50:24.739000 audit[3023]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=3006 pid=3023 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.739000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6236653334663134363662643937623761636138303538373461333364 Jan 24 11:50:24.739000 audit: BPF prog-id=139 op=UNLOAD Jan 24 11:50:24.739000 audit[3023]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3006 pid=3023 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.739000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6236653334663134363662643937623761636138303538373461333364 Jan 24 11:50:24.739000 audit: BPF prog-id=138 op=UNLOAD Jan 24 11:50:24.739000 audit[3023]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3006 pid=3023 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.739000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6236653334663134363662643937623761636138303538373461333364 Jan 24 11:50:24.739000 audit: BPF prog-id=140 op=LOAD Jan 24 11:50:24.739000 audit[3023]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=3006 pid=3023 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.739000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6236653334663134363662643937623761636138303538373461333364 Jan 24 11:50:24.759000 audit: BPF prog-id=141 op=LOAD Jan 24 11:50:24.759000 audit[3016]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=2959 pid=3016 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.759000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3338333032383637663033346563306664343230613864666630643066 Jan 24 11:50:24.759000 audit: BPF prog-id=142 op=LOAD Jan 24 11:50:24.759000 audit[3016]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=2959 pid=3016 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.759000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3338333032383637663033346563306664343230613864666630643066 Jan 24 11:50:24.759000 audit: BPF prog-id=142 op=UNLOAD Jan 24 11:50:24.759000 audit[3016]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2959 pid=3016 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.759000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3338333032383637663033346563306664343230613864666630643066 Jan 24 11:50:24.759000 audit: BPF prog-id=141 op=UNLOAD Jan 24 11:50:24.759000 audit[3016]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2959 pid=3016 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.759000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3338333032383637663033346563306664343230613864666630643066 Jan 24 11:50:24.760000 audit: BPF prog-id=143 op=LOAD Jan 24 11:50:24.760000 audit[3016]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=2959 pid=3016 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:24.760000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3338333032383637663033346563306664343230613864666630643066 Jan 24 11:50:24.844032 containerd[1647]: time="2026-01-24T11:50:24.843935499Z" level=info msg="StartContainer for \"38302867f034ec0fd420a8dff0d0fbfa3963ce6f6d596f8b36cdaaa4a435438e\" returns successfully" Jan 24 11:50:24.888610 containerd[1647]: time="2026-01-24T11:50:24.887754382Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-9rh4s,Uid:0c411ea2-67ac-46d0-ad61-7975c1c76616,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"b6e34f1466bd97b7aca805874a33df28311926530eda9465360444b60d41df9d\"" Jan 24 11:50:24.931209 containerd[1647]: time="2026-01-24T11:50:24.930200428Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\"" Jan 24 11:50:25.344334 kubelet[2895]: E0124 11:50:25.343687 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:25.362000 audit[3110]: NETFILTER_CFG table=mangle:54 family=10 entries=1 op=nft_register_chain pid=3110 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.366579 kernel: kauditd_printk_skb: 91 callbacks suppressed Jan 24 11:50:25.366824 kernel: audit: type=1325 audit(1769255425.362:451): table=mangle:54 family=10 entries=1 op=nft_register_chain pid=3110 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.375478 kernel: audit: type=1300 audit(1769255425.362:451): arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe98ebbf10 a2=0 a3=7ffe98ebbefc items=0 ppid=3040 pid=3110 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.362000 audit[3110]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe98ebbf10 a2=0 a3=7ffe98ebbefc items=0 ppid=3040 pid=3110 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.362000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Jan 24 11:50:25.399430 kernel: audit: type=1327 audit(1769255425.362:451): proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Jan 24 11:50:25.369000 audit[3113]: NETFILTER_CFG table=nat:55 family=10 entries=1 op=nft_register_chain pid=3113 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.369000 audit[3113]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd8633ae90 a2=0 a3=7ffd8633ae7c items=0 ppid=3040 pid=3113 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.442022 kernel: audit: type=1325 audit(1769255425.369:452): table=nat:55 family=10 entries=1 op=nft_register_chain pid=3113 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.442159 kernel: audit: type=1300 audit(1769255425.369:452): arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd8633ae90 a2=0 a3=7ffd8633ae7c items=0 ppid=3040 pid=3113 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.369000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Jan 24 11:50:25.372000 audit[3111]: NETFILTER_CFG table=mangle:56 family=2 entries=1 op=nft_register_chain pid=3111 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.461060 kernel: audit: type=1327 audit(1769255425.369:452): proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Jan 24 11:50:25.461138 kernel: audit: type=1325 audit(1769255425.372:453): table=mangle:56 family=2 entries=1 op=nft_register_chain pid=3111 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.372000 audit[3111]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffda2bf17e0 a2=0 a3=7ffda2bf17cc items=0 ppid=3040 pid=3111 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.472151 kernel: audit: type=1300 audit(1769255425.372:453): arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffda2bf17e0 a2=0 a3=7ffda2bf17cc items=0 ppid=3040 pid=3111 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.472230 kernel: audit: type=1327 audit(1769255425.372:453): proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Jan 24 11:50:25.372000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Jan 24 11:50:25.389000 audit[3116]: NETFILTER_CFG table=nat:57 family=2 entries=1 op=nft_register_chain pid=3116 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.485343 kernel: audit: type=1325 audit(1769255425.389:454): table=nat:57 family=2 entries=1 op=nft_register_chain pid=3116 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.389000 audit[3116]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff90e72d10 a2=0 a3=7fff90e72cfc items=0 ppid=3040 pid=3116 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.389000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Jan 24 11:50:25.396000 audit[3119]: NETFILTER_CFG table=filter:58 family=2 entries=1 op=nft_register_chain pid=3119 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.396000 audit[3119]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffd909b6e00 a2=0 a3=7ffd909b6dec items=0 ppid=3040 pid=3119 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.396000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Jan 24 11:50:25.400000 audit[3114]: NETFILTER_CFG table=filter:59 family=10 entries=1 op=nft_register_chain pid=3114 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.400000 audit[3114]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fff0c72c990 a2=0 a3=7fff0c72c97c items=0 ppid=3040 pid=3114 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.400000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Jan 24 11:50:25.476000 audit[3120]: NETFILTER_CFG table=filter:60 family=2 entries=1 op=nft_register_chain pid=3120 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.476000 audit[3120]: SYSCALL arch=c000003e syscall=46 success=yes exit=108 a0=3 a1=7ffde8ec5bc0 a2=0 a3=7ffde8ec5bac items=0 ppid=3040 pid=3120 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.476000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 Jan 24 11:50:25.486000 audit[3122]: NETFILTER_CFG table=filter:61 family=2 entries=1 op=nft_register_rule pid=3122 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.486000 audit[3122]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7fff3bf7cf40 a2=0 a3=7fff3bf7cf2c items=0 ppid=3040 pid=3122 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.486000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C652073657276696365 Jan 24 11:50:25.497000 audit[3125]: NETFILTER_CFG table=filter:62 family=2 entries=1 op=nft_register_rule pid=3125 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.497000 audit[3125]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffc9c41b710 a2=0 a3=7ffc9c41b6fc items=0 ppid=3040 pid=3125 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.497000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669 Jan 24 11:50:25.507000 audit[3126]: NETFILTER_CFG table=filter:63 family=2 entries=1 op=nft_register_chain pid=3126 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.507000 audit[3126]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffcceb39930 a2=0 a3=7ffcceb3991c items=0 ppid=3040 pid=3126 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.507000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4E4F4445504F525453002D740066696C746572 Jan 24 11:50:25.515000 audit[3129]: NETFILTER_CFG table=filter:64 family=2 entries=1 op=nft_register_rule pid=3129 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.515000 audit[3129]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7fff415c6d10 a2=0 a3=7fff415c6cfc items=0 ppid=3040 pid=3129 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.515000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 Jan 24 11:50:25.519000 audit[3130]: NETFILTER_CFG table=filter:65 family=2 entries=1 op=nft_register_chain pid=3130 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.519000 audit[3130]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff28f4ece0 a2=0 a3=7fff28f4eccc items=0 ppid=3040 pid=3130 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.519000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D740066696C746572 Jan 24 11:50:25.525000 audit[3132]: NETFILTER_CFG table=filter:66 family=2 entries=1 op=nft_register_rule pid=3132 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.525000 audit[3132]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffc4a020e50 a2=0 a3=7ffc4a020e3c items=0 ppid=3040 pid=3132 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.525000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D Jan 24 11:50:25.538000 audit[3135]: NETFILTER_CFG table=filter:67 family=2 entries=1 op=nft_register_rule pid=3135 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.538000 audit[3135]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffe24da78a0 a2=0 a3=7ffe24da788c items=0 ppid=3040 pid=3135 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.538000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D53 Jan 24 11:50:25.544000 audit[3136]: NETFILTER_CFG table=filter:68 family=2 entries=1 op=nft_register_chain pid=3136 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.544000 audit[3136]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fffced32760 a2=0 a3=7fffced3274c items=0 ppid=3040 pid=3136 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.544000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D464F5257415244002D740066696C746572 Jan 24 11:50:25.550000 audit[3138]: NETFILTER_CFG table=filter:69 family=2 entries=1 op=nft_register_rule pid=3138 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.550000 audit[3138]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffc602114d0 a2=0 a3=7ffc602114bc items=0 ppid=3040 pid=3138 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.550000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 Jan 24 11:50:25.554000 audit[3139]: NETFILTER_CFG table=filter:70 family=2 entries=1 op=nft_register_chain pid=3139 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.554000 audit[3139]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe05869160 a2=0 a3=7ffe0586914c items=0 ppid=3040 pid=3139 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.554000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 Jan 24 11:50:25.560000 audit[3141]: NETFILTER_CFG table=filter:71 family=2 entries=1 op=nft_register_rule pid=3141 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.560000 audit[3141]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffd66b232f0 a2=0 a3=7ffd66b232dc items=0 ppid=3040 pid=3141 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.560000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Jan 24 11:50:25.572000 audit[3144]: NETFILTER_CFG table=filter:72 family=2 entries=1 op=nft_register_rule pid=3144 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.572000 audit[3144]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffcaa720470 a2=0 a3=7ffcaa72045c items=0 ppid=3040 pid=3144 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.572000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Jan 24 11:50:25.589000 audit[3147]: NETFILTER_CFG table=filter:73 family=2 entries=1 op=nft_register_rule pid=3147 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.589000 audit[3147]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffcf03adad0 a2=0 a3=7ffcf03adabc items=0 ppid=3040 pid=3147 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.589000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D Jan 24 11:50:25.592000 audit[3148]: NETFILTER_CFG table=nat:74 family=2 entries=1 op=nft_register_chain pid=3148 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.592000 audit[3148]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffc2970de20 a2=0 a3=7ffc2970de0c items=0 ppid=3040 pid=3148 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.592000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D74006E6174 Jan 24 11:50:25.607000 audit[3150]: NETFILTER_CFG table=nat:75 family=2 entries=1 op=nft_register_rule pid=3150 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.607000 audit[3150]: SYSCALL arch=c000003e syscall=46 success=yes exit=524 a0=3 a1=7ffd0e0e2030 a2=0 a3=7ffd0e0e201c items=0 ppid=3040 pid=3150 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.607000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 24 11:50:25.619000 audit[3153]: NETFILTER_CFG table=nat:76 family=2 entries=1 op=nft_register_rule pid=3153 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.619000 audit[3153]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7fff074da2a0 a2=0 a3=7fff074da28c items=0 ppid=3040 pid=3153 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.619000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 24 11:50:25.625000 audit[3154]: NETFILTER_CFG table=nat:77 family=2 entries=1 op=nft_register_chain pid=3154 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.625000 audit[3154]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe5f3fc6a0 a2=0 a3=7ffe5f3fc68c items=0 ppid=3040 pid=3154 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.625000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Jan 24 11:50:25.633000 audit[3156]: NETFILTER_CFG table=nat:78 family=2 entries=1 op=nft_register_rule pid=3156 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 24 11:50:25.633000 audit[3156]: SYSCALL arch=c000003e syscall=46 success=yes exit=532 a0=3 a1=7ffcfa700bb0 a2=0 a3=7ffcfa700b9c items=0 ppid=3040 pid=3156 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.633000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 Jan 24 11:50:25.694000 audit[3162]: NETFILTER_CFG table=filter:79 family=2 entries=8 op=nft_register_rule pid=3162 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:25.694000 audit[3162]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffc17d0b4d0 a2=0 a3=7ffc17d0b4bc items=0 ppid=3040 pid=3162 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.694000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:25.712000 audit[3162]: NETFILTER_CFG table=nat:80 family=2 entries=14 op=nft_register_chain pid=3162 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:25.712000 audit[3162]: SYSCALL arch=c000003e syscall=46 success=yes exit=5508 a0=3 a1=7ffc17d0b4d0 a2=0 a3=7ffc17d0b4bc items=0 ppid=3040 pid=3162 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.712000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:25.717000 audit[3167]: NETFILTER_CFG table=filter:81 family=10 entries=1 op=nft_register_chain pid=3167 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.717000 audit[3167]: SYSCALL arch=c000003e syscall=46 success=yes exit=108 a0=3 a1=7fffad44fc60 a2=0 a3=7fffad44fc4c items=0 ppid=3040 pid=3167 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.717000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 Jan 24 11:50:25.724000 audit[3169]: NETFILTER_CFG table=filter:82 family=10 entries=2 op=nft_register_chain pid=3169 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.724000 audit[3169]: SYSCALL arch=c000003e syscall=46 success=yes exit=836 a0=3 a1=7ffef4eb6550 a2=0 a3=7ffef4eb653c items=0 ppid=3040 pid=3169 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.724000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C6520736572766963 Jan 24 11:50:25.733000 audit[3172]: NETFILTER_CFG table=filter:83 family=10 entries=1 op=nft_register_rule pid=3172 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.733000 audit[3172]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffcbfc18440 a2=0 a3=7ffcbfc1842c items=0 ppid=3040 pid=3172 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.733000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C652073657276 Jan 24 11:50:25.736000 audit[3173]: NETFILTER_CFG table=filter:84 family=10 entries=1 op=nft_register_chain pid=3173 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.736000 audit[3173]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff34e6eef0 a2=0 a3=7fff34e6eedc items=0 ppid=3040 pid=3173 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.736000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4E4F4445504F525453002D740066696C746572 Jan 24 11:50:25.742000 audit[3175]: NETFILTER_CFG table=filter:85 family=10 entries=1 op=nft_register_rule pid=3175 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.742000 audit[3175]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffca25ecdf0 a2=0 a3=7ffca25ecddc items=0 ppid=3040 pid=3175 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.742000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 Jan 24 11:50:25.749000 audit[3176]: NETFILTER_CFG table=filter:86 family=10 entries=1 op=nft_register_chain pid=3176 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.749000 audit[3176]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffeaf2ce4f0 a2=0 a3=7ffeaf2ce4dc items=0 ppid=3040 pid=3176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.749000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D740066696C746572 Jan 24 11:50:25.758000 audit[3178]: NETFILTER_CFG table=filter:87 family=10 entries=1 op=nft_register_rule pid=3178 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.758000 audit[3178]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffe2206b380 a2=0 a3=7ffe2206b36c items=0 ppid=3040 pid=3178 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.758000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B554245 Jan 24 11:50:25.772000 audit[3181]: NETFILTER_CFG table=filter:88 family=10 entries=2 op=nft_register_chain pid=3181 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.772000 audit[3181]: SYSCALL arch=c000003e syscall=46 success=yes exit=828 a0=3 a1=7fffa345f7d0 a2=0 a3=7fffa345f7bc items=0 ppid=3040 pid=3181 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.772000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D Jan 24 11:50:25.775000 audit[3182]: NETFILTER_CFG table=filter:89 family=10 entries=1 op=nft_register_chain pid=3182 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.775000 audit[3182]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc91b98320 a2=0 a3=7ffc91b9830c items=0 ppid=3040 pid=3182 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.775000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D464F5257415244002D740066696C746572 Jan 24 11:50:25.782000 audit[3184]: NETFILTER_CFG table=filter:90 family=10 entries=1 op=nft_register_rule pid=3184 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.782000 audit[3184]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffc0e155cf0 a2=0 a3=7ffc0e155cdc items=0 ppid=3040 pid=3184 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.782000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 Jan 24 11:50:25.788000 audit[3185]: NETFILTER_CFG table=filter:91 family=10 entries=1 op=nft_register_chain pid=3185 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.788000 audit[3185]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fffac358fa0 a2=0 a3=7fffac358f8c items=0 ppid=3040 pid=3185 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.788000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 Jan 24 11:50:25.798000 audit[3187]: NETFILTER_CFG table=filter:92 family=10 entries=1 op=nft_register_rule pid=3187 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.798000 audit[3187]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffda2e302d0 a2=0 a3=7ffda2e302bc items=0 ppid=3040 pid=3187 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.798000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Jan 24 11:50:25.810000 audit[3190]: NETFILTER_CFG table=filter:93 family=10 entries=1 op=nft_register_rule pid=3190 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.810000 audit[3190]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffd9168be10 a2=0 a3=7ffd9168bdfc items=0 ppid=3040 pid=3190 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.810000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D Jan 24 11:50:25.829000 audit[3197]: NETFILTER_CFG table=filter:94 family=10 entries=1 op=nft_register_rule pid=3197 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.829000 audit[3197]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffc3b44c540 a2=0 a3=7ffc3b44c52c items=0 ppid=3040 pid=3197 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.829000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C Jan 24 11:50:25.832000 audit[3198]: NETFILTER_CFG table=nat:95 family=10 entries=1 op=nft_register_chain pid=3198 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.832000 audit[3198]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffe7544d9c0 a2=0 a3=7ffe7544d9ac items=0 ppid=3040 pid=3198 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.832000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D74006E6174 Jan 24 11:50:25.840000 audit[3200]: NETFILTER_CFG table=nat:96 family=10 entries=1 op=nft_register_rule pid=3200 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.840000 audit[3200]: SYSCALL arch=c000003e syscall=46 success=yes exit=524 a0=3 a1=7ffd6579b020 a2=0 a3=7ffd6579b00c items=0 ppid=3040 pid=3200 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.840000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 24 11:50:25.858000 audit[3203]: NETFILTER_CFG table=nat:97 family=10 entries=1 op=nft_register_rule pid=3203 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.858000 audit[3203]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffec3bb8600 a2=0 a3=7ffec3bb85ec items=0 ppid=3040 pid=3203 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.858000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 24 11:50:25.862000 audit[3204]: NETFILTER_CFG table=nat:98 family=10 entries=1 op=nft_register_chain pid=3204 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.862000 audit[3204]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff9953ad50 a2=0 a3=7fff9953ad3c items=0 ppid=3040 pid=3204 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.862000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Jan 24 11:50:25.868000 audit[3206]: NETFILTER_CFG table=nat:99 family=10 entries=2 op=nft_register_chain pid=3206 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.868000 audit[3206]: SYSCALL arch=c000003e syscall=46 success=yes exit=612 a0=3 a1=7ffd3551e1b0 a2=0 a3=7ffd3551e19c items=0 ppid=3040 pid=3206 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.868000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 Jan 24 11:50:25.871000 audit[3207]: NETFILTER_CFG table=filter:100 family=10 entries=1 op=nft_register_chain pid=3207 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.871000 audit[3207]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffcb9034b30 a2=0 a3=7ffcb9034b1c items=0 ppid=3040 pid=3207 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.871000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 Jan 24 11:50:25.877000 audit[3209]: NETFILTER_CFG table=filter:101 family=10 entries=1 op=nft_register_rule pid=3209 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.877000 audit[3209]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7ffe6bde5260 a2=0 a3=7ffe6bde524c items=0 ppid=3040 pid=3209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.877000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 24 11:50:25.881989 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1997322761.mount: Deactivated successfully. Jan 24 11:50:25.892000 audit[3212]: NETFILTER_CFG table=filter:102 family=10 entries=1 op=nft_register_rule pid=3212 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 24 11:50:25.892000 audit[3212]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7fff803cdfb0 a2=0 a3=7fff803cdf9c items=0 ppid=3040 pid=3212 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.892000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 24 11:50:25.907000 audit[3214]: NETFILTER_CFG table=filter:103 family=10 entries=3 op=nft_register_rule pid=3214 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" Jan 24 11:50:25.907000 audit[3214]: SYSCALL arch=c000003e syscall=46 success=yes exit=2088 a0=3 a1=7ffcdf299700 a2=0 a3=7ffcdf2996ec items=0 ppid=3040 pid=3214 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.907000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:25.908000 audit[3214]: NETFILTER_CFG table=nat:104 family=10 entries=7 op=nft_register_chain pid=3214 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" Jan 24 11:50:25.908000 audit[3214]: SYSCALL arch=c000003e syscall=46 success=yes exit=2056 a0=3 a1=7ffcdf299700 a2=0 a3=7ffcdf2996ec items=0 ppid=3040 pid=3214 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:25.908000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:26.364912 kubelet[2895]: E0124 11:50:26.361368 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:28.061986 containerd[1647]: time="2026-01-24T11:50:28.061235653Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:50:28.066331 containerd[1647]: time="2026-01-24T11:50:28.064797085Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.7: active requests=0, bytes read=23558945" Jan 24 11:50:28.069751 containerd[1647]: time="2026-01-24T11:50:28.069690590Z" level=info msg="ImageCreate event name:\"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:50:28.072723 containerd[1647]: time="2026-01-24T11:50:28.072687491Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:50:28.074640 containerd[1647]: time="2026-01-24T11:50:28.074469359Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.7\" with image id \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\", repo tag \"quay.io/tigera/operator:v1.38.7\", repo digest \"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\", size \"25057686\" in 3.144222594s" Jan 24 11:50:28.074640 containerd[1647]: time="2026-01-24T11:50:28.074532526Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\" returns image reference \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\"" Jan 24 11:50:28.093827 containerd[1647]: time="2026-01-24T11:50:28.091357839Z" level=info msg="CreateContainer within sandbox \"b6e34f1466bd97b7aca805874a33df28311926530eda9465360444b60d41df9d\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Jan 24 11:50:28.123335 containerd[1647]: time="2026-01-24T11:50:28.122521342Z" level=info msg="Container 9aa5d59c7ad44f4002a12d402c2dda1b69ae1ac08f318da787aaffc1e16db2de: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:50:28.134844 containerd[1647]: time="2026-01-24T11:50:28.134786783Z" level=info msg="CreateContainer within sandbox \"b6e34f1466bd97b7aca805874a33df28311926530eda9465360444b60d41df9d\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"9aa5d59c7ad44f4002a12d402c2dda1b69ae1ac08f318da787aaffc1e16db2de\"" Jan 24 11:50:28.136907 containerd[1647]: time="2026-01-24T11:50:28.136833404Z" level=info msg="StartContainer for \"9aa5d59c7ad44f4002a12d402c2dda1b69ae1ac08f318da787aaffc1e16db2de\"" Jan 24 11:50:28.138972 containerd[1647]: time="2026-01-24T11:50:28.138759230Z" level=info msg="connecting to shim 9aa5d59c7ad44f4002a12d402c2dda1b69ae1ac08f318da787aaffc1e16db2de" address="unix:///run/containerd/s/e55f705d53dda65ce312b92ba2107ed2765c020c727089568e0ecae6d373ed99" protocol=ttrpc version=3 Jan 24 11:50:28.184900 systemd[1]: Started cri-containerd-9aa5d59c7ad44f4002a12d402c2dda1b69ae1ac08f318da787aaffc1e16db2de.scope - libcontainer container 9aa5d59c7ad44f4002a12d402c2dda1b69ae1ac08f318da787aaffc1e16db2de. Jan 24 11:50:28.218000 audit: BPF prog-id=144 op=LOAD Jan 24 11:50:28.220000 audit: BPF prog-id=145 op=LOAD Jan 24 11:50:28.220000 audit[3219]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=3006 pid=3219 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:28.220000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961613564353963376164343466343030326131326434303263326464 Jan 24 11:50:28.221000 audit: BPF prog-id=145 op=UNLOAD Jan 24 11:50:28.221000 audit[3219]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3006 pid=3219 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:28.221000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961613564353963376164343466343030326131326434303263326464 Jan 24 11:50:28.221000 audit: BPF prog-id=146 op=LOAD Jan 24 11:50:28.221000 audit[3219]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=3006 pid=3219 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:28.221000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961613564353963376164343466343030326131326434303263326464 Jan 24 11:50:28.221000 audit: BPF prog-id=147 op=LOAD Jan 24 11:50:28.221000 audit[3219]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=3006 pid=3219 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:28.221000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961613564353963376164343466343030326131326434303263326464 Jan 24 11:50:28.221000 audit: BPF prog-id=147 op=UNLOAD Jan 24 11:50:28.221000 audit[3219]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3006 pid=3219 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:28.221000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961613564353963376164343466343030326131326434303263326464 Jan 24 11:50:28.221000 audit: BPF prog-id=146 op=UNLOAD Jan 24 11:50:28.221000 audit[3219]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3006 pid=3219 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:28.221000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961613564353963376164343466343030326131326434303263326464 Jan 24 11:50:28.221000 audit: BPF prog-id=148 op=LOAD Jan 24 11:50:28.221000 audit[3219]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=3006 pid=3219 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:28.221000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3961613564353963376164343466343030326131326434303263326464 Jan 24 11:50:28.271244 containerd[1647]: time="2026-01-24T11:50:28.271045216Z" level=info msg="StartContainer for \"9aa5d59c7ad44f4002a12d402c2dda1b69ae1ac08f318da787aaffc1e16db2de\" returns successfully" Jan 24 11:50:28.491348 kubelet[2895]: I0124 11:50:28.486643 2895 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-fvhxl" podStartSLOduration=5.486617664 podStartE2EDuration="5.486617664s" podCreationTimestamp="2026-01-24 11:50:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 11:50:25.367777379 +0000 UTC m=+4.889570106" watchObservedRunningTime="2026-01-24 11:50:28.486617664 +0000 UTC m=+8.008410370" Jan 24 11:50:28.491348 kubelet[2895]: I0124 11:50:28.487051 2895 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7dcd859c48-9rh4s" podStartSLOduration=2.336518536 podStartE2EDuration="5.487037626s" podCreationTimestamp="2026-01-24 11:50:23 +0000 UTC" firstStartedPulling="2026-01-24 11:50:24.929577548 +0000 UTC m=+4.451370264" lastFinishedPulling="2026-01-24 11:50:28.080096658 +0000 UTC m=+7.601889354" observedRunningTime="2026-01-24 11:50:28.484194982 +0000 UTC m=+8.005987698" watchObservedRunningTime="2026-01-24 11:50:28.487037626 +0000 UTC m=+8.008830332" Jan 24 11:50:29.757053 kubelet[2895]: E0124 11:50:29.756848 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:30.432627 kubelet[2895]: E0124 11:50:30.432216 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:30.940012 kubelet[2895]: E0124 11:50:30.939051 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:31.254795 kubelet[2895]: E0124 11:50:31.253325 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:31.433910 kubelet[2895]: E0124 11:50:31.433585 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:31.434362 kubelet[2895]: E0124 11:50:31.434186 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:32.526659 systemd[1]: cri-containerd-9aa5d59c7ad44f4002a12d402c2dda1b69ae1ac08f318da787aaffc1e16db2de.scope: Deactivated successfully. Jan 24 11:50:32.534655 containerd[1647]: time="2026-01-24T11:50:32.534608064Z" level=info msg="received container exit event container_id:\"9aa5d59c7ad44f4002a12d402c2dda1b69ae1ac08f318da787aaffc1e16db2de\" id:\"9aa5d59c7ad44f4002a12d402c2dda1b69ae1ac08f318da787aaffc1e16db2de\" pid:3233 exit_status:1 exited_at:{seconds:1769255432 nanos:531474267}" Jan 24 11:50:32.541166 kernel: kauditd_printk_skb: 165 callbacks suppressed Jan 24 11:50:32.541259 kernel: audit: type=1334 audit(1769255432.533:510): prog-id=144 op=UNLOAD Jan 24 11:50:32.533000 audit: BPF prog-id=144 op=UNLOAD Jan 24 11:50:32.533000 audit: BPF prog-id=148 op=UNLOAD Jan 24 11:50:32.559163 kernel: audit: type=1334 audit(1769255432.533:511): prog-id=148 op=UNLOAD Jan 24 11:50:32.712430 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9aa5d59c7ad44f4002a12d402c2dda1b69ae1ac08f318da787aaffc1e16db2de-rootfs.mount: Deactivated successfully. Jan 24 11:50:33.502771 kubelet[2895]: I0124 11:50:33.500029 2895 scope.go:117] "RemoveContainer" containerID="9aa5d59c7ad44f4002a12d402c2dda1b69ae1ac08f318da787aaffc1e16db2de" Jan 24 11:50:33.510561 containerd[1647]: time="2026-01-24T11:50:33.510495057Z" level=info msg="CreateContainer within sandbox \"b6e34f1466bd97b7aca805874a33df28311926530eda9465360444b60d41df9d\" for container &ContainerMetadata{Name:tigera-operator,Attempt:1,}" Jan 24 11:50:33.570477 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2559830940.mount: Deactivated successfully. Jan 24 11:50:33.582117 containerd[1647]: time="2026-01-24T11:50:33.581739207Z" level=info msg="Container e4098641bbcd67eea0764ac500f3768a5d0a74492b08fd1ff054fcef59d4ba21: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:50:33.628545 containerd[1647]: time="2026-01-24T11:50:33.628078014Z" level=info msg="CreateContainer within sandbox \"b6e34f1466bd97b7aca805874a33df28311926530eda9465360444b60d41df9d\" for &ContainerMetadata{Name:tigera-operator,Attempt:1,} returns container id \"e4098641bbcd67eea0764ac500f3768a5d0a74492b08fd1ff054fcef59d4ba21\"" Jan 24 11:50:33.648375 containerd[1647]: time="2026-01-24T11:50:33.644806094Z" level=info msg="StartContainer for \"e4098641bbcd67eea0764ac500f3768a5d0a74492b08fd1ff054fcef59d4ba21\"" Jan 24 11:50:33.648375 containerd[1647]: time="2026-01-24T11:50:33.647098525Z" level=info msg="connecting to shim e4098641bbcd67eea0764ac500f3768a5d0a74492b08fd1ff054fcef59d4ba21" address="unix:///run/containerd/s/e55f705d53dda65ce312b92ba2107ed2765c020c727089568e0ecae6d373ed99" protocol=ttrpc version=3 Jan 24 11:50:33.737887 systemd[1]: Started cri-containerd-e4098641bbcd67eea0764ac500f3768a5d0a74492b08fd1ff054fcef59d4ba21.scope - libcontainer container e4098641bbcd67eea0764ac500f3768a5d0a74492b08fd1ff054fcef59d4ba21. Jan 24 11:50:33.787000 audit: BPF prog-id=149 op=LOAD Jan 24 11:50:33.792356 kernel: audit: type=1334 audit(1769255433.787:512): prog-id=149 op=LOAD Jan 24 11:50:33.790000 audit: BPF prog-id=150 op=LOAD Jan 24 11:50:33.797469 kernel: audit: type=1334 audit(1769255433.790:513): prog-id=150 op=LOAD Jan 24 11:50:33.825909 kernel: audit: type=1300 audit(1769255433.790:513): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8238 a2=98 a3=0 items=0 ppid=3006 pid=3281 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:33.790000 audit[3281]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8238 a2=98 a3=0 items=0 ppid=3006 pid=3281 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:33.790000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6534303938363431626263643637656561303736346163353030663337 Jan 24 11:50:33.845868 kernel: audit: type=1327 audit(1769255433.790:513): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6534303938363431626263643637656561303736346163353030663337 Jan 24 11:50:33.790000 audit: BPF prog-id=150 op=UNLOAD Jan 24 11:50:33.890974 kernel: audit: type=1334 audit(1769255433.790:514): prog-id=150 op=UNLOAD Jan 24 11:50:33.912494 kernel: audit: type=1300 audit(1769255433.790:514): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3006 pid=3281 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:33.790000 audit[3281]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3006 pid=3281 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:33.790000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6534303938363431626263643637656561303736346163353030663337 Jan 24 11:50:33.927568 kernel: audit: type=1327 audit(1769255433.790:514): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6534303938363431626263643637656561303736346163353030663337 Jan 24 11:50:33.791000 audit: BPF prog-id=151 op=LOAD Jan 24 11:50:33.791000 audit[3281]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8488 a2=98 a3=0 items=0 ppid=3006 pid=3281 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:33.791000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6534303938363431626263643637656561303736346163353030663337 Jan 24 11:50:33.791000 audit: BPF prog-id=152 op=LOAD Jan 24 11:50:33.791000 audit[3281]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a8218 a2=98 a3=0 items=0 ppid=3006 pid=3281 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:33.791000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6534303938363431626263643637656561303736346163353030663337 Jan 24 11:50:33.791000 audit: BPF prog-id=152 op=UNLOAD Jan 24 11:50:33.791000 audit[3281]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3006 pid=3281 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:33.791000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6534303938363431626263643637656561303736346163353030663337 Jan 24 11:50:33.791000 audit: BPF prog-id=151 op=UNLOAD Jan 24 11:50:33.791000 audit[3281]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3006 pid=3281 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:33.791000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6534303938363431626263643637656561303736346163353030663337 Jan 24 11:50:33.791000 audit: BPF prog-id=153 op=LOAD Jan 24 11:50:33.791000 audit[3281]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a86e8 a2=98 a3=0 items=0 ppid=3006 pid=3281 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:33.791000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6534303938363431626263643637656561303736346163353030663337 Jan 24 11:50:33.944235 kernel: audit: type=1334 audit(1769255433.791:515): prog-id=151 op=LOAD Jan 24 11:50:34.115688 containerd[1647]: time="2026-01-24T11:50:34.110240998Z" level=info msg="StartContainer for \"e4098641bbcd67eea0764ac500f3768a5d0a74492b08fd1ff054fcef59d4ba21\" returns successfully" Jan 24 11:50:35.729578 sudo[1856]: pam_unix(sudo:session): session closed for user root Jan 24 11:50:35.728000 audit[1856]: USER_END pid=1856 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:50:35.728000 audit[1856]: CRED_DISP pid=1856 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 24 11:50:35.738234 sshd[1855]: Connection closed by 10.0.0.1 port 51262 Jan 24 11:50:35.738037 sshd-session[1851]: pam_unix(sshd:session): session closed for user core Jan 24 11:50:35.748000 audit[1851]: USER_END pid=1851 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:50:35.748000 audit[1851]: CRED_DISP pid=1851 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:50:35.755000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-10.0.0.67:22-10.0.0.1:51262 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:50:35.756133 systemd[1]: sshd@6-10.0.0.67:22-10.0.0.1:51262.service: Deactivated successfully. Jan 24 11:50:35.761983 systemd[1]: session-8.scope: Deactivated successfully. Jan 24 11:50:35.762743 systemd[1]: session-8.scope: Consumed 14.902s CPU time, 213.4M memory peak. Jan 24 11:50:35.773723 systemd-logind[1621]: Session 8 logged out. Waiting for processes to exit. Jan 24 11:50:35.778157 systemd-logind[1621]: Removed session 8. Jan 24 11:50:40.877020 kernel: kauditd_printk_skb: 19 callbacks suppressed Jan 24 11:50:40.877480 kernel: audit: type=1325 audit(1769255440.865:525): table=filter:105 family=2 entries=15 op=nft_register_rule pid=3355 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:40.865000 audit[3355]: NETFILTER_CFG table=filter:105 family=2 entries=15 op=nft_register_rule pid=3355 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:40.865000 audit[3355]: SYSCALL arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7ffd5b80a510 a2=0 a3=7ffd5b80a4fc items=0 ppid=3040 pid=3355 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:40.905349 kernel: audit: type=1300 audit(1769255440.865:525): arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7ffd5b80a510 a2=0 a3=7ffd5b80a4fc items=0 ppid=3040 pid=3355 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:40.865000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:40.923640 kernel: audit: type=1327 audit(1769255440.865:525): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:40.923757 kernel: audit: type=1325 audit(1769255440.905:526): table=nat:106 family=2 entries=12 op=nft_register_rule pid=3355 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:40.905000 audit[3355]: NETFILTER_CFG table=nat:106 family=2 entries=12 op=nft_register_rule pid=3355 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:40.905000 audit[3355]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd5b80a510 a2=0 a3=0 items=0 ppid=3040 pid=3355 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:40.945088 kernel: audit: type=1300 audit(1769255440.905:526): arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd5b80a510 a2=0 a3=0 items=0 ppid=3040 pid=3355 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:40.945174 kernel: audit: type=1327 audit(1769255440.905:526): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:40.905000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:40.955000 audit[3357]: NETFILTER_CFG table=filter:107 family=2 entries=16 op=nft_register_rule pid=3357 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:40.985530 kernel: audit: type=1325 audit(1769255440.955:527): table=filter:107 family=2 entries=16 op=nft_register_rule pid=3357 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:40.985653 kernel: audit: type=1300 audit(1769255440.955:527): arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7ffc9d2ebf00 a2=0 a3=7ffc9d2ebeec items=0 ppid=3040 pid=3357 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:40.955000 audit[3357]: SYSCALL arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7ffc9d2ebf00 a2=0 a3=7ffc9d2ebeec items=0 ppid=3040 pid=3357 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:41.002216 kernel: audit: type=1327 audit(1769255440.955:527): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:40.955000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:41.009000 audit[3357]: NETFILTER_CFG table=nat:108 family=2 entries=12 op=nft_register_rule pid=3357 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:41.009000 audit[3357]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffc9d2ebf00 a2=0 a3=0 items=0 ppid=3040 pid=3357 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:41.009000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:41.019359 kernel: audit: type=1325 audit(1769255441.009:528): table=nat:108 family=2 entries=12 op=nft_register_rule pid=3357 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:47.512000 audit[3359]: NETFILTER_CFG table=filter:109 family=2 entries=16 op=nft_register_rule pid=3359 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:47.517387 kernel: kauditd_printk_skb: 2 callbacks suppressed Jan 24 11:50:47.517491 kernel: audit: type=1325 audit(1769255447.512:529): table=filter:109 family=2 entries=16 op=nft_register_rule pid=3359 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:47.512000 audit[3359]: SYSCALL arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7ffed24667a0 a2=0 a3=7ffed246678c items=0 ppid=3040 pid=3359 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:47.546575 kernel: audit: type=1300 audit(1769255447.512:529): arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7ffed24667a0 a2=0 a3=7ffed246678c items=0 ppid=3040 pid=3359 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:47.571050 kernel: audit: type=1327 audit(1769255447.512:529): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:47.573122 kernel: audit: type=1325 audit(1769255447.535:530): table=nat:110 family=2 entries=12 op=nft_register_rule pid=3359 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:47.512000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:47.535000 audit[3359]: NETFILTER_CFG table=nat:110 family=2 entries=12 op=nft_register_rule pid=3359 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:47.535000 audit[3359]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffed24667a0 a2=0 a3=0 items=0 ppid=3040 pid=3359 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:47.599583 kernel: audit: type=1300 audit(1769255447.535:530): arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffed24667a0 a2=0 a3=0 items=0 ppid=3040 pid=3359 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:47.599720 kernel: audit: type=1327 audit(1769255447.535:530): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:47.535000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:47.612000 audit[3361]: NETFILTER_CFG table=filter:111 family=2 entries=17 op=nft_register_rule pid=3361 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:47.612000 audit[3361]: SYSCALL arch=c000003e syscall=46 success=yes exit=6736 a0=3 a1=7ffd89343dd0 a2=0 a3=7ffd89343dbc items=0 ppid=3040 pid=3361 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:47.646372 kernel: audit: type=1325 audit(1769255447.612:531): table=filter:111 family=2 entries=17 op=nft_register_rule pid=3361 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:47.646547 kernel: audit: type=1300 audit(1769255447.612:531): arch=c000003e syscall=46 success=yes exit=6736 a0=3 a1=7ffd89343dd0 a2=0 a3=7ffd89343dbc items=0 ppid=3040 pid=3361 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:47.612000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:47.657000 audit[3361]: NETFILTER_CFG table=nat:112 family=2 entries=12 op=nft_register_rule pid=3361 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:47.673081 kernel: audit: type=1327 audit(1769255447.612:531): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:47.673206 kernel: audit: type=1325 audit(1769255447.657:532): table=nat:112 family=2 entries=12 op=nft_register_rule pid=3361 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:47.657000 audit[3361]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd89343dd0 a2=0 a3=0 items=0 ppid=3040 pid=3361 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:47.657000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:48.728000 audit[3363]: NETFILTER_CFG table=filter:113 family=2 entries=19 op=nft_register_rule pid=3363 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:48.728000 audit[3363]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7fffab4ad070 a2=0 a3=7fffab4ad05c items=0 ppid=3040 pid=3363 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:48.728000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:48.742000 audit[3363]: NETFILTER_CFG table=nat:114 family=2 entries=12 op=nft_register_rule pid=3363 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:48.742000 audit[3363]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7fffab4ad070 a2=0 a3=0 items=0 ppid=3040 pid=3363 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:48.742000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:52.325000 audit[3365]: NETFILTER_CFG table=filter:115 family=2 entries=21 op=nft_register_rule pid=3365 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:52.325000 audit[3365]: SYSCALL arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7ffd0b690fe0 a2=0 a3=7ffd0b690fcc items=0 ppid=3040 pid=3365 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:52.325000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:52.334000 audit[3365]: NETFILTER_CFG table=nat:116 family=2 entries=12 op=nft_register_rule pid=3365 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:52.334000 audit[3365]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd0b690fe0 a2=0 a3=0 items=0 ppid=3040 pid=3365 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:52.334000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:52.411000 audit[3367]: NETFILTER_CFG table=filter:117 family=2 entries=22 op=nft_register_rule pid=3367 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:52.411000 audit[3367]: SYSCALL arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7ffc6c97ab10 a2=0 a3=7ffc6c97aafc items=0 ppid=3040 pid=3367 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:52.411000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:52.430000 audit[3367]: NETFILTER_CFG table=nat:118 family=2 entries=12 op=nft_register_rule pid=3367 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:52.430000 audit[3367]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffc6c97ab10 a2=0 a3=0 items=0 ppid=3040 pid=3367 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:52.430000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:52.490105 systemd[1]: Created slice kubepods-besteffort-poddfb7b235_7402_4077_b56a_426749d4701b.slice - libcontainer container kubepods-besteffort-poddfb7b235_7402_4077_b56a_426749d4701b.slice. Jan 24 11:50:52.567742 kubelet[2895]: I0124 11:50:52.550203 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/dfb7b235-7402-4077-b56a-426749d4701b-typha-certs\") pod \"calico-typha-dfdc7d98b-tn6mr\" (UID: \"dfb7b235-7402-4077-b56a-426749d4701b\") " pod="calico-system/calico-typha-dfdc7d98b-tn6mr" Jan 24 11:50:52.567742 kubelet[2895]: I0124 11:50:52.552409 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c97t8\" (UniqueName: \"kubernetes.io/projected/dfb7b235-7402-4077-b56a-426749d4701b-kube-api-access-c97t8\") pod \"calico-typha-dfdc7d98b-tn6mr\" (UID: \"dfb7b235-7402-4077-b56a-426749d4701b\") " pod="calico-system/calico-typha-dfdc7d98b-tn6mr" Jan 24 11:50:52.567742 kubelet[2895]: I0124 11:50:52.552508 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dfb7b235-7402-4077-b56a-426749d4701b-tigera-ca-bundle\") pod \"calico-typha-dfdc7d98b-tn6mr\" (UID: \"dfb7b235-7402-4077-b56a-426749d4701b\") " pod="calico-system/calico-typha-dfdc7d98b-tn6mr" Jan 24 11:50:52.820065 kubelet[2895]: E0124 11:50:52.819619 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:52.829669 containerd[1647]: time="2026-01-24T11:50:52.827201653Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-dfdc7d98b-tn6mr,Uid:dfb7b235-7402-4077-b56a-426749d4701b,Namespace:calico-system,Attempt:0,}" Jan 24 11:50:53.059971 kubelet[2895]: I0124 11:50:53.058978 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/47c115f3-217f-4dd0-9700-791552d056e3-cni-bin-dir\") pod \"calico-node-878v5\" (UID: \"47c115f3-217f-4dd0-9700-791552d056e3\") " pod="calico-system/calico-node-878v5" Jan 24 11:50:53.063597 kubelet[2895]: I0124 11:50:53.061374 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/47c115f3-217f-4dd0-9700-791552d056e3-cni-net-dir\") pod \"calico-node-878v5\" (UID: \"47c115f3-217f-4dd0-9700-791552d056e3\") " pod="calico-system/calico-node-878v5" Jan 24 11:50:53.063597 kubelet[2895]: I0124 11:50:53.063581 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/47c115f3-217f-4dd0-9700-791552d056e3-lib-modules\") pod \"calico-node-878v5\" (UID: \"47c115f3-217f-4dd0-9700-791552d056e3\") " pod="calico-system/calico-node-878v5" Jan 24 11:50:53.063597 kubelet[2895]: I0124 11:50:53.063607 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/47c115f3-217f-4dd0-9700-791552d056e3-node-certs\") pod \"calico-node-878v5\" (UID: \"47c115f3-217f-4dd0-9700-791552d056e3\") " pod="calico-system/calico-node-878v5" Jan 24 11:50:53.063852 kubelet[2895]: I0124 11:50:53.063632 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/47c115f3-217f-4dd0-9700-791552d056e3-cni-log-dir\") pod \"calico-node-878v5\" (UID: \"47c115f3-217f-4dd0-9700-791552d056e3\") " pod="calico-system/calico-node-878v5" Jan 24 11:50:53.063852 kubelet[2895]: I0124 11:50:53.063655 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/47c115f3-217f-4dd0-9700-791552d056e3-policysync\") pod \"calico-node-878v5\" (UID: \"47c115f3-217f-4dd0-9700-791552d056e3\") " pod="calico-system/calico-node-878v5" Jan 24 11:50:53.063852 kubelet[2895]: I0124 11:50:53.063675 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47c115f3-217f-4dd0-9700-791552d056e3-tigera-ca-bundle\") pod \"calico-node-878v5\" (UID: \"47c115f3-217f-4dd0-9700-791552d056e3\") " pod="calico-system/calico-node-878v5" Jan 24 11:50:53.063852 kubelet[2895]: I0124 11:50:53.063694 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/47c115f3-217f-4dd0-9700-791552d056e3-var-run-calico\") pod \"calico-node-878v5\" (UID: \"47c115f3-217f-4dd0-9700-791552d056e3\") " pod="calico-system/calico-node-878v5" Jan 24 11:50:53.063852 kubelet[2895]: I0124 11:50:53.063716 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrq7l\" (UniqueName: \"kubernetes.io/projected/47c115f3-217f-4dd0-9700-791552d056e3-kube-api-access-mrq7l\") pod \"calico-node-878v5\" (UID: \"47c115f3-217f-4dd0-9700-791552d056e3\") " pod="calico-system/calico-node-878v5" Jan 24 11:50:53.064041 kubelet[2895]: I0124 11:50:53.063742 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/47c115f3-217f-4dd0-9700-791552d056e3-xtables-lock\") pod \"calico-node-878v5\" (UID: \"47c115f3-217f-4dd0-9700-791552d056e3\") " pod="calico-system/calico-node-878v5" Jan 24 11:50:53.064041 kubelet[2895]: I0124 11:50:53.063781 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/47c115f3-217f-4dd0-9700-791552d056e3-flexvol-driver-host\") pod \"calico-node-878v5\" (UID: \"47c115f3-217f-4dd0-9700-791552d056e3\") " pod="calico-system/calico-node-878v5" Jan 24 11:50:53.064041 kubelet[2895]: I0124 11:50:53.063803 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/47c115f3-217f-4dd0-9700-791552d056e3-var-lib-calico\") pod \"calico-node-878v5\" (UID: \"47c115f3-217f-4dd0-9700-791552d056e3\") " pod="calico-system/calico-node-878v5" Jan 24 11:50:53.064077 systemd[1]: Created slice kubepods-besteffort-pod47c115f3_217f_4dd0_9700_791552d056e3.slice - libcontainer container kubepods-besteffort-pod47c115f3_217f_4dd0_9700_791552d056e3.slice. Jan 24 11:50:53.181729 containerd[1647]: time="2026-01-24T11:50:53.179950597Z" level=info msg="connecting to shim 6e16b6d6761d0bd70ac87a5f62b4eff051dd2f16dab637fbd89535c45b7e5523" address="unix:///run/containerd/s/f85dd62405222e764ba5c06cbe1c5a27bf66ae66e377fe4ceac96b2710e96c59" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:50:53.237539 kubelet[2895]: E0124 11:50:53.234116 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.237539 kubelet[2895]: W0124 11:50:53.234154 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.237539 kubelet[2895]: E0124 11:50:53.234223 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.484731 kubelet[2895]: E0124 11:50:53.481860 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:50:53.484138 systemd[1]: Started cri-containerd-6e16b6d6761d0bd70ac87a5f62b4eff051dd2f16dab637fbd89535c45b7e5523.scope - libcontainer container 6e16b6d6761d0bd70ac87a5f62b4eff051dd2f16dab637fbd89535c45b7e5523. Jan 24 11:50:53.495820 kubelet[2895]: E0124 11:50:53.494756 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.495820 kubelet[2895]: W0124 11:50:53.494803 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.495820 kubelet[2895]: E0124 11:50:53.494836 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.577183 kubelet[2895]: E0124 11:50:53.577004 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.577818 kubelet[2895]: W0124 11:50:53.577221 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.577818 kubelet[2895]: E0124 11:50:53.577374 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.599066 kubelet[2895]: E0124 11:50:53.597060 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.599066 kubelet[2895]: W0124 11:50:53.598197 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.599066 kubelet[2895]: E0124 11:50:53.599000 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.599066 kubelet[2895]: E0124 11:50:53.600870 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.599066 kubelet[2895]: W0124 11:50:53.600886 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.599066 kubelet[2895]: E0124 11:50:53.600909 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.599066 kubelet[2895]: E0124 11:50:53.601815 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.599066 kubelet[2895]: W0124 11:50:53.601830 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.599066 kubelet[2895]: E0124 11:50:53.601846 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.599066 kubelet[2895]: E0124 11:50:53.602565 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.669974 kubelet[2895]: W0124 11:50:53.602580 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.669974 kubelet[2895]: E0124 11:50:53.602598 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.669974 kubelet[2895]: E0124 11:50:53.602866 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.669974 kubelet[2895]: W0124 11:50:53.602917 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.669974 kubelet[2895]: E0124 11:50:53.602938 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.696918 kubelet[2895]: E0124 11:50:53.695875 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.696918 kubelet[2895]: W0124 11:50:53.695969 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.696918 kubelet[2895]: E0124 11:50:53.696032 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.720617 kubelet[2895]: E0124 11:50:53.720322 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.720617 kubelet[2895]: W0124 11:50:53.720362 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.720617 kubelet[2895]: E0124 11:50:53.720394 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.722395 kubelet[2895]: E0124 11:50:53.721635 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.723789 kubelet[2895]: W0124 11:50:53.722937 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.725103 kubelet[2895]: E0124 11:50:53.724941 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.726117 kubelet[2895]: E0124 11:50:53.725788 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.726117 kubelet[2895]: W0124 11:50:53.725899 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.726117 kubelet[2895]: E0124 11:50:53.725919 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.727728 kubelet[2895]: E0124 11:50:53.726253 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.727728 kubelet[2895]: W0124 11:50:53.726518 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.727728 kubelet[2895]: E0124 11:50:53.726543 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.732673 kubelet[2895]: E0124 11:50:53.728944 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.732673 kubelet[2895]: W0124 11:50:53.728960 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.732673 kubelet[2895]: E0124 11:50:53.728974 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.732673 kubelet[2895]: E0124 11:50:53.729335 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:53.737350 kubelet[2895]: E0124 11:50:53.736177 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.737350 kubelet[2895]: W0124 11:50:53.736360 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.737350 kubelet[2895]: E0124 11:50:53.736385 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.746227 containerd[1647]: time="2026-01-24T11:50:53.742119044Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-878v5,Uid:47c115f3-217f-4dd0-9700-791552d056e3,Namespace:calico-system,Attempt:0,}" Jan 24 11:50:53.750415 kubelet[2895]: E0124 11:50:53.747788 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.750935 kubelet[2895]: W0124 11:50:53.750546 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.751500 kubelet[2895]: E0124 11:50:53.751233 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.754547 kubelet[2895]: E0124 11:50:53.754527 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.754898 kubelet[2895]: W0124 11:50:53.754772 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.755633 kubelet[2895]: E0124 11:50:53.754795 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.760483 kubelet[2895]: E0124 11:50:53.760460 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.761134 kubelet[2895]: W0124 11:50:53.761107 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.761710 kubelet[2895]: E0124 11:50:53.761684 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.763125 kubelet[2895]: E0124 11:50:53.763027 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.763125 kubelet[2895]: W0124 11:50:53.763047 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.764014 kubelet[2895]: E0124 11:50:53.763990 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.765332 kubelet[2895]: E0124 11:50:53.765162 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.765332 kubelet[2895]: W0124 11:50:53.765179 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.765332 kubelet[2895]: E0124 11:50:53.765198 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.765878 kubelet[2895]: E0124 11:50:53.765859 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.766049 kubelet[2895]: W0124 11:50:53.766019 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.766154 kubelet[2895]: E0124 11:50:53.766135 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.766922 kubelet[2895]: E0124 11:50:53.766862 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.766922 kubelet[2895]: W0124 11:50:53.766877 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.766922 kubelet[2895]: E0124 11:50:53.766893 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.766000 audit: BPF prog-id=154 op=LOAD Jan 24 11:50:53.769871 kubelet[2895]: E0124 11:50:53.769773 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.769871 kubelet[2895]: W0124 11:50:53.769791 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.770167 kubelet[2895]: E0124 11:50:53.770005 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.773819 kubelet[2895]: I0124 11:50:53.773669 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqn89\" (UniqueName: \"kubernetes.io/projected/999c48c5-f3b1-4bfd-91f4-26cafee1eef0-kube-api-access-lqn89\") pod \"csi-node-driver-2g45n\" (UID: \"999c48c5-f3b1-4bfd-91f4-26cafee1eef0\") " pod="calico-system/csi-node-driver-2g45n" Jan 24 11:50:53.776618 kernel: kauditd_printk_skb: 20 callbacks suppressed Jan 24 11:50:53.776716 kernel: audit: type=1334 audit(1769255453.766:539): prog-id=154 op=LOAD Jan 24 11:50:53.776942 kubelet[2895]: E0124 11:50:53.776882 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.776942 kubelet[2895]: W0124 11:50:53.776904 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.776942 kubelet[2895]: E0124 11:50:53.776922 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.779375 kubelet[2895]: E0124 11:50:53.779356 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.779661 kubelet[2895]: W0124 11:50:53.779488 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.779661 kubelet[2895]: E0124 11:50:53.779512 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.779991 kubelet[2895]: E0124 11:50:53.779975 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.780122 kubelet[2895]: W0124 11:50:53.780103 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.780337 kubelet[2895]: E0124 11:50:53.780247 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.780695 kubelet[2895]: I0124 11:50:53.780670 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/999c48c5-f3b1-4bfd-91f4-26cafee1eef0-varrun\") pod \"csi-node-driver-2g45n\" (UID: \"999c48c5-f3b1-4bfd-91f4-26cafee1eef0\") " pod="calico-system/csi-node-driver-2g45n" Jan 24 11:50:53.781233 kubelet[2895]: E0124 11:50:53.781215 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.781677 kubelet[2895]: W0124 11:50:53.781389 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.781677 kubelet[2895]: E0124 11:50:53.781408 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.782768 kubelet[2895]: E0124 11:50:53.782525 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.782768 kubelet[2895]: W0124 11:50:53.782544 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.782768 kubelet[2895]: E0124 11:50:53.782558 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.784237 kubelet[2895]: E0124 11:50:53.784142 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.784237 kubelet[2895]: W0124 11:50:53.784159 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.784687 kubelet[2895]: E0124 11:50:53.784523 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.784961 kubelet[2895]: I0124 11:50:53.784900 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/999c48c5-f3b1-4bfd-91f4-26cafee1eef0-kubelet-dir\") pod \"csi-node-driver-2g45n\" (UID: \"999c48c5-f3b1-4bfd-91f4-26cafee1eef0\") " pod="calico-system/csi-node-driver-2g45n" Jan 24 11:50:53.785613 kubelet[2895]: E0124 11:50:53.785415 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.785613 kubelet[2895]: W0124 11:50:53.785521 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.785613 kubelet[2895]: E0124 11:50:53.785554 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.784000 audit: BPF prog-id=155 op=LOAD Jan 24 11:50:53.796688 kubelet[2895]: E0124 11:50:53.791486 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.796688 kubelet[2895]: W0124 11:50:53.791517 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.796688 kubelet[2895]: E0124 11:50:53.791548 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.796688 kubelet[2895]: E0124 11:50:53.791919 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.796688 kubelet[2895]: W0124 11:50:53.791934 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.796688 kubelet[2895]: E0124 11:50:53.791952 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.796688 kubelet[2895]: I0124 11:50:53.792358 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/999c48c5-f3b1-4bfd-91f4-26cafee1eef0-registration-dir\") pod \"csi-node-driver-2g45n\" (UID: \"999c48c5-f3b1-4bfd-91f4-26cafee1eef0\") " pod="calico-system/csi-node-driver-2g45n" Jan 24 11:50:53.796688 kubelet[2895]: E0124 11:50:53.793339 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.796688 kubelet[2895]: W0124 11:50:53.793359 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.798007 kubelet[2895]: E0124 11:50:53.793377 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.798007 kubelet[2895]: E0124 11:50:53.793860 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.798007 kubelet[2895]: W0124 11:50:53.793873 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.798007 kubelet[2895]: E0124 11:50:53.793892 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.798007 kubelet[2895]: E0124 11:50:53.794726 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.798007 kubelet[2895]: W0124 11:50:53.794741 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.798007 kubelet[2895]: E0124 11:50:53.794758 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.798007 kubelet[2895]: I0124 11:50:53.794796 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/999c48c5-f3b1-4bfd-91f4-26cafee1eef0-socket-dir\") pod \"csi-node-driver-2g45n\" (UID: \"999c48c5-f3b1-4bfd-91f4-26cafee1eef0\") " pod="calico-system/csi-node-driver-2g45n" Jan 24 11:50:53.798007 kubelet[2895]: E0124 11:50:53.795137 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.798865 kubelet[2895]: W0124 11:50:53.795155 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.798865 kubelet[2895]: E0124 11:50:53.795172 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.798865 kubelet[2895]: E0124 11:50:53.795567 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.798865 kubelet[2895]: W0124 11:50:53.795579 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.798865 kubelet[2895]: E0124 11:50:53.795592 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.804689 kernel: audit: type=1334 audit(1769255453.784:540): prog-id=155 op=LOAD Jan 24 11:50:53.804792 kernel: audit: type=1300 audit(1769255453.784:540): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=3379 pid=3392 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:53.784000 audit[3392]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=3379 pid=3392 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:53.784000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665313662366436373631643062643730616338376135663632623465 Jan 24 11:50:53.845805 kernel: audit: type=1327 audit(1769255453.784:540): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665313662366436373631643062643730616338376135663632623465 Jan 24 11:50:53.849664 kernel: audit: type=1334 audit(1769255453.786:541): prog-id=155 op=UNLOAD Jan 24 11:50:53.786000 audit: BPF prog-id=155 op=UNLOAD Jan 24 11:50:53.866192 kernel: audit: type=1300 audit(1769255453.786:541): arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3379 pid=3392 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:53.786000 audit[3392]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3379 pid=3392 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:53.786000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665313662366436373631643062643730616338376135663632623465 Jan 24 11:50:53.891355 kernel: audit: type=1327 audit(1769255453.786:541): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665313662366436373631643062643730616338376135663632623465 Jan 24 11:50:53.786000 audit: BPF prog-id=156 op=LOAD Jan 24 11:50:53.895315 kernel: audit: type=1334 audit(1769255453.786:542): prog-id=156 op=LOAD Jan 24 11:50:53.786000 audit[3392]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3379 pid=3392 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:53.925161 kubelet[2895]: E0124 11:50:53.915885 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.925161 kubelet[2895]: W0124 11:50:53.915992 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.925161 kubelet[2895]: E0124 11:50:53.916027 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.925161 kubelet[2895]: E0124 11:50:53.917717 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.925161 kubelet[2895]: W0124 11:50:53.917846 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.925161 kubelet[2895]: E0124 11:50:53.923989 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.926779 kernel: audit: type=1300 audit(1769255453.786:542): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3379 pid=3392 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:53.786000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665313662366436373631643062643730616338376135663632623465 Jan 24 11:50:53.943811 kubelet[2895]: E0124 11:50:53.934815 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.943811 kubelet[2895]: W0124 11:50:53.934920 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.943811 kubelet[2895]: E0124 11:50:53.935035 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.943811 kubelet[2895]: E0124 11:50:53.939849 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.943811 kubelet[2895]: W0124 11:50:53.939980 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.943811 kubelet[2895]: E0124 11:50:53.940181 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.786000 audit: BPF prog-id=157 op=LOAD Jan 24 11:50:53.786000 audit[3392]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3379 pid=3392 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:53.944494 kernel: audit: type=1327 audit(1769255453.786:542): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665313662366436373631643062643730616338376135663632623465 Jan 24 11:50:53.786000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665313662366436373631643062643730616338376135663632623465 Jan 24 11:50:53.786000 audit: BPF prog-id=157 op=UNLOAD Jan 24 11:50:53.786000 audit[3392]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3379 pid=3392 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:53.786000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665313662366436373631643062643730616338376135663632623465 Jan 24 11:50:53.786000 audit: BPF prog-id=156 op=UNLOAD Jan 24 11:50:53.786000 audit[3392]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3379 pid=3392 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:53.786000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665313662366436373631643062643730616338376135663632623465 Jan 24 11:50:53.787000 audit: BPF prog-id=158 op=LOAD Jan 24 11:50:53.787000 audit[3392]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=3379 pid=3392 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:53.787000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665313662366436373631643062643730616338376135663632623465 Jan 24 11:50:53.789000 audit[3426]: NETFILTER_CFG table=filter:119 family=2 entries=22 op=nft_register_rule pid=3426 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:53.789000 audit[3426]: SYSCALL arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7ffc3d280530 a2=0 a3=7ffc3d28051c items=0 ppid=3040 pid=3426 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:53.789000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:53.937000 audit[3426]: NETFILTER_CFG table=nat:120 family=2 entries=12 op=nft_register_rule pid=3426 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:53.937000 audit[3426]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffc3d280530 a2=0 a3=0 items=0 ppid=3040 pid=3426 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:53.937000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:50:53.950116 kubelet[2895]: E0124 11:50:53.949063 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.950116 kubelet[2895]: W0124 11:50:53.949094 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.950116 kubelet[2895]: E0124 11:50:53.949118 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.955138 kubelet[2895]: E0124 11:50:53.955065 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.955138 kubelet[2895]: W0124 11:50:53.955120 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.955393 kubelet[2895]: E0124 11:50:53.955154 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.956716 kubelet[2895]: E0124 11:50:53.956658 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.956837 kubelet[2895]: W0124 11:50:53.956782 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.956957 kubelet[2895]: E0124 11:50:53.956902 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.960692 kubelet[2895]: E0124 11:50:53.960632 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.960692 kubelet[2895]: W0124 11:50:53.960678 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.960806 kubelet[2895]: E0124 11:50:53.960699 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.962476 kubelet[2895]: E0124 11:50:53.962375 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.962476 kubelet[2895]: W0124 11:50:53.962462 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.962574 kubelet[2895]: E0124 11:50:53.962485 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.965477 kubelet[2895]: E0124 11:50:53.965380 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.965477 kubelet[2895]: W0124 11:50:53.965462 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.965594 kubelet[2895]: E0124 11:50:53.965485 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.970756 kubelet[2895]: E0124 11:50:53.970691 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.970756 kubelet[2895]: W0124 11:50:53.970741 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.971387 kubelet[2895]: E0124 11:50:53.970765 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.971387 kubelet[2895]: E0124 11:50:53.971186 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.971387 kubelet[2895]: W0124 11:50:53.971202 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.971387 kubelet[2895]: E0124 11:50:53.971220 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.972993 kubelet[2895]: E0124 11:50:53.972923 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.972993 kubelet[2895]: W0124 11:50:53.972970 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.972993 kubelet[2895]: E0124 11:50:53.972988 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.973470 kubelet[2895]: E0124 11:50:53.973404 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.973470 kubelet[2895]: W0124 11:50:53.973467 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.973570 kubelet[2895]: E0124 11:50:53.973485 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.974033 kubelet[2895]: E0124 11:50:53.973969 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.974033 kubelet[2895]: W0124 11:50:53.974012 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.974033 kubelet[2895]: E0124 11:50:53.974028 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.974589 containerd[1647]: time="2026-01-24T11:50:53.974382113Z" level=info msg="connecting to shim 2c075fdffb700b0c1172313d3ff66de2f4728237f591f9d6d03d9c3131fb8564" address="unix:///run/containerd/s/5856dbc4464bd7263ae0bc1ab234b2764d81a51e8bf8fc1354a687699a56c458" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:50:53.982368 kubelet[2895]: E0124 11:50:53.974984 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.982368 kubelet[2895]: W0124 11:50:53.974996 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.982368 kubelet[2895]: E0124 11:50:53.975011 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.982368 kubelet[2895]: E0124 11:50:53.981680 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.982368 kubelet[2895]: W0124 11:50:53.981700 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.982368 kubelet[2895]: E0124 11:50:53.981720 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.983165 kubelet[2895]: E0124 11:50:53.982784 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.983165 kubelet[2895]: W0124 11:50:53.982804 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.983165 kubelet[2895]: E0124 11:50:53.982825 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.994357 kubelet[2895]: E0124 11:50:53.983521 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.994357 kubelet[2895]: W0124 11:50:53.983536 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.994357 kubelet[2895]: E0124 11:50:53.983553 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.994357 kubelet[2895]: E0124 11:50:53.984989 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.994357 kubelet[2895]: W0124 11:50:53.985088 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.994357 kubelet[2895]: E0124 11:50:53.985106 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:53.994357 kubelet[2895]: E0124 11:50:53.986133 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:53.994357 kubelet[2895]: W0124 11:50:53.986214 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:53.994357 kubelet[2895]: E0124 11:50:53.986232 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:54.010885 kubelet[2895]: E0124 11:50:54.010583 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:54.010885 kubelet[2895]: W0124 11:50:54.010617 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:54.010885 kubelet[2895]: E0124 11:50:54.010654 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:54.014504 kubelet[2895]: E0124 11:50:54.012949 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:54.014504 kubelet[2895]: W0124 11:50:54.012968 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:54.014504 kubelet[2895]: E0124 11:50:54.012997 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:54.014504 kubelet[2895]: E0124 11:50:54.013873 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:54.014504 kubelet[2895]: W0124 11:50:54.013890 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:54.014504 kubelet[2895]: E0124 11:50:54.013912 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:54.024087 kubelet[2895]: E0124 11:50:54.023166 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:54.024087 kubelet[2895]: W0124 11:50:54.023227 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:54.024087 kubelet[2895]: E0124 11:50:54.023320 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:54.055911 kubelet[2895]: E0124 11:50:54.055734 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:54.055911 kubelet[2895]: W0124 11:50:54.055782 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:54.055911 kubelet[2895]: E0124 11:50:54.055815 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:54.181256 containerd[1647]: time="2026-01-24T11:50:54.180627109Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-dfdc7d98b-tn6mr,Uid:dfb7b235-7402-4077-b56a-426749d4701b,Namespace:calico-system,Attempt:0,} returns sandbox id \"6e16b6d6761d0bd70ac87a5f62b4eff051dd2f16dab637fbd89535c45b7e5523\"" Jan 24 11:50:54.217339 kubelet[2895]: E0124 11:50:54.216853 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:54.218836 containerd[1647]: time="2026-01-24T11:50:54.218513616Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\"" Jan 24 11:50:54.254755 systemd[1]: Started cri-containerd-2c075fdffb700b0c1172313d3ff66de2f4728237f591f9d6d03d9c3131fb8564.scope - libcontainer container 2c075fdffb700b0c1172313d3ff66de2f4728237f591f9d6d03d9c3131fb8564. Jan 24 11:50:54.287000 audit: BPF prog-id=159 op=LOAD Jan 24 11:50:54.288000 audit: BPF prog-id=160 op=LOAD Jan 24 11:50:54.288000 audit[3519]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000186238 a2=98 a3=0 items=0 ppid=3491 pid=3519 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:54.288000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3263303735666466666237303062306331313732333133643366663636 Jan 24 11:50:54.288000 audit: BPF prog-id=160 op=UNLOAD Jan 24 11:50:54.288000 audit[3519]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3491 pid=3519 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:54.288000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3263303735666466666237303062306331313732333133643366663636 Jan 24 11:50:54.288000 audit: BPF prog-id=161 op=LOAD Jan 24 11:50:54.288000 audit[3519]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000186488 a2=98 a3=0 items=0 ppid=3491 pid=3519 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:54.288000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3263303735666466666237303062306331313732333133643366663636 Jan 24 11:50:54.288000 audit: BPF prog-id=162 op=LOAD Jan 24 11:50:54.288000 audit[3519]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000186218 a2=98 a3=0 items=0 ppid=3491 pid=3519 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:54.288000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3263303735666466666237303062306331313732333133643366663636 Jan 24 11:50:54.288000 audit: BPF prog-id=162 op=UNLOAD Jan 24 11:50:54.288000 audit[3519]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3491 pid=3519 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:54.288000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3263303735666466666237303062306331313732333133643366663636 Jan 24 11:50:54.288000 audit: BPF prog-id=161 op=UNLOAD Jan 24 11:50:54.288000 audit[3519]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3491 pid=3519 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:54.288000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3263303735666466666237303062306331313732333133643366663636 Jan 24 11:50:54.288000 audit: BPF prog-id=163 op=LOAD Jan 24 11:50:54.288000 audit[3519]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001866e8 a2=98 a3=0 items=0 ppid=3491 pid=3519 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:54.288000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3263303735666466666237303062306331313732333133643366663636 Jan 24 11:50:54.361341 containerd[1647]: time="2026-01-24T11:50:54.360946929Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-878v5,Uid:47c115f3-217f-4dd0-9700-791552d056e3,Namespace:calico-system,Attempt:0,} returns sandbox id \"2c075fdffb700b0c1172313d3ff66de2f4728237f591f9d6d03d9c3131fb8564\"" Jan 24 11:50:54.363329 kubelet[2895]: E0124 11:50:54.363111 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:55.133040 kubelet[2895]: E0124 11:50:55.131244 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:50:55.433562 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3909140454.mount: Deactivated successfully. Jan 24 11:50:57.168038 kubelet[2895]: E0124 11:50:57.165121 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:50:58.820888 containerd[1647]: time="2026-01-24T11:50:58.820553777Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:50:58.824754 containerd[1647]: time="2026-01-24T11:50:58.824583860Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.4: active requests=0, bytes read=33735893" Jan 24 11:50:58.825465 containerd[1647]: time="2026-01-24T11:50:58.825336072Z" level=info msg="ImageCreate event name:\"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:50:58.832000 containerd[1647]: time="2026-01-24T11:50:58.831914437Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:50:58.832811 containerd[1647]: time="2026-01-24T11:50:58.832739901Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.4\" with image id \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\", size \"35234482\" in 4.614171352s" Jan 24 11:50:58.832811 containerd[1647]: time="2026-01-24T11:50:58.832773313Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\" returns image reference \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\"" Jan 24 11:50:58.836694 containerd[1647]: time="2026-01-24T11:50:58.836553802Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\"" Jan 24 11:50:58.866775 containerd[1647]: time="2026-01-24T11:50:58.866617743Z" level=info msg="CreateContainer within sandbox \"6e16b6d6761d0bd70ac87a5f62b4eff051dd2f16dab637fbd89535c45b7e5523\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Jan 24 11:50:58.890715 containerd[1647]: time="2026-01-24T11:50:58.890593589Z" level=info msg="Container 768958dbe9370839c769b931f1dd2739b3550c64716377bb5f3669c9f6c7c9d8: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:50:58.917139 containerd[1647]: time="2026-01-24T11:50:58.917015270Z" level=info msg="CreateContainer within sandbox \"6e16b6d6761d0bd70ac87a5f62b4eff051dd2f16dab637fbd89535c45b7e5523\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"768958dbe9370839c769b931f1dd2739b3550c64716377bb5f3669c9f6c7c9d8\"" Jan 24 11:50:58.921568 containerd[1647]: time="2026-01-24T11:50:58.919630989Z" level=info msg="StartContainer for \"768958dbe9370839c769b931f1dd2739b3550c64716377bb5f3669c9f6c7c9d8\"" Jan 24 11:50:58.921568 containerd[1647]: time="2026-01-24T11:50:58.921409485Z" level=info msg="connecting to shim 768958dbe9370839c769b931f1dd2739b3550c64716377bb5f3669c9f6c7c9d8" address="unix:///run/containerd/s/f85dd62405222e764ba5c06cbe1c5a27bf66ae66e377fe4ceac96b2710e96c59" protocol=ttrpc version=3 Jan 24 11:50:58.999152 systemd[1]: Started cri-containerd-768958dbe9370839c769b931f1dd2739b3550c64716377bb5f3669c9f6c7c9d8.scope - libcontainer container 768958dbe9370839c769b931f1dd2739b3550c64716377bb5f3669c9f6c7c9d8. Jan 24 11:50:59.043572 kernel: kauditd_printk_skb: 40 callbacks suppressed Jan 24 11:50:59.043716 kernel: audit: type=1334 audit(1769255459.038:557): prog-id=164 op=LOAD Jan 24 11:50:59.038000 audit: BPF prog-id=164 op=LOAD Jan 24 11:50:59.055397 kernel: audit: type=1334 audit(1769255459.040:558): prog-id=165 op=LOAD Jan 24 11:50:59.040000 audit: BPF prog-id=165 op=LOAD Jan 24 11:50:59.040000 audit[3567]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=3379 pid=3567 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:59.040000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736383935386462653933373038333963373639623933316631646432 Jan 24 11:50:59.080726 kernel: audit: type=1300 audit(1769255459.040:558): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=3379 pid=3567 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:59.080924 kernel: audit: type=1327 audit(1769255459.040:558): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736383935386462653933373038333963373639623933316631646432 Jan 24 11:50:59.080989 kernel: audit: type=1334 audit(1769255459.040:559): prog-id=165 op=UNLOAD Jan 24 11:50:59.040000 audit: BPF prog-id=165 op=UNLOAD Jan 24 11:50:59.088985 kernel: audit: type=1300 audit(1769255459.040:559): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3379 pid=3567 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:59.040000 audit[3567]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3379 pid=3567 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:59.040000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736383935386462653933373038333963373639623933316631646432 Jan 24 11:50:59.122136 kernel: audit: type=1327 audit(1769255459.040:559): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736383935386462653933373038333963373639623933316631646432 Jan 24 11:50:59.122365 kernel: audit: type=1334 audit(1769255459.042:560): prog-id=166 op=LOAD Jan 24 11:50:59.042000 audit: BPF prog-id=166 op=LOAD Jan 24 11:50:59.042000 audit[3567]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=3379 pid=3567 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:59.134488 kubelet[2895]: E0124 11:50:59.131250 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:50:59.137331 kernel: audit: type=1300 audit(1769255459.042:560): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=3379 pid=3567 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:59.140017 kernel: audit: type=1327 audit(1769255459.042:560): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736383935386462653933373038333963373639623933316631646432 Jan 24 11:50:59.042000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736383935386462653933373038333963373639623933316631646432 Jan 24 11:50:59.042000 audit: BPF prog-id=167 op=LOAD Jan 24 11:50:59.042000 audit[3567]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=3379 pid=3567 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:59.042000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736383935386462653933373038333963373639623933316631646432 Jan 24 11:50:59.042000 audit: BPF prog-id=167 op=UNLOAD Jan 24 11:50:59.042000 audit[3567]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3379 pid=3567 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:59.042000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736383935386462653933373038333963373639623933316631646432 Jan 24 11:50:59.042000 audit: BPF prog-id=166 op=UNLOAD Jan 24 11:50:59.042000 audit[3567]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3379 pid=3567 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:59.042000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736383935386462653933373038333963373639623933316631646432 Jan 24 11:50:59.042000 audit: BPF prog-id=168 op=LOAD Jan 24 11:50:59.042000 audit[3567]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=3379 pid=3567 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:59.042000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736383935386462653933373038333963373639623933316631646432 Jan 24 11:50:59.259821 containerd[1647]: time="2026-01-24T11:50:59.259666656Z" level=info msg="StartContainer for \"768958dbe9370839c769b931f1dd2739b3550c64716377bb5f3669c9f6c7c9d8\" returns successfully" Jan 24 11:50:59.837556 kubelet[2895]: E0124 11:50:59.837223 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:50:59.899754 kubelet[2895]: E0124 11:50:59.899651 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.899754 kubelet[2895]: W0124 11:50:59.899685 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.903068 kubelet[2895]: E0124 11:50:59.902733 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.903323 kubelet[2895]: E0124 11:50:59.903228 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.903697 kubelet[2895]: W0124 11:50:59.903350 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.903697 kubelet[2895]: E0124 11:50:59.903378 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.904213 kubelet[2895]: E0124 11:50:59.904104 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.904213 kubelet[2895]: W0124 11:50:59.904123 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.905095 kubelet[2895]: E0124 11:50:59.904434 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.905233 kubelet[2895]: E0124 11:50:59.905216 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.905233 kubelet[2895]: W0124 11:50:59.905230 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.905233 kubelet[2895]: E0124 11:50:59.905247 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.908579 containerd[1647]: time="2026-01-24T11:50:59.908372887Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:50:59.909029 kubelet[2895]: E0124 11:50:59.908402 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.909029 kubelet[2895]: W0124 11:50:59.908425 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.909029 kubelet[2895]: E0124 11:50:59.908486 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.909029 kubelet[2895]: E0124 11:50:59.908773 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.909029 kubelet[2895]: W0124 11:50:59.908786 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.909029 kubelet[2895]: E0124 11:50:59.908802 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.909233 kubelet[2895]: E0124 11:50:59.909160 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.909233 kubelet[2895]: W0124 11:50:59.909173 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.909233 kubelet[2895]: E0124 11:50:59.909188 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.910846 kubelet[2895]: E0124 11:50:59.910733 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.910846 kubelet[2895]: W0124 11:50:59.910754 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.910846 kubelet[2895]: E0124 11:50:59.910769 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.911147 containerd[1647]: time="2026-01-24T11:50:59.910958357Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4: active requests=0, bytes read=4442579" Jan 24 11:50:59.912004 kubelet[2895]: E0124 11:50:59.911986 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.912350 kubelet[2895]: W0124 11:50:59.912090 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.912350 kubelet[2895]: E0124 11:50:59.912111 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.913777 kubelet[2895]: E0124 11:50:59.913759 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.913864 kubelet[2895]: W0124 11:50:59.913850 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.913991 kubelet[2895]: E0124 11:50:59.913925 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.915093 kubelet[2895]: E0124 11:50:59.914986 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.915093 kubelet[2895]: W0124 11:50:59.915006 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.915093 kubelet[2895]: E0124 11:50:59.915022 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.915637 kubelet[2895]: E0124 11:50:59.915620 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.915837 kubelet[2895]: W0124 11:50:59.915702 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.915837 kubelet[2895]: E0124 11:50:59.915723 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.916178 kubelet[2895]: E0124 11:50:59.916163 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.916250 kubelet[2895]: W0124 11:50:59.916236 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.916432 kubelet[2895]: E0124 11:50:59.916414 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.917072 kubelet[2895]: E0124 11:50:59.916980 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.917072 kubelet[2895]: W0124 11:50:59.916996 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.917072 kubelet[2895]: E0124 11:50:59.917010 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.917987 kubelet[2895]: E0124 11:50:59.917658 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.917987 kubelet[2895]: W0124 11:50:59.917674 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.917987 kubelet[2895]: E0124 11:50:59.917687 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.918122 containerd[1647]: time="2026-01-24T11:50:59.917650765Z" level=info msg="ImageCreate event name:\"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:50:59.918735 kubelet[2895]: E0124 11:50:59.918713 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.918822 kubelet[2895]: W0124 11:50:59.918805 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.918901 kubelet[2895]: E0124 11:50:59.918886 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.919578 kubelet[2895]: E0124 11:50:59.919559 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.919687 kubelet[2895]: W0124 11:50:59.919670 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.919760 kubelet[2895]: E0124 11:50:59.919745 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.920376 kubelet[2895]: E0124 11:50:59.920357 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.920496 kubelet[2895]: W0124 11:50:59.920478 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.920572 kubelet[2895]: E0124 11:50:59.920557 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.921565 kubelet[2895]: E0124 11:50:59.921174 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.921565 kubelet[2895]: W0124 11:50:59.921190 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.921565 kubelet[2895]: E0124 11:50:59.921204 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.921943 kubelet[2895]: E0124 11:50:59.921924 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.922027 kubelet[2895]: W0124 11:50:59.922013 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.922153 kubelet[2895]: E0124 11:50:59.922080 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.922779 kubelet[2895]: E0124 11:50:59.922758 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.923053 kubelet[2895]: W0124 11:50:59.922864 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.923053 kubelet[2895]: E0124 11:50:59.922885 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.923386 kubelet[2895]: E0124 11:50:59.923226 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.923514 kubelet[2895]: W0124 11:50:59.923496 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.923604 kubelet[2895]: E0124 11:50:59.923587 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.924229 kubelet[2895]: E0124 11:50:59.924211 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.924506 kubelet[2895]: W0124 11:50:59.924403 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.924506 kubelet[2895]: E0124 11:50:59.924480 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.925373 kubelet[2895]: E0124 11:50:59.924954 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.925373 kubelet[2895]: W0124 11:50:59.924967 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.925373 kubelet[2895]: E0124 11:50:59.924980 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.925562 containerd[1647]: time="2026-01-24T11:50:59.925008421Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:50:59.926089 containerd[1647]: time="2026-01-24T11:50:59.926019866Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" with image id \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\", size \"5941314\" in 1.089401362s" Jan 24 11:50:59.926089 containerd[1647]: time="2026-01-24T11:50:59.926081421Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" returns image reference \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\"" Jan 24 11:50:59.927299 kubelet[2895]: E0124 11:50:59.926797 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.927299 kubelet[2895]: W0124 11:50:59.926833 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.927299 kubelet[2895]: E0124 11:50:59.926848 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.927617 kubelet[2895]: E0124 11:50:59.927501 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.927617 kubelet[2895]: W0124 11:50:59.927541 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.927617 kubelet[2895]: E0124 11:50:59.927556 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.929386 kubelet[2895]: E0124 11:50:59.929093 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.929386 kubelet[2895]: W0124 11:50:59.929110 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.929386 kubelet[2895]: E0124 11:50:59.929126 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.931815 kubelet[2895]: E0124 11:50:59.931783 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.931815 kubelet[2895]: W0124 11:50:59.931803 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.931927 kubelet[2895]: E0124 11:50:59.931819 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.933601 kubelet[2895]: E0124 11:50:59.933570 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.933601 kubelet[2895]: W0124 11:50:59.933591 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.933689 kubelet[2895]: E0124 11:50:59.933607 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.934143 kubelet[2895]: I0124 11:50:59.933932 2895 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-dfdc7d98b-tn6mr" podStartSLOduration=3.315741992 podStartE2EDuration="7.933916008s" podCreationTimestamp="2026-01-24 11:50:52 +0000 UTC" firstStartedPulling="2026-01-24 11:50:54.217998916 +0000 UTC m=+33.739791622" lastFinishedPulling="2026-01-24 11:50:58.836172922 +0000 UTC m=+38.357965638" observedRunningTime="2026-01-24 11:50:59.895192889 +0000 UTC m=+39.416985615" watchObservedRunningTime="2026-01-24 11:50:59.933916008 +0000 UTC m=+39.455708724" Jan 24 11:50:59.935355 kubelet[2895]: E0124 11:50:59.935242 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.935836 kubelet[2895]: W0124 11:50:59.935631 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.935836 kubelet[2895]: E0124 11:50:59.935658 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.943776 kubelet[2895]: E0124 11:50:59.943140 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.943776 kubelet[2895]: W0124 11:50:59.943162 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.943776 kubelet[2895]: E0124 11:50:59.943182 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.944357 kubelet[2895]: E0124 11:50:59.944112 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.944357 kubelet[2895]: W0124 11:50:59.944158 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.944357 kubelet[2895]: E0124 11:50:59.944176 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.945045 kubelet[2895]: E0124 11:50:59.944671 2895 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 24 11:50:59.945045 kubelet[2895]: W0124 11:50:59.944684 2895 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 24 11:50:59.945045 kubelet[2895]: E0124 11:50:59.944699 2895 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 24 11:50:59.952589 containerd[1647]: time="2026-01-24T11:50:59.952403453Z" level=info msg="CreateContainer within sandbox \"2c075fdffb700b0c1172313d3ff66de2f4728237f591f9d6d03d9c3131fb8564\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Jan 24 11:50:59.989014 containerd[1647]: time="2026-01-24T11:50:59.988894530Z" level=info msg="Container 9392684f10a102713525bda642e35f8f28dd6872cdfe457ee861c5c621592826: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:50:59.998000 audit[3646]: NETFILTER_CFG table=filter:121 family=2 entries=21 op=nft_register_rule pid=3646 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:50:59.998000 audit[3646]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffcae240430 a2=0 a3=7ffcae24041c items=0 ppid=3040 pid=3646 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:50:59.998000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:00.014000 audit[3646]: NETFILTER_CFG table=nat:122 family=2 entries=19 op=nft_register_chain pid=3646 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:00.014000 audit[3646]: SYSCALL arch=c000003e syscall=46 success=yes exit=6276 a0=3 a1=7ffcae240430 a2=0 a3=7ffcae24041c items=0 ppid=3040 pid=3646 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:00.014000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:00.024064 containerd[1647]: time="2026-01-24T11:51:00.023939928Z" level=info msg="CreateContainer within sandbox \"2c075fdffb700b0c1172313d3ff66de2f4728237f591f9d6d03d9c3131fb8564\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"9392684f10a102713525bda642e35f8f28dd6872cdfe457ee861c5c621592826\"" Jan 24 11:51:00.029025 containerd[1647]: time="2026-01-24T11:51:00.028986465Z" level=info msg="StartContainer for \"9392684f10a102713525bda642e35f8f28dd6872cdfe457ee861c5c621592826\"" Jan 24 11:51:00.034072 containerd[1647]: time="2026-01-24T11:51:00.033697181Z" level=info msg="connecting to shim 9392684f10a102713525bda642e35f8f28dd6872cdfe457ee861c5c621592826" address="unix:///run/containerd/s/5856dbc4464bd7263ae0bc1ab234b2764d81a51e8bf8fc1354a687699a56c458" protocol=ttrpc version=3 Jan 24 11:51:00.098223 systemd[1]: Started cri-containerd-9392684f10a102713525bda642e35f8f28dd6872cdfe457ee861c5c621592826.scope - libcontainer container 9392684f10a102713525bda642e35f8f28dd6872cdfe457ee861c5c621592826. Jan 24 11:51:00.260000 audit: BPF prog-id=169 op=LOAD Jan 24 11:51:00.260000 audit[3647]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3491 pid=3647 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:00.260000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3933393236383466313061313032373133353235626461363432653335 Jan 24 11:51:00.260000 audit: BPF prog-id=170 op=LOAD Jan 24 11:51:00.260000 audit[3647]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3491 pid=3647 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:00.260000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3933393236383466313061313032373133353235626461363432653335 Jan 24 11:51:00.260000 audit: BPF prog-id=170 op=UNLOAD Jan 24 11:51:00.260000 audit[3647]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3491 pid=3647 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:00.260000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3933393236383466313061313032373133353235626461363432653335 Jan 24 11:51:00.260000 audit: BPF prog-id=169 op=UNLOAD Jan 24 11:51:00.260000 audit[3647]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3491 pid=3647 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:00.260000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3933393236383466313061313032373133353235626461363432653335 Jan 24 11:51:00.260000 audit: BPF prog-id=171 op=LOAD Jan 24 11:51:00.260000 audit[3647]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=3491 pid=3647 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:00.260000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3933393236383466313061313032373133353235626461363432653335 Jan 24 11:51:00.387968 containerd[1647]: time="2026-01-24T11:51:00.387435850Z" level=info msg="StartContainer for \"9392684f10a102713525bda642e35f8f28dd6872cdfe457ee861c5c621592826\" returns successfully" Jan 24 11:51:00.433166 systemd[1]: cri-containerd-9392684f10a102713525bda642e35f8f28dd6872cdfe457ee861c5c621592826.scope: Deactivated successfully. Jan 24 11:51:00.437611 containerd[1647]: time="2026-01-24T11:51:00.437382589Z" level=info msg="received container exit event container_id:\"9392684f10a102713525bda642e35f8f28dd6872cdfe457ee861c5c621592826\" id:\"9392684f10a102713525bda642e35f8f28dd6872cdfe457ee861c5c621592826\" pid:3660 exited_at:{seconds:1769255460 nanos:436737796}" Jan 24 11:51:00.440000 audit: BPF prog-id=171 op=UNLOAD Jan 24 11:51:00.547166 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9392684f10a102713525bda642e35f8f28dd6872cdfe457ee861c5c621592826-rootfs.mount: Deactivated successfully. Jan 24 11:51:00.845490 kubelet[2895]: E0124 11:51:00.845377 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:00.846919 kubelet[2895]: E0124 11:51:00.846729 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:00.848566 containerd[1647]: time="2026-01-24T11:51:00.848067391Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\"" Jan 24 11:51:01.132209 kubelet[2895]: E0124 11:51:01.131838 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:51:01.939402 kubelet[2895]: E0124 11:51:01.937331 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:03.133686 kubelet[2895]: E0124 11:51:03.133520 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:51:05.136628 kubelet[2895]: E0124 11:51:05.134717 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:51:07.135044 kubelet[2895]: E0124 11:51:07.134768 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:51:07.822651 containerd[1647]: time="2026-01-24T11:51:07.820379954Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:51:07.827715 containerd[1647]: time="2026-01-24T11:51:07.827551718Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.4: active requests=0, bytes read=70442291" Jan 24 11:51:07.833846 containerd[1647]: time="2026-01-24T11:51:07.833337938Z" level=info msg="ImageCreate event name:\"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:51:07.849529 containerd[1647]: time="2026-01-24T11:51:07.849011312Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:51:07.856557 containerd[1647]: time="2026-01-24T11:51:07.856503012Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.4\" with image id \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\", size \"71941459\" in 7.008343541s" Jan 24 11:51:07.859445 containerd[1647]: time="2026-01-24T11:51:07.858500743Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\" returns image reference \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\"" Jan 24 11:51:07.982889 containerd[1647]: time="2026-01-24T11:51:07.981216119Z" level=info msg="CreateContainer within sandbox \"2c075fdffb700b0c1172313d3ff66de2f4728237f591f9d6d03d9c3131fb8564\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jan 24 11:51:08.283425 containerd[1647]: time="2026-01-24T11:51:08.281763685Z" level=info msg="Container 2d01038755a5e554ee9dcd53417ffdd8d58632e05b70ed202cdadf0bc637d23a: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:51:08.845935 containerd[1647]: time="2026-01-24T11:51:08.842142953Z" level=info msg="CreateContainer within sandbox \"2c075fdffb700b0c1172313d3ff66de2f4728237f591f9d6d03d9c3131fb8564\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"2d01038755a5e554ee9dcd53417ffdd8d58632e05b70ed202cdadf0bc637d23a\"" Jan 24 11:51:08.852352 containerd[1647]: time="2026-01-24T11:51:08.849118291Z" level=info msg="StartContainer for \"2d01038755a5e554ee9dcd53417ffdd8d58632e05b70ed202cdadf0bc637d23a\"" Jan 24 11:51:08.875894 containerd[1647]: time="2026-01-24T11:51:08.874189046Z" level=info msg="connecting to shim 2d01038755a5e554ee9dcd53417ffdd8d58632e05b70ed202cdadf0bc637d23a" address="unix:///run/containerd/s/5856dbc4464bd7263ae0bc1ab234b2764d81a51e8bf8fc1354a687699a56c458" protocol=ttrpc version=3 Jan 24 11:51:09.137617 kubelet[2895]: E0124 11:51:09.137077 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:51:09.190386 systemd[1]: Started cri-containerd-2d01038755a5e554ee9dcd53417ffdd8d58632e05b70ed202cdadf0bc637d23a.scope - libcontainer container 2d01038755a5e554ee9dcd53417ffdd8d58632e05b70ed202cdadf0bc637d23a. Jan 24 11:51:09.379801 kernel: kauditd_printk_skb: 34 callbacks suppressed Jan 24 11:51:09.380179 kernel: audit: type=1334 audit(1769255469.369:573): prog-id=172 op=LOAD Jan 24 11:51:09.369000 audit: BPF prog-id=172 op=LOAD Jan 24 11:51:09.397002 kernel: audit: type=1300 audit(1769255469.369:573): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a8488 a2=98 a3=0 items=0 ppid=3491 pid=3707 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:09.369000 audit[3707]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a8488 a2=98 a3=0 items=0 ppid=3491 pid=3707 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:09.369000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3264303130333837353561356535353465653964636435333431376666 Jan 24 11:51:09.445209 kernel: audit: type=1327 audit(1769255469.369:573): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3264303130333837353561356535353465653964636435333431376666 Jan 24 11:51:09.446120 kernel: audit: type=1334 audit(1769255469.369:574): prog-id=173 op=LOAD Jan 24 11:51:09.446164 kernel: audit: type=1300 audit(1769255469.369:574): arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001a8218 a2=98 a3=0 items=0 ppid=3491 pid=3707 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:09.369000 audit: BPF prog-id=173 op=LOAD Jan 24 11:51:09.369000 audit[3707]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001a8218 a2=98 a3=0 items=0 ppid=3491 pid=3707 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:09.473958 kernel: audit: type=1327 audit(1769255469.369:574): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3264303130333837353561356535353465653964636435333431376666 Jan 24 11:51:09.369000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3264303130333837353561356535353465653964636435333431376666 Jan 24 11:51:09.369000 audit: BPF prog-id=173 op=UNLOAD Jan 24 11:51:09.369000 audit[3707]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3491 pid=3707 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:09.550800 kernel: audit: type=1334 audit(1769255469.369:575): prog-id=173 op=UNLOAD Jan 24 11:51:09.595091 kernel: audit: type=1300 audit(1769255469.369:575): arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3491 pid=3707 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:09.624444 kernel: audit: type=1327 audit(1769255469.369:575): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3264303130333837353561356535353465653964636435333431376666 Jan 24 11:51:09.625004 kernel: audit: type=1334 audit(1769255469.369:576): prog-id=172 op=UNLOAD Jan 24 11:51:09.369000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3264303130333837353561356535353465653964636435333431376666 Jan 24 11:51:09.369000 audit: BPF prog-id=172 op=UNLOAD Jan 24 11:51:09.369000 audit[3707]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3491 pid=3707 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:09.369000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3264303130333837353561356535353465653964636435333431376666 Jan 24 11:51:09.369000 audit: BPF prog-id=174 op=LOAD Jan 24 11:51:09.369000 audit[3707]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a86e8 a2=98 a3=0 items=0 ppid=3491 pid=3707 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:09.369000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3264303130333837353561356535353465653964636435333431376666 Jan 24 11:51:09.950343 containerd[1647]: time="2026-01-24T11:51:09.950212707Z" level=info msg="StartContainer for \"2d01038755a5e554ee9dcd53417ffdd8d58632e05b70ed202cdadf0bc637d23a\" returns successfully" Jan 24 11:51:10.265840 kubelet[2895]: E0124 11:51:10.265702 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:11.131380 kubelet[2895]: E0124 11:51:11.130895 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:51:11.262080 kubelet[2895]: E0124 11:51:11.261346 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:12.588589 systemd[1]: cri-containerd-2d01038755a5e554ee9dcd53417ffdd8d58632e05b70ed202cdadf0bc637d23a.scope: Deactivated successfully. Jan 24 11:51:12.589601 systemd[1]: cri-containerd-2d01038755a5e554ee9dcd53417ffdd8d58632e05b70ed202cdadf0bc637d23a.scope: Consumed 1.794s CPU time, 177.7M memory peak, 2.9M read from disk, 171.3M written to disk. Jan 24 11:51:12.597334 containerd[1647]: time="2026-01-24T11:51:12.596040332Z" level=info msg="received container exit event container_id:\"2d01038755a5e554ee9dcd53417ffdd8d58632e05b70ed202cdadf0bc637d23a\" id:\"2d01038755a5e554ee9dcd53417ffdd8d58632e05b70ed202cdadf0bc637d23a\" pid:3718 exited_at:{seconds:1769255472 nanos:595654342}" Jan 24 11:51:12.597000 audit: BPF prog-id=174 op=UNLOAD Jan 24 11:51:12.760533 kubelet[2895]: I0124 11:51:12.760427 2895 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Jan 24 11:51:12.764846 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2d01038755a5e554ee9dcd53417ffdd8d58632e05b70ed202cdadf0bc637d23a-rootfs.mount: Deactivated successfully. Jan 24 11:51:12.941643 kubelet[2895]: I0124 11:51:12.941019 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/c28e9f44-9bc9-4328-bf7a-f4f9d016a87e-calico-apiserver-certs\") pod \"calico-apiserver-f485b4455-qhcml\" (UID: \"c28e9f44-9bc9-4328-bf7a-f4f9d016a87e\") " pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" Jan 24 11:51:12.941643 kubelet[2895]: I0124 11:51:12.941094 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41fe2d74-dafe-45a2-82d2-5fd56447e59d-tigera-ca-bundle\") pod \"calico-kube-controllers-66955b7fc7-xmscg\" (UID: \"41fe2d74-dafe-45a2-82d2-5fd56447e59d\") " pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" Jan 24 11:51:12.941643 kubelet[2895]: I0124 11:51:12.941130 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l682s\" (UniqueName: \"kubernetes.io/projected/c28e9f44-9bc9-4328-bf7a-f4f9d016a87e-kube-api-access-l682s\") pod \"calico-apiserver-f485b4455-qhcml\" (UID: \"c28e9f44-9bc9-4328-bf7a-f4f9d016a87e\") " pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" Jan 24 11:51:12.941643 kubelet[2895]: I0124 11:51:12.941173 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5kgf\" (UniqueName: \"kubernetes.io/projected/41fe2d74-dafe-45a2-82d2-5fd56447e59d-kube-api-access-p5kgf\") pod \"calico-kube-controllers-66955b7fc7-xmscg\" (UID: \"41fe2d74-dafe-45a2-82d2-5fd56447e59d\") " pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" Jan 24 11:51:12.947551 systemd[1]: Created slice kubepods-besteffort-pod41fe2d74_dafe_45a2_82d2_5fd56447e59d.slice - libcontainer container kubepods-besteffort-pod41fe2d74_dafe_45a2_82d2_5fd56447e59d.slice. Jan 24 11:51:12.983753 systemd[1]: Created slice kubepods-besteffort-podd47a09e2_52ab_4e49_942e_05f2cedb67ae.slice - libcontainer container kubepods-besteffort-podd47a09e2_52ab_4e49_942e_05f2cedb67ae.slice. Jan 24 11:51:13.020252 systemd[1]: Created slice kubepods-besteffort-pod628627fd_f89d_4b0b_af23_27eb9101b7cb.slice - libcontainer container kubepods-besteffort-pod628627fd_f89d_4b0b_af23_27eb9101b7cb.slice. Jan 24 11:51:13.045217 kubelet[2895]: I0124 11:51:13.044931 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4s6n\" (UniqueName: \"kubernetes.io/projected/d47a09e2-52ab-4e49-942e-05f2cedb67ae-kube-api-access-r4s6n\") pod \"calico-apiserver-f485b4455-nc2lh\" (UID: \"d47a09e2-52ab-4e49-942e-05f2cedb67ae\") " pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" Jan 24 11:51:13.045217 kubelet[2895]: I0124 11:51:13.044996 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/628627fd-f89d-4b0b-af23-27eb9101b7cb-whisker-ca-bundle\") pod \"whisker-c7c4667dc-5v9ql\" (UID: \"628627fd-f89d-4b0b-af23-27eb9101b7cb\") " pod="calico-system/whisker-c7c4667dc-5v9ql" Jan 24 11:51:13.045217 kubelet[2895]: I0124 11:51:13.045025 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/293f1d4a-795c-44fc-bda8-3c550f56df72-goldmane-key-pair\") pod \"goldmane-666569f655-jfngl\" (UID: \"293f1d4a-795c-44fc-bda8-3c550f56df72\") " pod="calico-system/goldmane-666569f655-jfngl" Jan 24 11:51:13.045217 kubelet[2895]: I0124 11:51:13.045052 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xpg5\" (UniqueName: \"kubernetes.io/projected/f46cf494-de53-4117-858a-b0e56c0ba7ac-kube-api-access-9xpg5\") pod \"coredns-674b8bbfcf-k9kt6\" (UID: \"f46cf494-de53-4117-858a-b0e56c0ba7ac\") " pod="kube-system/coredns-674b8bbfcf-k9kt6" Jan 24 11:51:13.045217 kubelet[2895]: I0124 11:51:13.045106 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/628627fd-f89d-4b0b-af23-27eb9101b7cb-whisker-backend-key-pair\") pod \"whisker-c7c4667dc-5v9ql\" (UID: \"628627fd-f89d-4b0b-af23-27eb9101b7cb\") " pod="calico-system/whisker-c7c4667dc-5v9ql" Jan 24 11:51:13.046812 kubelet[2895]: I0124 11:51:13.045129 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f46cf494-de53-4117-858a-b0e56c0ba7ac-config-volume\") pod \"coredns-674b8bbfcf-k9kt6\" (UID: \"f46cf494-de53-4117-858a-b0e56c0ba7ac\") " pod="kube-system/coredns-674b8bbfcf-k9kt6" Jan 24 11:51:13.046812 kubelet[2895]: I0124 11:51:13.045195 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmh49\" (UniqueName: \"kubernetes.io/projected/7a087061-e7ac-47a8-a3c4-b753efe3e78f-kube-api-access-rmh49\") pod \"coredns-674b8bbfcf-dnscb\" (UID: \"7a087061-e7ac-47a8-a3c4-b753efe3e78f\") " pod="kube-system/coredns-674b8bbfcf-dnscb" Jan 24 11:51:13.046812 kubelet[2895]: I0124 11:51:13.045225 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a087061-e7ac-47a8-a3c4-b753efe3e78f-config-volume\") pod \"coredns-674b8bbfcf-dnscb\" (UID: \"7a087061-e7ac-47a8-a3c4-b753efe3e78f\") " pod="kube-system/coredns-674b8bbfcf-dnscb" Jan 24 11:51:13.046812 kubelet[2895]: I0124 11:51:13.045422 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/d47a09e2-52ab-4e49-942e-05f2cedb67ae-calico-apiserver-certs\") pod \"calico-apiserver-f485b4455-nc2lh\" (UID: \"d47a09e2-52ab-4e49-942e-05f2cedb67ae\") " pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" Jan 24 11:51:13.046812 kubelet[2895]: I0124 11:51:13.045457 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2g54\" (UniqueName: \"kubernetes.io/projected/293f1d4a-795c-44fc-bda8-3c550f56df72-kube-api-access-k2g54\") pod \"goldmane-666569f655-jfngl\" (UID: \"293f1d4a-795c-44fc-bda8-3c550f56df72\") " pod="calico-system/goldmane-666569f655-jfngl" Jan 24 11:51:13.047455 kubelet[2895]: I0124 11:51:13.045528 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b48zv\" (UniqueName: \"kubernetes.io/projected/628627fd-f89d-4b0b-af23-27eb9101b7cb-kube-api-access-b48zv\") pod \"whisker-c7c4667dc-5v9ql\" (UID: \"628627fd-f89d-4b0b-af23-27eb9101b7cb\") " pod="calico-system/whisker-c7c4667dc-5v9ql" Jan 24 11:51:13.047455 kubelet[2895]: I0124 11:51:13.045560 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/293f1d4a-795c-44fc-bda8-3c550f56df72-config\") pod \"goldmane-666569f655-jfngl\" (UID: \"293f1d4a-795c-44fc-bda8-3c550f56df72\") " pod="calico-system/goldmane-666569f655-jfngl" Jan 24 11:51:13.047455 kubelet[2895]: I0124 11:51:13.045582 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/293f1d4a-795c-44fc-bda8-3c550f56df72-goldmane-ca-bundle\") pod \"goldmane-666569f655-jfngl\" (UID: \"293f1d4a-795c-44fc-bda8-3c550f56df72\") " pod="calico-system/goldmane-666569f655-jfngl" Jan 24 11:51:13.061417 systemd[1]: Created slice kubepods-besteffort-podc28e9f44_9bc9_4328_bf7a_f4f9d016a87e.slice - libcontainer container kubepods-besteffort-podc28e9f44_9bc9_4328_bf7a_f4f9d016a87e.slice. Jan 24 11:51:13.078718 systemd[1]: Created slice kubepods-burstable-podf46cf494_de53_4117_858a_b0e56c0ba7ac.slice - libcontainer container kubepods-burstable-podf46cf494_de53_4117_858a_b0e56c0ba7ac.slice. Jan 24 11:51:13.183518 systemd[1]: Created slice kubepods-burstable-pod7a087061_e7ac_47a8_a3c4_b753efe3e78f.slice - libcontainer container kubepods-burstable-pod7a087061_e7ac_47a8_a3c4_b753efe3e78f.slice. Jan 24 11:51:13.222589 systemd[1]: Created slice kubepods-besteffort-pod293f1d4a_795c_44fc_bda8_3c550f56df72.slice - libcontainer container kubepods-besteffort-pod293f1d4a_795c_44fc_bda8_3c550f56df72.slice. Jan 24 11:51:13.234362 containerd[1647]: time="2026-01-24T11:51:13.234213687Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-jfngl,Uid:293f1d4a-795c-44fc-bda8-3c550f56df72,Namespace:calico-system,Attempt:0,}" Jan 24 11:51:13.253038 systemd[1]: Created slice kubepods-besteffort-pod999c48c5_f3b1_4bfd_91f4_26cafee1eef0.slice - libcontainer container kubepods-besteffort-pod999c48c5_f3b1_4bfd_91f4_26cafee1eef0.slice. Jan 24 11:51:13.270935 containerd[1647]: time="2026-01-24T11:51:13.270885548Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-2g45n,Uid:999c48c5-f3b1-4bfd-91f4-26cafee1eef0,Namespace:calico-system,Attempt:0,}" Jan 24 11:51:13.271600 containerd[1647]: time="2026-01-24T11:51:13.270982292Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-66955b7fc7-xmscg,Uid:41fe2d74-dafe-45a2-82d2-5fd56447e59d,Namespace:calico-system,Attempt:0,}" Jan 24 11:51:13.287234 kubelet[2895]: E0124 11:51:13.284787 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:13.288953 containerd[1647]: time="2026-01-24T11:51:13.288770647Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\"" Jan 24 11:51:13.300355 containerd[1647]: time="2026-01-24T11:51:13.299138834Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f485b4455-nc2lh,Uid:d47a09e2-52ab-4e49-942e-05f2cedb67ae,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:51:13.359141 containerd[1647]: time="2026-01-24T11:51:13.359047879Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-c7c4667dc-5v9ql,Uid:628627fd-f89d-4b0b-af23-27eb9101b7cb,Namespace:calico-system,Attempt:0,}" Jan 24 11:51:13.398720 containerd[1647]: time="2026-01-24T11:51:13.397611220Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f485b4455-qhcml,Uid:c28e9f44-9bc9-4328-bf7a-f4f9d016a87e,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:51:13.449513 kubelet[2895]: E0124 11:51:13.449408 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:13.467873 containerd[1647]: time="2026-01-24T11:51:13.467653351Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-k9kt6,Uid:f46cf494-de53-4117-858a-b0e56c0ba7ac,Namespace:kube-system,Attempt:0,}" Jan 24 11:51:13.543420 kubelet[2895]: E0124 11:51:13.534557 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:13.609179 containerd[1647]: time="2026-01-24T11:51:13.609113260Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-dnscb,Uid:7a087061-e7ac-47a8-a3c4-b753efe3e78f,Namespace:kube-system,Attempt:0,}" Jan 24 11:51:14.091727 containerd[1647]: time="2026-01-24T11:51:14.090002099Z" level=error msg="Failed to destroy network for sandbox \"3b1fc45b01af4cec8780c990e8b104ecb93ee0a8d93313bff256c980732ea835\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.095518 systemd[1]: run-netns-cni\x2d42bddb34\x2dc23d\x2dbaf4\x2d10fd\x2d5a714d9107ce.mount: Deactivated successfully. Jan 24 11:51:14.117891 containerd[1647]: time="2026-01-24T11:51:14.117695158Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-2g45n,Uid:999c48c5-f3b1-4bfd-91f4-26cafee1eef0,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"3b1fc45b01af4cec8780c990e8b104ecb93ee0a8d93313bff256c980732ea835\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.118220 kubelet[2895]: E0124 11:51:14.118165 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3b1fc45b01af4cec8780c990e8b104ecb93ee0a8d93313bff256c980732ea835\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.119384 kubelet[2895]: E0124 11:51:14.118322 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3b1fc45b01af4cec8780c990e8b104ecb93ee0a8d93313bff256c980732ea835\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-2g45n" Jan 24 11:51:14.119384 kubelet[2895]: E0124 11:51:14.118361 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3b1fc45b01af4cec8780c990e8b104ecb93ee0a8d93313bff256c980732ea835\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-2g45n" Jan 24 11:51:14.119384 kubelet[2895]: E0124 11:51:14.118427 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-2g45n_calico-system(999c48c5-f3b1-4bfd-91f4-26cafee1eef0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-2g45n_calico-system(999c48c5-f3b1-4bfd-91f4-26cafee1eef0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3b1fc45b01af4cec8780c990e8b104ecb93ee0a8d93313bff256c980732ea835\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:51:14.135695 containerd[1647]: time="2026-01-24T11:51:14.134391072Z" level=error msg="Failed to destroy network for sandbox \"b910316cf6575019fbb3ea4220aaa7c45b6315c78d2b55dd59b440c9788ac331\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.145871 systemd[1]: run-netns-cni\x2dd50ec354\x2d620f\x2d1164\x2dcc0d\x2df9bda30501f0.mount: Deactivated successfully. Jan 24 11:51:14.155079 containerd[1647]: time="2026-01-24T11:51:14.154735268Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-k9kt6,Uid:f46cf494-de53-4117-858a-b0e56c0ba7ac,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"b910316cf6575019fbb3ea4220aaa7c45b6315c78d2b55dd59b440c9788ac331\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.156763 kubelet[2895]: E0124 11:51:14.155715 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b910316cf6575019fbb3ea4220aaa7c45b6315c78d2b55dd59b440c9788ac331\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.156763 kubelet[2895]: E0124 11:51:14.155906 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b910316cf6575019fbb3ea4220aaa7c45b6315c78d2b55dd59b440c9788ac331\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-k9kt6" Jan 24 11:51:14.156763 kubelet[2895]: E0124 11:51:14.155953 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b910316cf6575019fbb3ea4220aaa7c45b6315c78d2b55dd59b440c9788ac331\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-k9kt6" Jan 24 11:51:14.157141 kubelet[2895]: E0124 11:51:14.156033 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-k9kt6_kube-system(f46cf494-de53-4117-858a-b0e56c0ba7ac)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-k9kt6_kube-system(f46cf494-de53-4117-858a-b0e56c0ba7ac)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b910316cf6575019fbb3ea4220aaa7c45b6315c78d2b55dd59b440c9788ac331\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-k9kt6" podUID="f46cf494-de53-4117-858a-b0e56c0ba7ac" Jan 24 11:51:14.166247 containerd[1647]: time="2026-01-24T11:51:14.166010783Z" level=error msg="Failed to destroy network for sandbox \"e53fd5d1064e2a394c4ca18d0053b94922ea670fb8183454c159309f45c910a4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.174104 systemd[1]: run-netns-cni\x2d7ef22832\x2d88a6\x2d12b4\x2ded33\x2da5f88dbb22b9.mount: Deactivated successfully. Jan 24 11:51:14.177931 containerd[1647]: time="2026-01-24T11:51:14.177795983Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-jfngl,Uid:293f1d4a-795c-44fc-bda8-3c550f56df72,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e53fd5d1064e2a394c4ca18d0053b94922ea670fb8183454c159309f45c910a4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.178394 kubelet[2895]: E0124 11:51:14.178239 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e53fd5d1064e2a394c4ca18d0053b94922ea670fb8183454c159309f45c910a4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.178394 kubelet[2895]: E0124 11:51:14.178359 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e53fd5d1064e2a394c4ca18d0053b94922ea670fb8183454c159309f45c910a4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-jfngl" Jan 24 11:51:14.178394 kubelet[2895]: E0124 11:51:14.178384 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e53fd5d1064e2a394c4ca18d0053b94922ea670fb8183454c159309f45c910a4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-jfngl" Jan 24 11:51:14.178622 kubelet[2895]: E0124 11:51:14.178431 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-jfngl_calico-system(293f1d4a-795c-44fc-bda8-3c550f56df72)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-jfngl_calico-system(293f1d4a-795c-44fc-bda8-3c550f56df72)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e53fd5d1064e2a394c4ca18d0053b94922ea670fb8183454c159309f45c910a4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:51:14.189381 containerd[1647]: time="2026-01-24T11:51:14.189238060Z" level=error msg="Failed to destroy network for sandbox \"a767b275f6727aabbeb1c92366199ed67165d975464ee2e45e53d12ba169a397\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.193641 systemd[1]: run-netns-cni\x2d384d0540\x2d4478\x2d7114\x2dd875\x2d78530c731c16.mount: Deactivated successfully. Jan 24 11:51:14.214040 containerd[1647]: time="2026-01-24T11:51:14.213980540Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f485b4455-nc2lh,Uid:d47a09e2-52ab-4e49-942e-05f2cedb67ae,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"a767b275f6727aabbeb1c92366199ed67165d975464ee2e45e53d12ba169a397\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.214539 containerd[1647]: time="2026-01-24T11:51:14.214126264Z" level=error msg="Failed to destroy network for sandbox \"540c256f0f57391cc4fd8eb9d98ee45bf8ec96a3ac41001ea25aa04e69a1b7cf\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.214643 kubelet[2895]: E0124 11:51:14.214232 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a767b275f6727aabbeb1c92366199ed67165d975464ee2e45e53d12ba169a397\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.214643 kubelet[2895]: E0124 11:51:14.214467 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a767b275f6727aabbeb1c92366199ed67165d975464ee2e45e53d12ba169a397\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" Jan 24 11:51:14.214643 kubelet[2895]: E0124 11:51:14.214528 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a767b275f6727aabbeb1c92366199ed67165d975464ee2e45e53d12ba169a397\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" Jan 24 11:51:14.214796 kubelet[2895]: E0124 11:51:14.214600 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-f485b4455-nc2lh_calico-apiserver(d47a09e2-52ab-4e49-942e-05f2cedb67ae)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-f485b4455-nc2lh_calico-apiserver(d47a09e2-52ab-4e49-942e-05f2cedb67ae)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a767b275f6727aabbeb1c92366199ed67165d975464ee2e45e53d12ba169a397\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:51:14.220454 containerd[1647]: time="2026-01-24T11:51:14.220238217Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-c7c4667dc-5v9ql,Uid:628627fd-f89d-4b0b-af23-27eb9101b7cb,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"540c256f0f57391cc4fd8eb9d98ee45bf8ec96a3ac41001ea25aa04e69a1b7cf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.222343 kubelet[2895]: E0124 11:51:14.222143 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"540c256f0f57391cc4fd8eb9d98ee45bf8ec96a3ac41001ea25aa04e69a1b7cf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.222343 kubelet[2895]: E0124 11:51:14.222202 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"540c256f0f57391cc4fd8eb9d98ee45bf8ec96a3ac41001ea25aa04e69a1b7cf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-c7c4667dc-5v9ql" Jan 24 11:51:14.222343 kubelet[2895]: E0124 11:51:14.222239 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"540c256f0f57391cc4fd8eb9d98ee45bf8ec96a3ac41001ea25aa04e69a1b7cf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-c7c4667dc-5v9ql" Jan 24 11:51:14.222644 kubelet[2895]: E0124 11:51:14.222366 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-c7c4667dc-5v9ql_calico-system(628627fd-f89d-4b0b-af23-27eb9101b7cb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-c7c4667dc-5v9ql_calico-system(628627fd-f89d-4b0b-af23-27eb9101b7cb)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"540c256f0f57391cc4fd8eb9d98ee45bf8ec96a3ac41001ea25aa04e69a1b7cf\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-c7c4667dc-5v9ql" podUID="628627fd-f89d-4b0b-af23-27eb9101b7cb" Jan 24 11:51:14.233424 containerd[1647]: time="2026-01-24T11:51:14.232549773Z" level=error msg="Failed to destroy network for sandbox \"9466cd09b1c4872a1aa5f4c15085937b7f9f8868652a7b0c65dcb74b5afa7963\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.233424 containerd[1647]: time="2026-01-24T11:51:14.233214142Z" level=error msg="Failed to destroy network for sandbox \"0a508fd267cef3053118cc3596a99f173d9b8f74772d410d2ba2c64d5a830217\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.238377 containerd[1647]: time="2026-01-24T11:51:14.238323582Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-dnscb,Uid:7a087061-e7ac-47a8-a3c4-b753efe3e78f,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"9466cd09b1c4872a1aa5f4c15085937b7f9f8868652a7b0c65dcb74b5afa7963\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.239879 kubelet[2895]: E0124 11:51:14.239399 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9466cd09b1c4872a1aa5f4c15085937b7f9f8868652a7b0c65dcb74b5afa7963\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.240397 kubelet[2895]: E0124 11:51:14.240366 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9466cd09b1c4872a1aa5f4c15085937b7f9f8868652a7b0c65dcb74b5afa7963\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-dnscb" Jan 24 11:51:14.243355 kubelet[2895]: E0124 11:51:14.240646 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9466cd09b1c4872a1aa5f4c15085937b7f9f8868652a7b0c65dcb74b5afa7963\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-dnscb" Jan 24 11:51:14.243355 kubelet[2895]: E0124 11:51:14.241966 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-dnscb_kube-system(7a087061-e7ac-47a8-a3c4-b753efe3e78f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-dnscb_kube-system(7a087061-e7ac-47a8-a3c4-b753efe3e78f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9466cd09b1c4872a1aa5f4c15085937b7f9f8868652a7b0c65dcb74b5afa7963\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-dnscb" podUID="7a087061-e7ac-47a8-a3c4-b753efe3e78f" Jan 24 11:51:14.247348 containerd[1647]: time="2026-01-24T11:51:14.246595680Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-66955b7fc7-xmscg,Uid:41fe2d74-dafe-45a2-82d2-5fd56447e59d,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"0a508fd267cef3053118cc3596a99f173d9b8f74772d410d2ba2c64d5a830217\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.248128 kubelet[2895]: E0124 11:51:14.247880 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0a508fd267cef3053118cc3596a99f173d9b8f74772d410d2ba2c64d5a830217\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.248478 kubelet[2895]: E0124 11:51:14.248427 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0a508fd267cef3053118cc3596a99f173d9b8f74772d410d2ba2c64d5a830217\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" Jan 24 11:51:14.248587 kubelet[2895]: E0124 11:51:14.248517 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0a508fd267cef3053118cc3596a99f173d9b8f74772d410d2ba2c64d5a830217\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" Jan 24 11:51:14.248646 kubelet[2895]: E0124 11:51:14.248579 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-66955b7fc7-xmscg_calico-system(41fe2d74-dafe-45a2-82d2-5fd56447e59d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-66955b7fc7-xmscg_calico-system(41fe2d74-dafe-45a2-82d2-5fd56447e59d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0a508fd267cef3053118cc3596a99f173d9b8f74772d410d2ba2c64d5a830217\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:51:14.259336 containerd[1647]: time="2026-01-24T11:51:14.259204847Z" level=error msg="Failed to destroy network for sandbox \"8c78d3775b70d51f72f9a221656f23c30a0c7bc42aa995ed126c1b3d7eeca791\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.267202 containerd[1647]: time="2026-01-24T11:51:14.267117051Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f485b4455-qhcml,Uid:c28e9f44-9bc9-4328-bf7a-f4f9d016a87e,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"8c78d3775b70d51f72f9a221656f23c30a0c7bc42aa995ed126c1b3d7eeca791\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.268109 kubelet[2895]: E0124 11:51:14.267826 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8c78d3775b70d51f72f9a221656f23c30a0c7bc42aa995ed126c1b3d7eeca791\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:14.268109 kubelet[2895]: E0124 11:51:14.268016 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8c78d3775b70d51f72f9a221656f23c30a0c7bc42aa995ed126c1b3d7eeca791\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" Jan 24 11:51:14.268109 kubelet[2895]: E0124 11:51:14.268041 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8c78d3775b70d51f72f9a221656f23c30a0c7bc42aa995ed126c1b3d7eeca791\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" Jan 24 11:51:14.268333 kubelet[2895]: E0124 11:51:14.268222 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-f485b4455-qhcml_calico-apiserver(c28e9f44-9bc9-4328-bf7a-f4f9d016a87e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-f485b4455-qhcml_calico-apiserver(c28e9f44-9bc9-4328-bf7a-f4f9d016a87e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8c78d3775b70d51f72f9a221656f23c30a0c7bc42aa995ed126c1b3d7eeca791\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:51:14.770863 systemd[1]: run-netns-cni\x2dc423c682\x2d9de0\x2df2cb\x2d26c7\x2d3df2388c0845.mount: Deactivated successfully. Jan 24 11:51:14.771055 systemd[1]: run-netns-cni\x2d226fb090\x2dbc9c\x2ddf38\x2d1c9d\x2d454f2aadd1cd.mount: Deactivated successfully. Jan 24 11:51:14.771161 systemd[1]: run-netns-cni\x2df110743b\x2de23d\x2da268\x2ddcf6\x2d5ba720a58325.mount: Deactivated successfully. Jan 24 11:51:14.771319 systemd[1]: run-netns-cni\x2dc1c87cf9\x2dee9d\x2db582\x2d9899\x2d2881cb49bb52.mount: Deactivated successfully. Jan 24 11:51:25.148806 containerd[1647]: time="2026-01-24T11:51:25.145503687Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-c7c4667dc-5v9ql,Uid:628627fd-f89d-4b0b-af23-27eb9101b7cb,Namespace:calico-system,Attempt:0,}" Jan 24 11:51:25.505360 containerd[1647]: time="2026-01-24T11:51:25.502429113Z" level=error msg="Failed to destroy network for sandbox \"2058f79065dbebee4487e4b95bfe324bcfeca274b1a6e20e9cda41bd82dde426\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:25.509159 systemd[1]: run-netns-cni\x2db61765f3\x2d3322\x2d974d\x2da0e8\x2d7cd4578dd7b4.mount: Deactivated successfully. Jan 24 11:51:25.532504 containerd[1647]: time="2026-01-24T11:51:25.531247676Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-c7c4667dc-5v9ql,Uid:628627fd-f89d-4b0b-af23-27eb9101b7cb,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"2058f79065dbebee4487e4b95bfe324bcfeca274b1a6e20e9cda41bd82dde426\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:25.532786 kubelet[2895]: E0124 11:51:25.531831 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2058f79065dbebee4487e4b95bfe324bcfeca274b1a6e20e9cda41bd82dde426\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:25.532786 kubelet[2895]: E0124 11:51:25.531988 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2058f79065dbebee4487e4b95bfe324bcfeca274b1a6e20e9cda41bd82dde426\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-c7c4667dc-5v9ql" Jan 24 11:51:25.532786 kubelet[2895]: E0124 11:51:25.532022 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2058f79065dbebee4487e4b95bfe324bcfeca274b1a6e20e9cda41bd82dde426\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-c7c4667dc-5v9ql" Jan 24 11:51:25.536049 kubelet[2895]: E0124 11:51:25.532083 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-c7c4667dc-5v9ql_calico-system(628627fd-f89d-4b0b-af23-27eb9101b7cb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-c7c4667dc-5v9ql_calico-system(628627fd-f89d-4b0b-af23-27eb9101b7cb)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2058f79065dbebee4487e4b95bfe324bcfeca274b1a6e20e9cda41bd82dde426\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-c7c4667dc-5v9ql" podUID="628627fd-f89d-4b0b-af23-27eb9101b7cb" Jan 24 11:51:27.158753 kubelet[2895]: E0124 11:51:27.142598 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:27.158753 kubelet[2895]: E0124 11:51:27.143113 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:27.159885 containerd[1647]: time="2026-01-24T11:51:27.147738144Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f485b4455-qhcml,Uid:c28e9f44-9bc9-4328-bf7a-f4f9d016a87e,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:51:27.159885 containerd[1647]: time="2026-01-24T11:51:27.148674079Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-dnscb,Uid:7a087061-e7ac-47a8-a3c4-b753efe3e78f,Namespace:kube-system,Attempt:0,}" Jan 24 11:51:27.159885 containerd[1647]: time="2026-01-24T11:51:27.149169624Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-k9kt6,Uid:f46cf494-de53-4117-858a-b0e56c0ba7ac,Namespace:kube-system,Attempt:0,}" Jan 24 11:51:27.159885 containerd[1647]: time="2026-01-24T11:51:27.149244953Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-66955b7fc7-xmscg,Uid:41fe2d74-dafe-45a2-82d2-5fd56447e59d,Namespace:calico-system,Attempt:0,}" Jan 24 11:51:27.448584 containerd[1647]: time="2026-01-24T11:51:27.447888065Z" level=error msg="Failed to destroy network for sandbox \"b3cf5fe4ad03a55ab60a3773acf2fe7e12a4cab11b4dfac7a304598b4c0876b1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:27.453404 systemd[1]: run-netns-cni\x2d097bd65c\x2d833b\x2d0d88\x2ddb60\x2d55596fc263aa.mount: Deactivated successfully. Jan 24 11:51:27.454716 containerd[1647]: time="2026-01-24T11:51:27.454624588Z" level=error msg="Failed to destroy network for sandbox \"2b9cce8b4c78e2717739155ff5599c772f57cfa6398e2b54fbe70033d1c75855\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:27.456258 containerd[1647]: time="2026-01-24T11:51:27.456146267Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-k9kt6,Uid:f46cf494-de53-4117-858a-b0e56c0ba7ac,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"b3cf5fe4ad03a55ab60a3773acf2fe7e12a4cab11b4dfac7a304598b4c0876b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:27.456618 kubelet[2895]: E0124 11:51:27.456492 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b3cf5fe4ad03a55ab60a3773acf2fe7e12a4cab11b4dfac7a304598b4c0876b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:27.456717 kubelet[2895]: E0124 11:51:27.456641 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b3cf5fe4ad03a55ab60a3773acf2fe7e12a4cab11b4dfac7a304598b4c0876b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-k9kt6" Jan 24 11:51:27.456717 kubelet[2895]: E0124 11:51:27.456674 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b3cf5fe4ad03a55ab60a3773acf2fe7e12a4cab11b4dfac7a304598b4c0876b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-k9kt6" Jan 24 11:51:27.456776 kubelet[2895]: E0124 11:51:27.456739 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-k9kt6_kube-system(f46cf494-de53-4117-858a-b0e56c0ba7ac)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-k9kt6_kube-system(f46cf494-de53-4117-858a-b0e56c0ba7ac)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b3cf5fe4ad03a55ab60a3773acf2fe7e12a4cab11b4dfac7a304598b4c0876b1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-k9kt6" podUID="f46cf494-de53-4117-858a-b0e56c0ba7ac" Jan 24 11:51:27.466884 containerd[1647]: time="2026-01-24T11:51:27.466584345Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f485b4455-qhcml,Uid:c28e9f44-9bc9-4328-bf7a-f4f9d016a87e,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"2b9cce8b4c78e2717739155ff5599c772f57cfa6398e2b54fbe70033d1c75855\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:27.469349 kubelet[2895]: E0124 11:51:27.469243 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2b9cce8b4c78e2717739155ff5599c772f57cfa6398e2b54fbe70033d1c75855\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:27.469790 kubelet[2895]: E0124 11:51:27.469461 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2b9cce8b4c78e2717739155ff5599c772f57cfa6398e2b54fbe70033d1c75855\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" Jan 24 11:51:27.469790 kubelet[2895]: E0124 11:51:27.469489 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2b9cce8b4c78e2717739155ff5599c772f57cfa6398e2b54fbe70033d1c75855\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" Jan 24 11:51:27.469790 kubelet[2895]: E0124 11:51:27.469592 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-f485b4455-qhcml_calico-apiserver(c28e9f44-9bc9-4328-bf7a-f4f9d016a87e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-f485b4455-qhcml_calico-apiserver(c28e9f44-9bc9-4328-bf7a-f4f9d016a87e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2b9cce8b4c78e2717739155ff5599c772f57cfa6398e2b54fbe70033d1c75855\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:51:27.526483 containerd[1647]: time="2026-01-24T11:51:27.525999647Z" level=error msg="Failed to destroy network for sandbox \"ab02ec50565604891d3c5166385470b90ce634326da13a944567b70dcb53287c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:27.526483 containerd[1647]: time="2026-01-24T11:51:27.526225248Z" level=error msg="Failed to destroy network for sandbox \"781766c37da8e8ad87fd28fda78d5726b2895e4eb3abc243fa4e255a09fdd3ed\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:27.537771 containerd[1647]: time="2026-01-24T11:51:27.537637207Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-66955b7fc7-xmscg,Uid:41fe2d74-dafe-45a2-82d2-5fd56447e59d,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ab02ec50565604891d3c5166385470b90ce634326da13a944567b70dcb53287c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:27.539899 kubelet[2895]: E0124 11:51:27.539801 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ab02ec50565604891d3c5166385470b90ce634326da13a944567b70dcb53287c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:27.543352 kubelet[2895]: E0124 11:51:27.543251 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ab02ec50565604891d3c5166385470b90ce634326da13a944567b70dcb53287c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" Jan 24 11:51:27.547585 kubelet[2895]: E0124 11:51:27.543485 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ab02ec50565604891d3c5166385470b90ce634326da13a944567b70dcb53287c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" Jan 24 11:51:27.547696 containerd[1647]: time="2026-01-24T11:51:27.543958298Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-dnscb,Uid:7a087061-e7ac-47a8-a3c4-b753efe3e78f,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"781766c37da8e8ad87fd28fda78d5726b2895e4eb3abc243fa4e255a09fdd3ed\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:27.547957 kubelet[2895]: E0124 11:51:27.547362 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-66955b7fc7-xmscg_calico-system(41fe2d74-dafe-45a2-82d2-5fd56447e59d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-66955b7fc7-xmscg_calico-system(41fe2d74-dafe-45a2-82d2-5fd56447e59d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ab02ec50565604891d3c5166385470b90ce634326da13a944567b70dcb53287c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:51:27.550642 kubelet[2895]: E0124 11:51:27.550014 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"781766c37da8e8ad87fd28fda78d5726b2895e4eb3abc243fa4e255a09fdd3ed\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:27.550642 kubelet[2895]: E0124 11:51:27.550066 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"781766c37da8e8ad87fd28fda78d5726b2895e4eb3abc243fa4e255a09fdd3ed\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-dnscb" Jan 24 11:51:27.550642 kubelet[2895]: E0124 11:51:27.550091 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"781766c37da8e8ad87fd28fda78d5726b2895e4eb3abc243fa4e255a09fdd3ed\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-dnscb" Jan 24 11:51:27.550806 kubelet[2895]: E0124 11:51:27.550141 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-dnscb_kube-system(7a087061-e7ac-47a8-a3c4-b753efe3e78f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-dnscb_kube-system(7a087061-e7ac-47a8-a3c4-b753efe3e78f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"781766c37da8e8ad87fd28fda78d5726b2895e4eb3abc243fa4e255a09fdd3ed\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-dnscb" podUID="7a087061-e7ac-47a8-a3c4-b753efe3e78f" Jan 24 11:51:28.132171 containerd[1647]: time="2026-01-24T11:51:28.132081863Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-2g45n,Uid:999c48c5-f3b1-4bfd-91f4-26cafee1eef0,Namespace:calico-system,Attempt:0,}" Jan 24 11:51:28.240384 systemd[1]: run-netns-cni\x2d1b9ed891\x2de73a\x2df96d\x2d2493\x2d1ede628a450b.mount: Deactivated successfully. Jan 24 11:51:28.240902 systemd[1]: run-netns-cni\x2df7c94ba0\x2d3518\x2d8bbe\x2d31eb\x2d1fcbea628811.mount: Deactivated successfully. Jan 24 11:51:28.241146 systemd[1]: run-netns-cni\x2d7ac465cd\x2d24dc\x2dc7df\x2dcbbe\x2d312a23056365.mount: Deactivated successfully. Jan 24 11:51:28.381406 containerd[1647]: time="2026-01-24T11:51:28.376877294Z" level=error msg="Failed to destroy network for sandbox \"72931dd6edaeccea370dbc2696b4ebf267c318f1e4069a19a8e34d99122565d9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:28.384728 systemd[1]: run-netns-cni\x2ddf0c1cdd\x2d1cd9\x2d07ae\x2d53cd\x2dbc24a75f0716.mount: Deactivated successfully. Jan 24 11:51:28.391927 containerd[1647]: time="2026-01-24T11:51:28.391712830Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-2g45n,Uid:999c48c5-f3b1-4bfd-91f4-26cafee1eef0,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"72931dd6edaeccea370dbc2696b4ebf267c318f1e4069a19a8e34d99122565d9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:28.393075 kubelet[2895]: E0124 11:51:28.392965 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"72931dd6edaeccea370dbc2696b4ebf267c318f1e4069a19a8e34d99122565d9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:28.393715 kubelet[2895]: E0124 11:51:28.393168 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"72931dd6edaeccea370dbc2696b4ebf267c318f1e4069a19a8e34d99122565d9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-2g45n" Jan 24 11:51:28.393715 kubelet[2895]: E0124 11:51:28.393205 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"72931dd6edaeccea370dbc2696b4ebf267c318f1e4069a19a8e34d99122565d9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-2g45n" Jan 24 11:51:28.394567 kubelet[2895]: E0124 11:51:28.394189 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-2g45n_calico-system(999c48c5-f3b1-4bfd-91f4-26cafee1eef0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-2g45n_calico-system(999c48c5-f3b1-4bfd-91f4-26cafee1eef0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"72931dd6edaeccea370dbc2696b4ebf267c318f1e4069a19a8e34d99122565d9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:51:29.140006 containerd[1647]: time="2026-01-24T11:51:29.135592989Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f485b4455-nc2lh,Uid:d47a09e2-52ab-4e49-942e-05f2cedb67ae,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:51:29.140912 containerd[1647]: time="2026-01-24T11:51:29.140872139Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-jfngl,Uid:293f1d4a-795c-44fc-bda8-3c550f56df72,Namespace:calico-system,Attempt:0,}" Jan 24 11:51:29.450452 containerd[1647]: time="2026-01-24T11:51:29.448918877Z" level=error msg="Failed to destroy network for sandbox \"f5e2f3d9cede61ba78a36e8100eda0e28dc990425c983bf0ae9f308eeb3c5f73\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:29.456718 systemd[1]: run-netns-cni\x2da08c6f3b\x2dc853\x2d75fd\x2de21b\x2d28a2823f41b9.mount: Deactivated successfully. Jan 24 11:51:29.476032 containerd[1647]: time="2026-01-24T11:51:29.469207962Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f485b4455-nc2lh,Uid:d47a09e2-52ab-4e49-942e-05f2cedb67ae,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"f5e2f3d9cede61ba78a36e8100eda0e28dc990425c983bf0ae9f308eeb3c5f73\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:29.476651 kubelet[2895]: E0124 11:51:29.471577 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f5e2f3d9cede61ba78a36e8100eda0e28dc990425c983bf0ae9f308eeb3c5f73\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:29.476651 kubelet[2895]: E0124 11:51:29.471653 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f5e2f3d9cede61ba78a36e8100eda0e28dc990425c983bf0ae9f308eeb3c5f73\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" Jan 24 11:51:29.476651 kubelet[2895]: E0124 11:51:29.471681 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f5e2f3d9cede61ba78a36e8100eda0e28dc990425c983bf0ae9f308eeb3c5f73\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" Jan 24 11:51:29.479649 kubelet[2895]: E0124 11:51:29.471739 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-f485b4455-nc2lh_calico-apiserver(d47a09e2-52ab-4e49-942e-05f2cedb67ae)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-f485b4455-nc2lh_calico-apiserver(d47a09e2-52ab-4e49-942e-05f2cedb67ae)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f5e2f3d9cede61ba78a36e8100eda0e28dc990425c983bf0ae9f308eeb3c5f73\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:51:29.526368 containerd[1647]: time="2026-01-24T11:51:29.525054861Z" level=error msg="Failed to destroy network for sandbox \"07057c18003e49b2a277548bb8cabe17ca2ebbb6143598e0a3173d79168a361c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:29.533670 systemd[1]: run-netns-cni\x2d1409ff6e\x2d4ae2\x2dd453\x2d2b76\x2d65e149f2ff72.mount: Deactivated successfully. Jan 24 11:51:29.549729 containerd[1647]: time="2026-01-24T11:51:29.545849589Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-jfngl,Uid:293f1d4a-795c-44fc-bda8-3c550f56df72,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"07057c18003e49b2a277548bb8cabe17ca2ebbb6143598e0a3173d79168a361c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:29.550022 kubelet[2895]: E0124 11:51:29.546428 2895 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"07057c18003e49b2a277548bb8cabe17ca2ebbb6143598e0a3173d79168a361c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 24 11:51:29.550022 kubelet[2895]: E0124 11:51:29.546510 2895 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"07057c18003e49b2a277548bb8cabe17ca2ebbb6143598e0a3173d79168a361c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-jfngl" Jan 24 11:51:29.550022 kubelet[2895]: E0124 11:51:29.546578 2895 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"07057c18003e49b2a277548bb8cabe17ca2ebbb6143598e0a3173d79168a361c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-jfngl" Jan 24 11:51:29.550201 kubelet[2895]: E0124 11:51:29.546647 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-jfngl_calico-system(293f1d4a-795c-44fc-bda8-3c550f56df72)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-jfngl_calico-system(293f1d4a-795c-44fc-bda8-3c550f56df72)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"07057c18003e49b2a277548bb8cabe17ca2ebbb6143598e0a3173d79168a361c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:51:31.101735 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3990517324.mount: Deactivated successfully. Jan 24 11:51:31.181660 containerd[1647]: time="2026-01-24T11:51:31.181462137Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:51:31.183245 containerd[1647]: time="2026-01-24T11:51:31.183140518Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.4: active requests=0, bytes read=156880025" Jan 24 11:51:31.185832 containerd[1647]: time="2026-01-24T11:51:31.185358305Z" level=info msg="ImageCreate event name:\"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:51:31.192250 containerd[1647]: time="2026-01-24T11:51:31.192178726Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 24 11:51:31.193583 containerd[1647]: time="2026-01-24T11:51:31.192908569Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.4\" with image id \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\", size \"156883537\" in 17.904087769s" Jan 24 11:51:31.193583 containerd[1647]: time="2026-01-24T11:51:31.192949676Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\" returns image reference \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\"" Jan 24 11:51:31.250962 containerd[1647]: time="2026-01-24T11:51:31.250905160Z" level=info msg="CreateContainer within sandbox \"2c075fdffb700b0c1172313d3ff66de2f4728237f591f9d6d03d9c3131fb8564\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Jan 24 11:51:31.278645 containerd[1647]: time="2026-01-24T11:51:31.278506139Z" level=info msg="Container 984970cf2ea7928e62b4ac6ddd8d512268cb6b7383b67d99807d264d6dfa09b9: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:51:31.314249 containerd[1647]: time="2026-01-24T11:51:31.298121179Z" level=info msg="CreateContainer within sandbox \"2c075fdffb700b0c1172313d3ff66de2f4728237f591f9d6d03d9c3131fb8564\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"984970cf2ea7928e62b4ac6ddd8d512268cb6b7383b67d99807d264d6dfa09b9\"" Jan 24 11:51:31.336926 containerd[1647]: time="2026-01-24T11:51:31.333499185Z" level=info msg="StartContainer for \"984970cf2ea7928e62b4ac6ddd8d512268cb6b7383b67d99807d264d6dfa09b9\"" Jan 24 11:51:31.345313 containerd[1647]: time="2026-01-24T11:51:31.345151508Z" level=info msg="connecting to shim 984970cf2ea7928e62b4ac6ddd8d512268cb6b7383b67d99807d264d6dfa09b9" address="unix:///run/containerd/s/5856dbc4464bd7263ae0bc1ab234b2764d81a51e8bf8fc1354a687699a56c458" protocol=ttrpc version=3 Jan 24 11:51:31.414843 systemd[1]: Started cri-containerd-984970cf2ea7928e62b4ac6ddd8d512268cb6b7383b67d99807d264d6dfa09b9.scope - libcontainer container 984970cf2ea7928e62b4ac6ddd8d512268cb6b7383b67d99807d264d6dfa09b9. Jan 24 11:51:31.551000 audit: BPF prog-id=175 op=LOAD Jan 24 11:51:31.562877 kernel: kauditd_printk_skb: 6 callbacks suppressed Jan 24 11:51:31.566723 kernel: audit: type=1334 audit(1769255491.551:579): prog-id=175 op=LOAD Jan 24 11:51:31.566928 kernel: audit: type=1300 audit(1769255491.551:579): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=3491 pid=4289 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:31.551000 audit[4289]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=3491 pid=4289 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:31.551000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3938343937306366326561373932386536326234616336646464386435 Jan 24 11:51:31.594926 kernel: audit: type=1327 audit(1769255491.551:579): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3938343937306366326561373932386536326234616336646464386435 Jan 24 11:51:31.551000 audit: BPF prog-id=176 op=LOAD Jan 24 11:51:31.598992 kernel: audit: type=1334 audit(1769255491.551:580): prog-id=176 op=LOAD Jan 24 11:51:31.599063 kernel: audit: type=1300 audit(1769255491.551:580): arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=3491 pid=4289 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:31.551000 audit[4289]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=3491 pid=4289 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:31.623811 kernel: audit: type=1327 audit(1769255491.551:580): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3938343937306366326561373932386536326234616336646464386435 Jan 24 11:51:31.551000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3938343937306366326561373932386536326234616336646464386435 Jan 24 11:51:31.551000 audit: BPF prog-id=176 op=UNLOAD Jan 24 11:51:31.644527 kernel: audit: type=1334 audit(1769255491.551:581): prog-id=176 op=UNLOAD Jan 24 11:51:31.644649 kernel: audit: type=1300 audit(1769255491.551:581): arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3491 pid=4289 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:31.551000 audit[4289]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3491 pid=4289 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:31.551000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3938343937306366326561373932386536326234616336646464386435 Jan 24 11:51:31.682969 kernel: audit: type=1327 audit(1769255491.551:581): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3938343937306366326561373932386536326234616336646464386435 Jan 24 11:51:31.551000 audit: BPF prog-id=175 op=UNLOAD Jan 24 11:51:31.687326 kernel: audit: type=1334 audit(1769255491.551:582): prog-id=175 op=UNLOAD Jan 24 11:51:31.551000 audit[4289]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3491 pid=4289 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:31.551000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3938343937306366326561373932386536326234616336646464386435 Jan 24 11:51:31.551000 audit: BPF prog-id=177 op=LOAD Jan 24 11:51:31.551000 audit[4289]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=3491 pid=4289 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:31.551000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3938343937306366326561373932386536326234616336646464386435 Jan 24 11:51:31.731233 containerd[1647]: time="2026-01-24T11:51:31.731108706Z" level=info msg="StartContainer for \"984970cf2ea7928e62b4ac6ddd8d512268cb6b7383b67d99807d264d6dfa09b9\" returns successfully" Jan 24 11:51:32.384769 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Jan 24 11:51:32.385116 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Jan 24 11:51:32.666674 kubelet[2895]: E0124 11:51:32.663675 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:33.188366 kubelet[2895]: I0124 11:51:33.187883 2895 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-878v5" podStartSLOduration=4.341043968 podStartE2EDuration="41.187798148s" podCreationTimestamp="2026-01-24 11:50:52 +0000 UTC" firstStartedPulling="2026-01-24 11:50:54.367744683 +0000 UTC m=+33.889537389" lastFinishedPulling="2026-01-24 11:51:31.214498862 +0000 UTC m=+70.736291569" observedRunningTime="2026-01-24 11:51:33.183136178 +0000 UTC m=+72.704928965" watchObservedRunningTime="2026-01-24 11:51:33.187798148 +0000 UTC m=+72.709590854" Jan 24 11:51:33.664713 kubelet[2895]: E0124 11:51:33.663999 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:33.841677 kubelet[2895]: I0124 11:51:33.838741 2895 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/628627fd-f89d-4b0b-af23-27eb9101b7cb-whisker-backend-key-pair\") pod \"628627fd-f89d-4b0b-af23-27eb9101b7cb\" (UID: \"628627fd-f89d-4b0b-af23-27eb9101b7cb\") " Jan 24 11:51:33.841677 kubelet[2895]: I0124 11:51:33.841403 2895 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/628627fd-f89d-4b0b-af23-27eb9101b7cb-whisker-ca-bundle\") pod \"628627fd-f89d-4b0b-af23-27eb9101b7cb\" (UID: \"628627fd-f89d-4b0b-af23-27eb9101b7cb\") " Jan 24 11:51:33.841677 kubelet[2895]: I0124 11:51:33.841617 2895 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b48zv\" (UniqueName: \"kubernetes.io/projected/628627fd-f89d-4b0b-af23-27eb9101b7cb-kube-api-access-b48zv\") pod \"628627fd-f89d-4b0b-af23-27eb9101b7cb\" (UID: \"628627fd-f89d-4b0b-af23-27eb9101b7cb\") " Jan 24 11:51:33.849778 kubelet[2895]: I0124 11:51:33.849715 2895 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/628627fd-f89d-4b0b-af23-27eb9101b7cb-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "628627fd-f89d-4b0b-af23-27eb9101b7cb" (UID: "628627fd-f89d-4b0b-af23-27eb9101b7cb"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Jan 24 11:51:33.886591 kubelet[2895]: I0124 11:51:33.886325 2895 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/628627fd-f89d-4b0b-af23-27eb9101b7cb-kube-api-access-b48zv" (OuterVolumeSpecName: "kube-api-access-b48zv") pod "628627fd-f89d-4b0b-af23-27eb9101b7cb" (UID: "628627fd-f89d-4b0b-af23-27eb9101b7cb"). InnerVolumeSpecName "kube-api-access-b48zv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Jan 24 11:51:33.888455 systemd[1]: var-lib-kubelet-pods-628627fd\x2df89d\x2d4b0b\x2daf23\x2d27eb9101b7cb-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2db48zv.mount: Deactivated successfully. Jan 24 11:51:33.889547 systemd[1]: var-lib-kubelet-pods-628627fd\x2df89d\x2d4b0b\x2daf23\x2d27eb9101b7cb-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Jan 24 11:51:33.892671 kubelet[2895]: I0124 11:51:33.889018 2895 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/628627fd-f89d-4b0b-af23-27eb9101b7cb-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "628627fd-f89d-4b0b-af23-27eb9101b7cb" (UID: "628627fd-f89d-4b0b-af23-27eb9101b7cb"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Jan 24 11:51:33.948638 kubelet[2895]: I0124 11:51:33.947947 2895 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-b48zv\" (UniqueName: \"kubernetes.io/projected/628627fd-f89d-4b0b-af23-27eb9101b7cb-kube-api-access-b48zv\") on node \"localhost\" DevicePath \"\"" Jan 24 11:51:33.948638 kubelet[2895]: I0124 11:51:33.947991 2895 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/628627fd-f89d-4b0b-af23-27eb9101b7cb-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Jan 24 11:51:33.948638 kubelet[2895]: I0124 11:51:33.948007 2895 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/628627fd-f89d-4b0b-af23-27eb9101b7cb-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Jan 24 11:51:34.745946 systemd[1]: Removed slice kubepods-besteffort-pod628627fd_f89d_4b0b_af23_27eb9101b7cb.slice - libcontainer container kubepods-besteffort-pod628627fd_f89d_4b0b_af23_27eb9101b7cb.slice. Jan 24 11:51:35.016795 systemd[1]: Created slice kubepods-besteffort-pod5cdab4ea_dcce_4467_b450_26abe2130c8d.slice - libcontainer container kubepods-besteffort-pod5cdab4ea_dcce_4467_b450_26abe2130c8d.slice. Jan 24 11:51:35.122684 kubelet[2895]: I0124 11:51:35.121874 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cdab4ea-dcce-4467-b450-26abe2130c8d-whisker-ca-bundle\") pod \"whisker-5c99cdc8ff-zh5gl\" (UID: \"5cdab4ea-dcce-4467-b450-26abe2130c8d\") " pod="calico-system/whisker-5c99cdc8ff-zh5gl" Jan 24 11:51:35.122684 kubelet[2895]: I0124 11:51:35.122039 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/5cdab4ea-dcce-4467-b450-26abe2130c8d-whisker-backend-key-pair\") pod \"whisker-5c99cdc8ff-zh5gl\" (UID: \"5cdab4ea-dcce-4467-b450-26abe2130c8d\") " pod="calico-system/whisker-5c99cdc8ff-zh5gl" Jan 24 11:51:35.141446 kubelet[2895]: I0124 11:51:35.141402 2895 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="628627fd-f89d-4b0b-af23-27eb9101b7cb" path="/var/lib/kubelet/pods/628627fd-f89d-4b0b-af23-27eb9101b7cb/volumes" Jan 24 11:51:35.228082 kubelet[2895]: I0124 11:51:35.224615 2895 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n6cl\" (UniqueName: \"kubernetes.io/projected/5cdab4ea-dcce-4467-b450-26abe2130c8d-kube-api-access-5n6cl\") pod \"whisker-5c99cdc8ff-zh5gl\" (UID: \"5cdab4ea-dcce-4467-b450-26abe2130c8d\") " pod="calico-system/whisker-5c99cdc8ff-zh5gl" Jan 24 11:51:35.435717 containerd[1647]: time="2026-01-24T11:51:35.434353821Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5c99cdc8ff-zh5gl,Uid:5cdab4ea-dcce-4467-b450-26abe2130c8d,Namespace:calico-system,Attempt:0,}" Jan 24 11:51:36.230000 audit: BPF prog-id=178 op=LOAD Jan 24 11:51:36.230000 audit[4558]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffeb6481880 a2=98 a3=1fffffffffffffff items=0 ppid=4422 pid=4558 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.230000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 24 11:51:36.231000 audit: BPF prog-id=178 op=UNLOAD Jan 24 11:51:36.231000 audit[4558]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffeb6481850 a3=0 items=0 ppid=4422 pid=4558 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.231000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 24 11:51:36.231000 audit: BPF prog-id=179 op=LOAD Jan 24 11:51:36.231000 audit[4558]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffeb6481760 a2=94 a3=3 items=0 ppid=4422 pid=4558 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.231000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 24 11:51:36.232000 audit: BPF prog-id=179 op=UNLOAD Jan 24 11:51:36.232000 audit[4558]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffeb6481760 a2=94 a3=3 items=0 ppid=4422 pid=4558 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.232000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 24 11:51:36.232000 audit: BPF prog-id=180 op=LOAD Jan 24 11:51:36.232000 audit[4558]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffeb64817a0 a2=94 a3=7ffeb6481980 items=0 ppid=4422 pid=4558 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.232000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 24 11:51:36.232000 audit: BPF prog-id=180 op=UNLOAD Jan 24 11:51:36.232000 audit[4558]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffeb64817a0 a2=94 a3=7ffeb6481980 items=0 ppid=4422 pid=4558 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.232000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 24 11:51:36.244000 audit: BPF prog-id=181 op=LOAD Jan 24 11:51:36.244000 audit[4559]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7fff4f898db0 a2=98 a3=3 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.244000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.244000 audit: BPF prog-id=181 op=UNLOAD Jan 24 11:51:36.244000 audit[4559]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7fff4f898d80 a3=0 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.244000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.245000 audit: BPF prog-id=182 op=LOAD Jan 24 11:51:36.245000 audit[4559]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fff4f898ba0 a2=94 a3=54428f items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.245000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.245000 audit: BPF prog-id=182 op=UNLOAD Jan 24 11:51:36.245000 audit[4559]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7fff4f898ba0 a2=94 a3=54428f items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.245000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.245000 audit: BPF prog-id=183 op=LOAD Jan 24 11:51:36.245000 audit[4559]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fff4f898bd0 a2=94 a3=2 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.245000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.245000 audit: BPF prog-id=183 op=UNLOAD Jan 24 11:51:36.245000 audit[4559]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7fff4f898bd0 a2=0 a3=2 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.245000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.451219 systemd-networkd[1544]: calicf40c7cce41: Link UP Jan 24 11:51:36.453437 systemd-networkd[1544]: calicf40c7cce41: Gained carrier Jan 24 11:51:36.489621 containerd[1647]: 2026-01-24 11:51:35.668 [INFO][4460] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 24 11:51:36.489621 containerd[1647]: 2026-01-24 11:51:35.816 [INFO][4460] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--5c99cdc8ff--zh5gl-eth0 whisker-5c99cdc8ff- calico-system 5cdab4ea-dcce-4467-b450-26abe2130c8d 1074 0 2026-01-24 11:51:34 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:5c99cdc8ff projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-5c99cdc8ff-zh5gl eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calicf40c7cce41 [] [] }} ContainerID="7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" Namespace="calico-system" Pod="whisker-5c99cdc8ff-zh5gl" WorkloadEndpoint="localhost-k8s-whisker--5c99cdc8ff--zh5gl-" Jan 24 11:51:36.489621 containerd[1647]: 2026-01-24 11:51:35.817 [INFO][4460] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" Namespace="calico-system" Pod="whisker-5c99cdc8ff-zh5gl" WorkloadEndpoint="localhost-k8s-whisker--5c99cdc8ff--zh5gl-eth0" Jan 24 11:51:36.489621 containerd[1647]: 2026-01-24 11:51:36.250 [INFO][4519] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" HandleID="k8s-pod-network.7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" Workload="localhost-k8s-whisker--5c99cdc8ff--zh5gl-eth0" Jan 24 11:51:36.494017 containerd[1647]: 2026-01-24 11:51:36.253 [INFO][4519] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" HandleID="k8s-pod-network.7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" Workload="localhost-k8s-whisker--5c99cdc8ff--zh5gl-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0000bf000), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-5c99cdc8ff-zh5gl", "timestamp":"2026-01-24 11:51:36.250416235 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:51:36.494017 containerd[1647]: 2026-01-24 11:51:36.253 [INFO][4519] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:51:36.494017 containerd[1647]: 2026-01-24 11:51:36.255 [INFO][4519] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:51:36.494017 containerd[1647]: 2026-01-24 11:51:36.256 [INFO][4519] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:51:36.494017 containerd[1647]: 2026-01-24 11:51:36.289 [INFO][4519] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" host="localhost" Jan 24 11:51:36.494017 containerd[1647]: 2026-01-24 11:51:36.321 [INFO][4519] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:51:36.494017 containerd[1647]: 2026-01-24 11:51:36.334 [INFO][4519] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:51:36.494017 containerd[1647]: 2026-01-24 11:51:36.344 [INFO][4519] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:36.494017 containerd[1647]: 2026-01-24 11:51:36.349 [INFO][4519] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:36.494017 containerd[1647]: 2026-01-24 11:51:36.350 [INFO][4519] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" host="localhost" Jan 24 11:51:36.494524 containerd[1647]: 2026-01-24 11:51:36.354 [INFO][4519] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793 Jan 24 11:51:36.494524 containerd[1647]: 2026-01-24 11:51:36.369 [INFO][4519] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" host="localhost" Jan 24 11:51:36.494524 containerd[1647]: 2026-01-24 11:51:36.395 [INFO][4519] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" host="localhost" Jan 24 11:51:36.494524 containerd[1647]: 2026-01-24 11:51:36.395 [INFO][4519] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" host="localhost" Jan 24 11:51:36.494524 containerd[1647]: 2026-01-24 11:51:36.395 [INFO][4519] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:51:36.494524 containerd[1647]: 2026-01-24 11:51:36.395 [INFO][4519] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" HandleID="k8s-pod-network.7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" Workload="localhost-k8s-whisker--5c99cdc8ff--zh5gl-eth0" Jan 24 11:51:36.494757 containerd[1647]: 2026-01-24 11:51:36.415 [INFO][4460] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" Namespace="calico-system" Pod="whisker-5c99cdc8ff-zh5gl" WorkloadEndpoint="localhost-k8s-whisker--5c99cdc8ff--zh5gl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--5c99cdc8ff--zh5gl-eth0", GenerateName:"whisker-5c99cdc8ff-", Namespace:"calico-system", SelfLink:"", UID:"5cdab4ea-dcce-4467-b450-26abe2130c8d", ResourceVersion:"1074", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 51, 34, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"5c99cdc8ff", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-5c99cdc8ff-zh5gl", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calicf40c7cce41", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:36.494757 containerd[1647]: 2026-01-24 11:51:36.415 [INFO][4460] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" Namespace="calico-system" Pod="whisker-5c99cdc8ff-zh5gl" WorkloadEndpoint="localhost-k8s-whisker--5c99cdc8ff--zh5gl-eth0" Jan 24 11:51:36.494961 containerd[1647]: 2026-01-24 11:51:36.415 [INFO][4460] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calicf40c7cce41 ContainerID="7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" Namespace="calico-system" Pod="whisker-5c99cdc8ff-zh5gl" WorkloadEndpoint="localhost-k8s-whisker--5c99cdc8ff--zh5gl-eth0" Jan 24 11:51:36.494961 containerd[1647]: 2026-01-24 11:51:36.457 [INFO][4460] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" Namespace="calico-system" Pod="whisker-5c99cdc8ff-zh5gl" WorkloadEndpoint="localhost-k8s-whisker--5c99cdc8ff--zh5gl-eth0" Jan 24 11:51:36.495049 containerd[1647]: 2026-01-24 11:51:36.459 [INFO][4460] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" Namespace="calico-system" Pod="whisker-5c99cdc8ff-zh5gl" WorkloadEndpoint="localhost-k8s-whisker--5c99cdc8ff--zh5gl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--5c99cdc8ff--zh5gl-eth0", GenerateName:"whisker-5c99cdc8ff-", Namespace:"calico-system", SelfLink:"", UID:"5cdab4ea-dcce-4467-b450-26abe2130c8d", ResourceVersion:"1074", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 51, 34, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"5c99cdc8ff", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793", Pod:"whisker-5c99cdc8ff-zh5gl", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calicf40c7cce41", MAC:"26:56:25:36:a2:5d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:36.495165 containerd[1647]: 2026-01-24 11:51:36.478 [INFO][4460] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" Namespace="calico-system" Pod="whisker-5c99cdc8ff-zh5gl" WorkloadEndpoint="localhost-k8s-whisker--5c99cdc8ff--zh5gl-eth0" Jan 24 11:51:36.589000 audit: BPF prog-id=184 op=LOAD Jan 24 11:51:36.595013 kernel: kauditd_printk_skb: 41 callbacks suppressed Jan 24 11:51:36.595181 kernel: audit: type=1334 audit(1769255496.589:596): prog-id=184 op=LOAD Jan 24 11:51:36.589000 audit[4559]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fff4f898a90 a2=94 a3=1 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.617886 kernel: audit: type=1300 audit(1769255496.589:596): arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fff4f898a90 a2=94 a3=1 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.589000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.590000 audit: BPF prog-id=184 op=UNLOAD Jan 24 11:51:36.627746 kernel: audit: type=1327 audit(1769255496.589:596): proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.627799 kernel: audit: type=1334 audit(1769255496.590:597): prog-id=184 op=UNLOAD Jan 24 11:51:36.590000 audit[4559]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7fff4f898a90 a2=94 a3=1 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.643996 kernel: audit: type=1300 audit(1769255496.590:597): arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7fff4f898a90 a2=94 a3=1 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.648542 kernel: audit: type=1327 audit(1769255496.590:597): proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.590000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.650408 kernel: audit: type=1334 audit(1769255496.603:598): prog-id=185 op=LOAD Jan 24 11:51:36.603000 audit: BPF prog-id=185 op=LOAD Jan 24 11:51:36.603000 audit[4559]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7fff4f898a80 a2=94 a3=4 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.603000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.670987 kernel: audit: type=1300 audit(1769255496.603:598): arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7fff4f898a80 a2=94 a3=4 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.671083 kernel: audit: type=1327 audit(1769255496.603:598): proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.671136 kernel: audit: type=1334 audit(1769255496.603:599): prog-id=185 op=UNLOAD Jan 24 11:51:36.603000 audit: BPF prog-id=185 op=UNLOAD Jan 24 11:51:36.603000 audit[4559]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7fff4f898a80 a2=0 a3=4 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.603000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.604000 audit: BPF prog-id=186 op=LOAD Jan 24 11:51:36.604000 audit[4559]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fff4f8988e0 a2=94 a3=5 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.604000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.604000 audit: BPF prog-id=186 op=UNLOAD Jan 24 11:51:36.604000 audit[4559]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7fff4f8988e0 a2=0 a3=5 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.604000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.604000 audit: BPF prog-id=187 op=LOAD Jan 24 11:51:36.604000 audit[4559]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7fff4f898b00 a2=94 a3=6 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.604000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.605000 audit: BPF prog-id=187 op=UNLOAD Jan 24 11:51:36.605000 audit[4559]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7fff4f898b00 a2=0 a3=6 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.605000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.605000 audit: BPF prog-id=188 op=LOAD Jan 24 11:51:36.605000 audit[4559]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7fff4f8982b0 a2=94 a3=88 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.605000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.610000 audit: BPF prog-id=189 op=LOAD Jan 24 11:51:36.610000 audit[4559]: SYSCALL arch=c000003e syscall=321 success=yes exit=7 a0=5 a1=7fff4f898130 a2=94 a3=2 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.610000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.610000 audit: BPF prog-id=189 op=UNLOAD Jan 24 11:51:36.610000 audit[4559]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=7 a1=7fff4f898160 a2=0 a3=7fff4f898260 items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.610000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.612000 audit: BPF prog-id=188 op=UNLOAD Jan 24 11:51:36.612000 audit[4559]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=f7c9d10 a2=0 a3=3725f38418b9c28d items=0 ppid=4422 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.612000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 24 11:51:36.632000 audit: BPF prog-id=190 op=LOAD Jan 24 11:51:36.632000 audit[4575]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffc7703bee0 a2=98 a3=1999999999999999 items=0 ppid=4422 pid=4575 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.632000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 24 11:51:36.632000 audit: BPF prog-id=190 op=UNLOAD Jan 24 11:51:36.632000 audit[4575]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffc7703beb0 a3=0 items=0 ppid=4422 pid=4575 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.632000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 24 11:51:36.632000 audit: BPF prog-id=191 op=LOAD Jan 24 11:51:36.632000 audit[4575]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffc7703bdc0 a2=94 a3=ffff items=0 ppid=4422 pid=4575 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.632000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 24 11:51:36.632000 audit: BPF prog-id=191 op=UNLOAD Jan 24 11:51:36.632000 audit[4575]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffc7703bdc0 a2=94 a3=ffff items=0 ppid=4422 pid=4575 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.632000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 24 11:51:36.632000 audit: BPF prog-id=192 op=LOAD Jan 24 11:51:36.632000 audit[4575]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffc7703be00 a2=94 a3=7ffc7703bfe0 items=0 ppid=4422 pid=4575 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.632000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 24 11:51:36.632000 audit: BPF prog-id=192 op=UNLOAD Jan 24 11:51:36.632000 audit[4575]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffc7703be00 a2=94 a3=7ffc7703bfe0 items=0 ppid=4422 pid=4575 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:36.632000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 24 11:51:36.776655 containerd[1647]: time="2026-01-24T11:51:36.769723916Z" level=info msg="connecting to shim 7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793" address="unix:///run/containerd/s/e67f60ef2d7763857dcd7dba98201868499209bd99eab782210a4b87952187fc" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:51:37.007753 systemd-networkd[1544]: vxlan.calico: Link UP Jan 24 11:51:37.007769 systemd-networkd[1544]: vxlan.calico: Gained carrier Jan 24 11:51:37.108327 systemd[1]: Started cri-containerd-7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793.scope - libcontainer container 7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793. Jan 24 11:51:37.164000 audit: BPF prog-id=193 op=LOAD Jan 24 11:51:37.164000 audit[4634]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffcba961650 a2=98 a3=0 items=0 ppid=4422 pid=4634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.164000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:51:37.164000 audit: BPF prog-id=193 op=UNLOAD Jan 24 11:51:37.164000 audit[4634]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffcba961620 a3=0 items=0 ppid=4422 pid=4634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.164000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:51:37.164000 audit: BPF prog-id=194 op=LOAD Jan 24 11:51:37.164000 audit[4634]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffcba961460 a2=94 a3=54428f items=0 ppid=4422 pid=4634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.164000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:51:37.164000 audit: BPF prog-id=194 op=UNLOAD Jan 24 11:51:37.164000 audit[4634]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffcba961460 a2=94 a3=54428f items=0 ppid=4422 pid=4634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.164000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:51:37.164000 audit: BPF prog-id=195 op=LOAD Jan 24 11:51:37.164000 audit[4634]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffcba961490 a2=94 a3=2 items=0 ppid=4422 pid=4634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.164000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:51:37.164000 audit: BPF prog-id=195 op=UNLOAD Jan 24 11:51:37.164000 audit[4634]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffcba961490 a2=0 a3=2 items=0 ppid=4422 pid=4634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.164000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:51:37.164000 audit: BPF prog-id=196 op=LOAD Jan 24 11:51:37.164000 audit[4634]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffcba961240 a2=94 a3=4 items=0 ppid=4422 pid=4634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.164000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:51:37.164000 audit: BPF prog-id=196 op=UNLOAD Jan 24 11:51:37.164000 audit[4634]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffcba961240 a2=94 a3=4 items=0 ppid=4422 pid=4634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.164000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:51:37.164000 audit: BPF prog-id=197 op=LOAD Jan 24 11:51:37.164000 audit[4634]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffcba961340 a2=94 a3=7ffcba9614c0 items=0 ppid=4422 pid=4634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.164000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:51:37.164000 audit: BPF prog-id=197 op=UNLOAD Jan 24 11:51:37.164000 audit[4634]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffcba961340 a2=0 a3=7ffcba9614c0 items=0 ppid=4422 pid=4634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.164000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:51:37.170000 audit: BPF prog-id=198 op=LOAD Jan 24 11:51:37.170000 audit[4634]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffcba960a70 a2=94 a3=2 items=0 ppid=4422 pid=4634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.170000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:51:37.170000 audit: BPF prog-id=198 op=UNLOAD Jan 24 11:51:37.170000 audit[4634]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffcba960a70 a2=0 a3=2 items=0 ppid=4422 pid=4634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.170000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:51:37.170000 audit: BPF prog-id=199 op=LOAD Jan 24 11:51:37.170000 audit[4634]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffcba960b70 a2=94 a3=30 items=0 ppid=4422 pid=4634 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.170000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 24 11:51:37.259000 audit: BPF prog-id=200 op=LOAD Jan 24 11:51:37.259000 audit[4643]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe82ea7c00 a2=98 a3=0 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.259000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.259000 audit: BPF prog-id=200 op=UNLOAD Jan 24 11:51:37.259000 audit[4643]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffe82ea7bd0 a3=0 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.259000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.259000 audit: BPF prog-id=201 op=LOAD Jan 24 11:51:37.259000 audit[4643]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe82ea79f0 a2=94 a3=54428f items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.259000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.259000 audit: BPF prog-id=201 op=UNLOAD Jan 24 11:51:37.259000 audit[4643]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe82ea79f0 a2=94 a3=54428f items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.259000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.259000 audit: BPF prog-id=202 op=LOAD Jan 24 11:51:37.259000 audit[4643]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe82ea7a20 a2=94 a3=2 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.259000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.259000 audit: BPF prog-id=202 op=UNLOAD Jan 24 11:51:37.259000 audit[4643]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe82ea7a20 a2=0 a3=2 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.259000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.273000 audit: BPF prog-id=203 op=LOAD Jan 24 11:51:37.276000 audit: BPF prog-id=204 op=LOAD Jan 24 11:51:37.276000 audit[4609]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8238 a2=98 a3=0 items=0 ppid=4598 pid=4609 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.276000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3762396361613465623239376230663338396465376665666536363466 Jan 24 11:51:37.298000 audit: BPF prog-id=204 op=UNLOAD Jan 24 11:51:37.298000 audit[4609]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4598 pid=4609 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.298000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3762396361613465623239376230663338396465376665666536363466 Jan 24 11:51:37.326000 audit: BPF prog-id=205 op=LOAD Jan 24 11:51:37.326000 audit[4609]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8488 a2=98 a3=0 items=0 ppid=4598 pid=4609 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.326000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3762396361613465623239376230663338396465376665666536363466 Jan 24 11:51:37.329000 audit: BPF prog-id=206 op=LOAD Jan 24 11:51:37.329000 audit[4609]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a8218 a2=98 a3=0 items=0 ppid=4598 pid=4609 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.329000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3762396361613465623239376230663338396465376665666536363466 Jan 24 11:51:37.329000 audit: BPF prog-id=206 op=UNLOAD Jan 24 11:51:37.329000 audit[4609]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4598 pid=4609 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.329000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3762396361613465623239376230663338396465376665666536363466 Jan 24 11:51:37.330000 audit: BPF prog-id=205 op=UNLOAD Jan 24 11:51:37.330000 audit[4609]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4598 pid=4609 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.330000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3762396361613465623239376230663338396465376665666536363466 Jan 24 11:51:37.333000 audit: BPF prog-id=207 op=LOAD Jan 24 11:51:37.333000 audit[4609]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a86e8 a2=98 a3=0 items=0 ppid=4598 pid=4609 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.333000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3762396361613465623239376230663338396465376665666536363466 Jan 24 11:51:37.349928 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:51:37.545854 containerd[1647]: time="2026-01-24T11:51:37.542407600Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5c99cdc8ff-zh5gl,Uid:5cdab4ea-dcce-4467-b450-26abe2130c8d,Namespace:calico-system,Attempt:0,} returns sandbox id \"7b9caa4eb297b0f389de7fefe664fca9f84d6ff235c5f18c130b7fed73dc5793\"" Jan 24 11:51:37.554111 containerd[1647]: time="2026-01-24T11:51:37.553485774Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 24 11:51:37.635960 containerd[1647]: time="2026-01-24T11:51:37.635858027Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:51:37.640304 containerd[1647]: time="2026-01-24T11:51:37.639466739Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 24 11:51:37.640304 containerd[1647]: time="2026-01-24T11:51:37.639687871Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 24 11:51:37.640508 kubelet[2895]: E0124 11:51:37.640076 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:51:37.640508 kubelet[2895]: E0124 11:51:37.640433 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:51:37.642963 kubelet[2895]: E0124 11:51:37.642806 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:d91f0bbcac534ce79a38827d5dd29c41,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5n6cl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c99cdc8ff-zh5gl_calico-system(5cdab4ea-dcce-4467-b450-26abe2130c8d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 24 11:51:37.645029 containerd[1647]: time="2026-01-24T11:51:37.644986616Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 24 11:51:37.736000 audit: BPF prog-id=208 op=LOAD Jan 24 11:51:37.736000 audit[4643]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe82ea78e0 a2=94 a3=1 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.736000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.736000 audit: BPF prog-id=208 op=UNLOAD Jan 24 11:51:37.736000 audit[4643]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe82ea78e0 a2=94 a3=1 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.736000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.741808 containerd[1647]: time="2026-01-24T11:51:37.741451654Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:51:37.751412 containerd[1647]: time="2026-01-24T11:51:37.751227609Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 24 11:51:37.751412 containerd[1647]: time="2026-01-24T11:51:37.751340170Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 24 11:51:37.754093 kubelet[2895]: E0124 11:51:37.753811 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:51:37.754093 kubelet[2895]: E0124 11:51:37.754058 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:51:37.755970 kubelet[2895]: E0124 11:51:37.755837 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5n6cl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c99cdc8ff-zh5gl_calico-system(5cdab4ea-dcce-4467-b450-26abe2130c8d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 24 11:51:37.754000 audit: BPF prog-id=209 op=LOAD Jan 24 11:51:37.754000 audit[4643]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe82ea78d0 a2=94 a3=4 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.754000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.754000 audit: BPF prog-id=209 op=UNLOAD Jan 24 11:51:37.754000 audit[4643]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffe82ea78d0 a2=0 a3=4 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.754000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.754000 audit: BPF prog-id=210 op=LOAD Jan 24 11:51:37.754000 audit[4643]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffe82ea7730 a2=94 a3=5 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.754000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.754000 audit: BPF prog-id=210 op=UNLOAD Jan 24 11:51:37.754000 audit[4643]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffe82ea7730 a2=0 a3=5 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.754000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.755000 audit: BPF prog-id=211 op=LOAD Jan 24 11:51:37.755000 audit[4643]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe82ea7950 a2=94 a3=6 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.755000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.755000 audit: BPF prog-id=211 op=UNLOAD Jan 24 11:51:37.755000 audit[4643]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffe82ea7950 a2=0 a3=6 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.755000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.755000 audit: BPF prog-id=212 op=LOAD Jan 24 11:51:37.755000 audit[4643]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe82ea7100 a2=94 a3=88 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.755000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.756000 audit: BPF prog-id=213 op=LOAD Jan 24 11:51:37.756000 audit[4643]: SYSCALL arch=c000003e syscall=321 success=yes exit=7 a0=5 a1=7ffe82ea6f80 a2=94 a3=2 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.756000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.756000 audit: BPF prog-id=213 op=UNLOAD Jan 24 11:51:37.756000 audit[4643]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=7 a1=7ffe82ea6fb0 a2=0 a3=7ffe82ea70b0 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.756000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.756000 audit: BPF prog-id=212 op=UNLOAD Jan 24 11:51:37.756000 audit[4643]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=2a629d10 a2=0 a3=14eab16330b70cb4 items=0 ppid=4422 pid=4643 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.756000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 24 11:51:37.758953 kubelet[2895]: E0124 11:51:37.758714 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:51:37.770000 audit: BPF prog-id=199 op=UNLOAD Jan 24 11:51:37.770000 audit[4422]: SYSCALL arch=c000003e syscall=263 success=yes exit=0 a0=ffffffffffffff9c a1=c000c2c740 a2=0 a3=0 items=0 ppid=4409 pid=4422 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="calico-node" exe="/usr/bin/calico-node" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.770000 audit: PROCTITLE proctitle=63616C69636F2D6E6F6465002D66656C6978 Jan 24 11:51:37.883511 systemd-networkd[1544]: calicf40c7cce41: Gained IPv6LL Jan 24 11:51:37.976000 audit[4681]: NETFILTER_CFG table=nat:123 family=2 entries=15 op=nft_register_chain pid=4681 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:51:37.976000 audit[4681]: SYSCALL arch=c000003e syscall=46 success=yes exit=5084 a0=3 a1=7ffca3a7d950 a2=0 a3=7ffca3a7d93c items=0 ppid=4422 pid=4681 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.976000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:51:37.980000 audit[4680]: NETFILTER_CFG table=mangle:124 family=2 entries=16 op=nft_register_chain pid=4680 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:51:37.980000 audit[4680]: SYSCALL arch=c000003e syscall=46 success=yes exit=6868 a0=3 a1=7fffac68b6f0 a2=0 a3=7fffac68b6dc items=0 ppid=4422 pid=4680 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:37.980000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:51:38.046000 audit[4679]: NETFILTER_CFG table=raw:125 family=2 entries=21 op=nft_register_chain pid=4679 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:51:38.046000 audit[4679]: SYSCALL arch=c000003e syscall=46 success=yes exit=8452 a0=3 a1=7ffea0d64980 a2=0 a3=7ffea0d6496c items=0 ppid=4422 pid=4679 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:38.046000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:51:38.049000 audit[4682]: NETFILTER_CFG table=filter:126 family=2 entries=94 op=nft_register_chain pid=4682 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:51:38.049000 audit[4682]: SYSCALL arch=c000003e syscall=46 success=yes exit=53116 a0=3 a1=7ffc1b07eaa0 a2=0 a3=7ffc1b07ea8c items=0 ppid=4422 pid=4682 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:38.049000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:51:38.587768 systemd-networkd[1544]: vxlan.calico: Gained IPv6LL Jan 24 11:51:38.761657 kubelet[2895]: E0124 11:51:38.761434 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:51:38.806000 audit[4694]: NETFILTER_CFG table=filter:127 family=2 entries=20 op=nft_register_rule pid=4694 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:38.806000 audit[4694]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7fffff23dec0 a2=0 a3=7fffff23deac items=0 ppid=3040 pid=4694 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:38.806000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:38.813000 audit[4694]: NETFILTER_CFG table=nat:128 family=2 entries=14 op=nft_register_rule pid=4694 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:38.813000 audit[4694]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7fffff23dec0 a2=0 a3=0 items=0 ppid=3040 pid=4694 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:38.813000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:39.141475 containerd[1647]: time="2026-01-24T11:51:39.141069456Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-66955b7fc7-xmscg,Uid:41fe2d74-dafe-45a2-82d2-5fd56447e59d,Namespace:calico-system,Attempt:0,}" Jan 24 11:51:39.511027 systemd-networkd[1544]: cali9f21c845a2e: Link UP Jan 24 11:51:39.521416 systemd-networkd[1544]: cali9f21c845a2e: Gained carrier Jan 24 11:51:39.568176 containerd[1647]: 2026-01-24 11:51:39.265 [INFO][4701] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--66955b7fc7--xmscg-eth0 calico-kube-controllers-66955b7fc7- calico-system 41fe2d74-dafe-45a2-82d2-5fd56447e59d 956 0 2026-01-24 11:50:53 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:66955b7fc7 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-66955b7fc7-xmscg eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali9f21c845a2e [] [] }} ContainerID="91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" Namespace="calico-system" Pod="calico-kube-controllers-66955b7fc7-xmscg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66955b7fc7--xmscg-" Jan 24 11:51:39.568176 containerd[1647]: 2026-01-24 11:51:39.265 [INFO][4701] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" Namespace="calico-system" Pod="calico-kube-controllers-66955b7fc7-xmscg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66955b7fc7--xmscg-eth0" Jan 24 11:51:39.568176 containerd[1647]: 2026-01-24 11:51:39.340 [INFO][4709] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" HandleID="k8s-pod-network.91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" Workload="localhost-k8s-calico--kube--controllers--66955b7fc7--xmscg-eth0" Jan 24 11:51:39.568570 containerd[1647]: 2026-01-24 11:51:39.340 [INFO][4709] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" HandleID="k8s-pod-network.91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" Workload="localhost-k8s-calico--kube--controllers--66955b7fc7--xmscg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001237b0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-66955b7fc7-xmscg", "timestamp":"2026-01-24 11:51:39.340072161 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:51:39.568570 containerd[1647]: 2026-01-24 11:51:39.344 [INFO][4709] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:51:39.568570 containerd[1647]: 2026-01-24 11:51:39.345 [INFO][4709] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:51:39.568570 containerd[1647]: 2026-01-24 11:51:39.346 [INFO][4709] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:51:39.568570 containerd[1647]: 2026-01-24 11:51:39.374 [INFO][4709] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" host="localhost" Jan 24 11:51:39.568570 containerd[1647]: 2026-01-24 11:51:39.399 [INFO][4709] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:51:39.568570 containerd[1647]: 2026-01-24 11:51:39.419 [INFO][4709] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:51:39.568570 containerd[1647]: 2026-01-24 11:51:39.426 [INFO][4709] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:39.568570 containerd[1647]: 2026-01-24 11:51:39.433 [INFO][4709] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:39.568570 containerd[1647]: 2026-01-24 11:51:39.433 [INFO][4709] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" host="localhost" Jan 24 11:51:39.569841 containerd[1647]: 2026-01-24 11:51:39.441 [INFO][4709] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287 Jan 24 11:51:39.569841 containerd[1647]: 2026-01-24 11:51:39.467 [INFO][4709] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" host="localhost" Jan 24 11:51:39.569841 containerd[1647]: 2026-01-24 11:51:39.489 [INFO][4709] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" host="localhost" Jan 24 11:51:39.569841 containerd[1647]: 2026-01-24 11:51:39.490 [INFO][4709] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" host="localhost" Jan 24 11:51:39.569841 containerd[1647]: 2026-01-24 11:51:39.490 [INFO][4709] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:51:39.569841 containerd[1647]: 2026-01-24 11:51:39.490 [INFO][4709] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" HandleID="k8s-pod-network.91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" Workload="localhost-k8s-calico--kube--controllers--66955b7fc7--xmscg-eth0" Jan 24 11:51:39.569961 containerd[1647]: 2026-01-24 11:51:39.501 [INFO][4701] cni-plugin/k8s.go 418: Populated endpoint ContainerID="91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" Namespace="calico-system" Pod="calico-kube-controllers-66955b7fc7-xmscg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66955b7fc7--xmscg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--66955b7fc7--xmscg-eth0", GenerateName:"calico-kube-controllers-66955b7fc7-", Namespace:"calico-system", SelfLink:"", UID:"41fe2d74-dafe-45a2-82d2-5fd56447e59d", ResourceVersion:"956", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 50, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"66955b7fc7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-66955b7fc7-xmscg", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali9f21c845a2e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:39.570047 containerd[1647]: 2026-01-24 11:51:39.501 [INFO][4701] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" Namespace="calico-system" Pod="calico-kube-controllers-66955b7fc7-xmscg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66955b7fc7--xmscg-eth0" Jan 24 11:51:39.570047 containerd[1647]: 2026-01-24 11:51:39.501 [INFO][4701] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9f21c845a2e ContainerID="91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" Namespace="calico-system" Pod="calico-kube-controllers-66955b7fc7-xmscg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66955b7fc7--xmscg-eth0" Jan 24 11:51:39.570047 containerd[1647]: 2026-01-24 11:51:39.518 [INFO][4701] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" Namespace="calico-system" Pod="calico-kube-controllers-66955b7fc7-xmscg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66955b7fc7--xmscg-eth0" Jan 24 11:51:39.570109 containerd[1647]: 2026-01-24 11:51:39.522 [INFO][4701] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" Namespace="calico-system" Pod="calico-kube-controllers-66955b7fc7-xmscg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66955b7fc7--xmscg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--66955b7fc7--xmscg-eth0", GenerateName:"calico-kube-controllers-66955b7fc7-", Namespace:"calico-system", SelfLink:"", UID:"41fe2d74-dafe-45a2-82d2-5fd56447e59d", ResourceVersion:"956", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 50, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"66955b7fc7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287", Pod:"calico-kube-controllers-66955b7fc7-xmscg", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali9f21c845a2e", MAC:"a6:ab:6d:77:aa:76", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:39.570184 containerd[1647]: 2026-01-24 11:51:39.563 [INFO][4701] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" Namespace="calico-system" Pod="calico-kube-controllers-66955b7fc7-xmscg" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66955b7fc7--xmscg-eth0" Jan 24 11:51:39.596000 audit[4726]: NETFILTER_CFG table=filter:129 family=2 entries=36 op=nft_register_chain pid=4726 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:51:39.596000 audit[4726]: SYSCALL arch=c000003e syscall=46 success=yes exit=19576 a0=3 a1=7ffe6b5b3fd0 a2=0 a3=7ffe6b5b3fbc items=0 ppid=4422 pid=4726 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:39.596000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:51:39.668941 containerd[1647]: time="2026-01-24T11:51:39.668829166Z" level=info msg="connecting to shim 91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287" address="unix:///run/containerd/s/965b4173309bbb121507e45d1e2baf820bae4c65dae6ee13746f0a0c6e65b621" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:51:39.748712 systemd[1]: Started cri-containerd-91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287.scope - libcontainer container 91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287. Jan 24 11:51:39.774000 audit: BPF prog-id=214 op=LOAD Jan 24 11:51:39.776000 audit: BPF prog-id=215 op=LOAD Jan 24 11:51:39.776000 audit[4747]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=4736 pid=4747 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:39.776000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3931383133623165303138383734333639613737373438303962303332 Jan 24 11:51:39.776000 audit: BPF prog-id=215 op=UNLOAD Jan 24 11:51:39.776000 audit[4747]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4736 pid=4747 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:39.776000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3931383133623165303138383734333639613737373438303962303332 Jan 24 11:51:39.776000 audit: BPF prog-id=216 op=LOAD Jan 24 11:51:39.776000 audit[4747]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=4736 pid=4747 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:39.776000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3931383133623165303138383734333639613737373438303962303332 Jan 24 11:51:39.776000 audit: BPF prog-id=217 op=LOAD Jan 24 11:51:39.776000 audit[4747]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=4736 pid=4747 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:39.776000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3931383133623165303138383734333639613737373438303962303332 Jan 24 11:51:39.776000 audit: BPF prog-id=217 op=UNLOAD Jan 24 11:51:39.776000 audit[4747]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4736 pid=4747 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:39.776000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3931383133623165303138383734333639613737373438303962303332 Jan 24 11:51:39.776000 audit: BPF prog-id=216 op=UNLOAD Jan 24 11:51:39.776000 audit[4747]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4736 pid=4747 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:39.776000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3931383133623165303138383734333639613737373438303962303332 Jan 24 11:51:39.777000 audit: BPF prog-id=218 op=LOAD Jan 24 11:51:39.777000 audit[4747]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=4736 pid=4747 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:39.777000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3931383133623165303138383734333639613737373438303962303332 Jan 24 11:51:39.790334 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:51:39.908084 containerd[1647]: time="2026-01-24T11:51:39.907990876Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-66955b7fc7-xmscg,Uid:41fe2d74-dafe-45a2-82d2-5fd56447e59d,Namespace:calico-system,Attempt:0,} returns sandbox id \"91813b1e018874369a7774809b0329717999c3bc33207956d80a2834d026f287\"" Jan 24 11:51:39.911150 containerd[1647]: time="2026-01-24T11:51:39.911064876Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 24 11:51:39.976416 containerd[1647]: time="2026-01-24T11:51:39.976353775Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:51:39.982324 containerd[1647]: time="2026-01-24T11:51:39.982089042Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 24 11:51:39.983468 containerd[1647]: time="2026-01-24T11:51:39.982142053Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 24 11:51:39.983648 kubelet[2895]: E0124 11:51:39.983184 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:51:39.983648 kubelet[2895]: E0124 11:51:39.983252 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:51:39.984538 kubelet[2895]: E0124 11:51:39.984054 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p5kgf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-66955b7fc7-xmscg_calico-system(41fe2d74-dafe-45a2-82d2-5fd56447e59d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 24 11:51:39.985351 kubelet[2895]: E0124 11:51:39.985183 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:51:40.130325 kubelet[2895]: E0124 11:51:40.129923 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:40.130484 containerd[1647]: time="2026-01-24T11:51:40.130164368Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-2g45n,Uid:999c48c5-f3b1-4bfd-91f4-26cafee1eef0,Namespace:calico-system,Attempt:0,}" Jan 24 11:51:40.130801 containerd[1647]: time="2026-01-24T11:51:40.130737076Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-k9kt6,Uid:f46cf494-de53-4117-858a-b0e56c0ba7ac,Namespace:kube-system,Attempt:0,}" Jan 24 11:51:40.487458 systemd-networkd[1544]: cali1066188789e: Link UP Jan 24 11:51:40.491689 systemd-networkd[1544]: cali1066188789e: Gained carrier Jan 24 11:51:40.561490 containerd[1647]: 2026-01-24 11:51:40.251 [INFO][4774] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--2g45n-eth0 csi-node-driver- calico-system 999c48c5-f3b1-4bfd-91f4-26cafee1eef0 829 0 2026-01-24 11:50:53 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:857b56db8f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-2g45n eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali1066188789e [] [] }} ContainerID="1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" Namespace="calico-system" Pod="csi-node-driver-2g45n" WorkloadEndpoint="localhost-k8s-csi--node--driver--2g45n-" Jan 24 11:51:40.561490 containerd[1647]: 2026-01-24 11:51:40.252 [INFO][4774] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" Namespace="calico-system" Pod="csi-node-driver-2g45n" WorkloadEndpoint="localhost-k8s-csi--node--driver--2g45n-eth0" Jan 24 11:51:40.561490 containerd[1647]: 2026-01-24 11:51:40.335 [INFO][4805] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" HandleID="k8s-pod-network.1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" Workload="localhost-k8s-csi--node--driver--2g45n-eth0" Jan 24 11:51:40.562898 containerd[1647]: 2026-01-24 11:51:40.335 [INFO][4805] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" HandleID="k8s-pod-network.1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" Workload="localhost-k8s-csi--node--driver--2g45n-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00034c5a0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-2g45n", "timestamp":"2026-01-24 11:51:40.335113136 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:51:40.562898 containerd[1647]: 2026-01-24 11:51:40.335 [INFO][4805] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:51:40.562898 containerd[1647]: 2026-01-24 11:51:40.335 [INFO][4805] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:51:40.562898 containerd[1647]: 2026-01-24 11:51:40.335 [INFO][4805] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:51:40.562898 containerd[1647]: 2026-01-24 11:51:40.362 [INFO][4805] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" host="localhost" Jan 24 11:51:40.562898 containerd[1647]: 2026-01-24 11:51:40.385 [INFO][4805] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:51:40.562898 containerd[1647]: 2026-01-24 11:51:40.422 [INFO][4805] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:51:40.562898 containerd[1647]: 2026-01-24 11:51:40.435 [INFO][4805] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:40.562898 containerd[1647]: 2026-01-24 11:51:40.441 [INFO][4805] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:40.562898 containerd[1647]: 2026-01-24 11:51:40.442 [INFO][4805] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" host="localhost" Jan 24 11:51:40.563399 containerd[1647]: 2026-01-24 11:51:40.447 [INFO][4805] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4 Jan 24 11:51:40.563399 containerd[1647]: 2026-01-24 11:51:40.462 [INFO][4805] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" host="localhost" Jan 24 11:51:40.563399 containerd[1647]: 2026-01-24 11:51:40.474 [INFO][4805] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" host="localhost" Jan 24 11:51:40.563399 containerd[1647]: 2026-01-24 11:51:40.474 [INFO][4805] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" host="localhost" Jan 24 11:51:40.563399 containerd[1647]: 2026-01-24 11:51:40.474 [INFO][4805] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:51:40.563399 containerd[1647]: 2026-01-24 11:51:40.475 [INFO][4805] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" HandleID="k8s-pod-network.1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" Workload="localhost-k8s-csi--node--driver--2g45n-eth0" Jan 24 11:51:40.563708 containerd[1647]: 2026-01-24 11:51:40.480 [INFO][4774] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" Namespace="calico-system" Pod="csi-node-driver-2g45n" WorkloadEndpoint="localhost-k8s-csi--node--driver--2g45n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--2g45n-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"999c48c5-f3b1-4bfd-91f4-26cafee1eef0", ResourceVersion:"829", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 50, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-2g45n", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali1066188789e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:40.563838 containerd[1647]: 2026-01-24 11:51:40.481 [INFO][4774] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" Namespace="calico-system" Pod="csi-node-driver-2g45n" WorkloadEndpoint="localhost-k8s-csi--node--driver--2g45n-eth0" Jan 24 11:51:40.563838 containerd[1647]: 2026-01-24 11:51:40.481 [INFO][4774] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali1066188789e ContainerID="1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" Namespace="calico-system" Pod="csi-node-driver-2g45n" WorkloadEndpoint="localhost-k8s-csi--node--driver--2g45n-eth0" Jan 24 11:51:40.563838 containerd[1647]: 2026-01-24 11:51:40.494 [INFO][4774] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" Namespace="calico-system" Pod="csi-node-driver-2g45n" WorkloadEndpoint="localhost-k8s-csi--node--driver--2g45n-eth0" Jan 24 11:51:40.563977 containerd[1647]: 2026-01-24 11:51:40.499 [INFO][4774] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" Namespace="calico-system" Pod="csi-node-driver-2g45n" WorkloadEndpoint="localhost-k8s-csi--node--driver--2g45n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--2g45n-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"999c48c5-f3b1-4bfd-91f4-26cafee1eef0", ResourceVersion:"829", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 50, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4", Pod:"csi-node-driver-2g45n", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali1066188789e", MAC:"92:9a:0a:2c:21:1a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:40.564101 containerd[1647]: 2026-01-24 11:51:40.544 [INFO][4774] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" Namespace="calico-system" Pod="csi-node-driver-2g45n" WorkloadEndpoint="localhost-k8s-csi--node--driver--2g45n-eth0" Jan 24 11:51:40.635000 audit[4830]: NETFILTER_CFG table=filter:130 family=2 entries=40 op=nft_register_chain pid=4830 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:51:40.635000 audit[4830]: SYSCALL arch=c000003e syscall=46 success=yes exit=20764 a0=3 a1=7ffcbf840420 a2=0 a3=7ffcbf84040c items=0 ppid=4422 pid=4830 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:40.635000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:51:40.674747 systemd-networkd[1544]: cali33ec91e494e: Link UP Jan 24 11:51:40.676632 systemd-networkd[1544]: cali33ec91e494e: Gained carrier Jan 24 11:51:40.733512 containerd[1647]: 2026-01-24 11:51:40.255 [INFO][4784] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--674b8bbfcf--k9kt6-eth0 coredns-674b8bbfcf- kube-system f46cf494-de53-4117-858a-b0e56c0ba7ac 961 0 2026-01-24 11:50:23 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-674b8bbfcf-k9kt6 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali33ec91e494e [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" Namespace="kube-system" Pod="coredns-674b8bbfcf-k9kt6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--k9kt6-" Jan 24 11:51:40.733512 containerd[1647]: 2026-01-24 11:51:40.256 [INFO][4784] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" Namespace="kube-system" Pod="coredns-674b8bbfcf-k9kt6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--k9kt6-eth0" Jan 24 11:51:40.733512 containerd[1647]: 2026-01-24 11:51:40.387 [INFO][4811] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" HandleID="k8s-pod-network.7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" Workload="localhost-k8s-coredns--674b8bbfcf--k9kt6-eth0" Jan 24 11:51:40.733874 containerd[1647]: 2026-01-24 11:51:40.387 [INFO][4811] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" HandleID="k8s-pod-network.7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" Workload="localhost-k8s-coredns--674b8bbfcf--k9kt6-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00032e2a0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-674b8bbfcf-k9kt6", "timestamp":"2026-01-24 11:51:40.387430635 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:51:40.733874 containerd[1647]: 2026-01-24 11:51:40.387 [INFO][4811] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:51:40.733874 containerd[1647]: 2026-01-24 11:51:40.474 [INFO][4811] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:51:40.733874 containerd[1647]: 2026-01-24 11:51:40.475 [INFO][4811] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:51:40.733874 containerd[1647]: 2026-01-24 11:51:40.508 [INFO][4811] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" host="localhost" Jan 24 11:51:40.733874 containerd[1647]: 2026-01-24 11:51:40.535 [INFO][4811] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:51:40.733874 containerd[1647]: 2026-01-24 11:51:40.580 [INFO][4811] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:51:40.733874 containerd[1647]: 2026-01-24 11:51:40.593 [INFO][4811] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:40.733874 containerd[1647]: 2026-01-24 11:51:40.605 [INFO][4811] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:40.733874 containerd[1647]: 2026-01-24 11:51:40.605 [INFO][4811] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" host="localhost" Jan 24 11:51:40.734399 containerd[1647]: 2026-01-24 11:51:40.609 [INFO][4811] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43 Jan 24 11:51:40.734399 containerd[1647]: 2026-01-24 11:51:40.629 [INFO][4811] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" host="localhost" Jan 24 11:51:40.734399 containerd[1647]: 2026-01-24 11:51:40.654 [INFO][4811] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" host="localhost" Jan 24 11:51:40.734399 containerd[1647]: 2026-01-24 11:51:40.654 [INFO][4811] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" host="localhost" Jan 24 11:51:40.734399 containerd[1647]: 2026-01-24 11:51:40.654 [INFO][4811] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:51:40.734399 containerd[1647]: 2026-01-24 11:51:40.654 [INFO][4811] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" HandleID="k8s-pod-network.7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" Workload="localhost-k8s-coredns--674b8bbfcf--k9kt6-eth0" Jan 24 11:51:40.734642 containerd[1647]: 2026-01-24 11:51:40.665 [INFO][4784] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" Namespace="kube-system" Pod="coredns-674b8bbfcf-k9kt6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--k9kt6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--k9kt6-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"f46cf494-de53-4117-858a-b0e56c0ba7ac", ResourceVersion:"961", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 50, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-674b8bbfcf-k9kt6", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali33ec91e494e", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:40.734770 containerd[1647]: 2026-01-24 11:51:40.665 [INFO][4784] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" Namespace="kube-system" Pod="coredns-674b8bbfcf-k9kt6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--k9kt6-eth0" Jan 24 11:51:40.734770 containerd[1647]: 2026-01-24 11:51:40.665 [INFO][4784] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali33ec91e494e ContainerID="7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" Namespace="kube-system" Pod="coredns-674b8bbfcf-k9kt6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--k9kt6-eth0" Jan 24 11:51:40.734770 containerd[1647]: 2026-01-24 11:51:40.677 [INFO][4784] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" Namespace="kube-system" Pod="coredns-674b8bbfcf-k9kt6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--k9kt6-eth0" Jan 24 11:51:40.734898 containerd[1647]: 2026-01-24 11:51:40.681 [INFO][4784] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" Namespace="kube-system" Pod="coredns-674b8bbfcf-k9kt6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--k9kt6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--k9kt6-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"f46cf494-de53-4117-858a-b0e56c0ba7ac", ResourceVersion:"961", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 50, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43", Pod:"coredns-674b8bbfcf-k9kt6", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali33ec91e494e", MAC:"ae:0e:58:d3:bc:9e", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:40.734898 containerd[1647]: 2026-01-24 11:51:40.716 [INFO][4784] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" Namespace="kube-system" Pod="coredns-674b8bbfcf-k9kt6" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--k9kt6-eth0" Jan 24 11:51:40.756918 containerd[1647]: time="2026-01-24T11:51:40.756077656Z" level=info msg="connecting to shim 1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4" address="unix:///run/containerd/s/c10fb047b6d127fc3939fb2cc4d646665899f7819e22b10725ebc2e92062d501" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:51:40.788529 kubelet[2895]: E0124 11:51:40.788214 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:51:40.852000 audit[4864]: NETFILTER_CFG table=filter:131 family=2 entries=50 op=nft_register_chain pid=4864 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:51:40.852000 audit[4864]: SYSCALL arch=c000003e syscall=46 success=yes exit=24928 a0=3 a1=7ffc5d7129a0 a2=0 a3=7ffc5d71298c items=0 ppid=4422 pid=4864 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:40.852000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:51:41.027073 containerd[1647]: time="2026-01-24T11:51:41.025850804Z" level=info msg="connecting to shim 7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43" address="unix:///run/containerd/s/5526e1b56d9e5954e265a3b822732f6640217f1147903f5c34f48b074101fe6d" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:51:41.181497 containerd[1647]: time="2026-01-24T11:51:41.180575476Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-jfngl,Uid:293f1d4a-795c-44fc-bda8-3c550f56df72,Namespace:calico-system,Attempt:0,}" Jan 24 11:51:41.196111 systemd[1]: Started cri-containerd-1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4.scope - libcontainer container 1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4. Jan 24 11:51:41.215739 systemd-networkd[1544]: cali9f21c845a2e: Gained IPv6LL Jan 24 11:51:41.733831 systemd-networkd[1544]: cali1066188789e: Gained IPv6LL Jan 24 11:51:41.744737 systemd[1]: Started cri-containerd-7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43.scope - libcontainer container 7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43. Jan 24 11:51:41.830036 kubelet[2895]: E0124 11:51:41.829936 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:51:41.879341 kernel: kauditd_printk_skb: 211 callbacks suppressed Jan 24 11:51:41.879526 kernel: audit: type=1334 audit(1769255501.871:671): prog-id=219 op=LOAD Jan 24 11:51:41.871000 audit: BPF prog-id=219 op=LOAD Jan 24 11:51:41.878000 audit: BPF prog-id=220 op=LOAD Jan 24 11:51:41.890176 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:51:41.897016 kernel: audit: type=1334 audit(1769255501.878:672): prog-id=220 op=LOAD Jan 24 11:51:41.897117 kernel: audit: type=1300 audit(1769255501.878:672): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=4879 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.878000 audit[4893]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=4879 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.878000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735373234393335333936333463373263393435666639303138373839 Jan 24 11:51:41.879000 audit: BPF prog-id=220 op=UNLOAD Jan 24 11:51:41.994390 kernel: audit: type=1327 audit(1769255501.878:672): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735373234393335333936333463373263393435666639303138373839 Jan 24 11:51:41.994523 kernel: audit: type=1334 audit(1769255501.879:673): prog-id=220 op=UNLOAD Jan 24 11:51:41.994558 kernel: audit: type=1300 audit(1769255501.879:673): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4879 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.879000 audit[4893]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4879 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.978908 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:51:42.025829 kernel: audit: type=1327 audit(1769255501.879:673): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735373234393335333936333463373263393435666639303138373839 Jan 24 11:51:42.119382 kernel: audit: type=1334 audit(1769255501.879:674): prog-id=221 op=LOAD Jan 24 11:51:42.275209 kernel: audit: type=1300 audit(1769255501.879:674): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=4879 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:42.331577 kernel: audit: type=1327 audit(1769255501.879:674): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735373234393335333936333463373263393435666639303138373839 Jan 24 11:51:41.879000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735373234393335333936333463373263393435666639303138373839 Jan 24 11:51:41.879000 audit: BPF prog-id=221 op=LOAD Jan 24 11:51:41.879000 audit[4893]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=4879 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.879000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735373234393335333936333463373263393435666639303138373839 Jan 24 11:51:41.879000 audit: BPF prog-id=222 op=LOAD Jan 24 11:51:41.879000 audit[4893]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=4879 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.879000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735373234393335333936333463373263393435666639303138373839 Jan 24 11:51:41.879000 audit: BPF prog-id=222 op=UNLOAD Jan 24 11:51:41.879000 audit[4893]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4879 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.879000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735373234393335333936333463373263393435666639303138373839 Jan 24 11:51:41.879000 audit: BPF prog-id=221 op=UNLOAD Jan 24 11:51:41.879000 audit[4893]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4879 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.879000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735373234393335333936333463373263393435666639303138373839 Jan 24 11:51:41.879000 audit: BPF prog-id=223 op=LOAD Jan 24 11:51:41.879000 audit[4893]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=4879 pid=4893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.879000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735373234393335333936333463373263393435666639303138373839 Jan 24 11:51:41.886000 audit: BPF prog-id=224 op=LOAD Jan 24 11:51:41.898000 audit: BPF prog-id=225 op=LOAD Jan 24 11:51:41.898000 audit[4862]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=4852 pid=4862 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.898000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3135313165376233343664653163316664313664366664313235346439 Jan 24 11:51:41.898000 audit: BPF prog-id=225 op=UNLOAD Jan 24 11:51:41.898000 audit[4862]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4852 pid=4862 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.898000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3135313165376233343664653163316664313664366664313235346439 Jan 24 11:51:41.898000 audit: BPF prog-id=226 op=LOAD Jan 24 11:51:41.898000 audit[4862]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=4852 pid=4862 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.898000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3135313165376233343664653163316664313664366664313235346439 Jan 24 11:51:41.901000 audit: BPF prog-id=227 op=LOAD Jan 24 11:51:41.901000 audit[4862]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=4852 pid=4862 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.901000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3135313165376233343664653163316664313664366664313235346439 Jan 24 11:51:41.916000 audit: BPF prog-id=227 op=UNLOAD Jan 24 11:51:41.916000 audit[4862]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4852 pid=4862 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.916000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3135313165376233343664653163316664313664366664313235346439 Jan 24 11:51:41.916000 audit: BPF prog-id=226 op=UNLOAD Jan 24 11:51:41.916000 audit[4862]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4852 pid=4862 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.916000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3135313165376233343664653163316664313664366664313235346439 Jan 24 11:51:41.916000 audit: BPF prog-id=228 op=LOAD Jan 24 11:51:41.916000 audit[4862]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=4852 pid=4862 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:41.916000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3135313165376233343664653163316664313664366664313235346439 Jan 24 11:51:42.324200 systemd-networkd[1544]: cali33ec91e494e: Gained IPv6LL Jan 24 11:51:42.653347 kubelet[2895]: E0124 11:51:42.652717 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:42.655968 containerd[1647]: time="2026-01-24T11:51:42.653100228Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f485b4455-qhcml,Uid:c28e9f44-9bc9-4328-bf7a-f4f9d016a87e,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:51:42.660382 containerd[1647]: time="2026-01-24T11:51:42.657961874Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-dnscb,Uid:7a087061-e7ac-47a8-a3c4-b753efe3e78f,Namespace:kube-system,Attempt:0,}" Jan 24 11:51:42.849366 containerd[1647]: time="2026-01-24T11:51:42.848358022Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-k9kt6,Uid:f46cf494-de53-4117-858a-b0e56c0ba7ac,Namespace:kube-system,Attempt:0,} returns sandbox id \"7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43\"" Jan 24 11:51:42.859937 kubelet[2895]: E0124 11:51:42.859834 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:42.998075 containerd[1647]: time="2026-01-24T11:51:42.997373583Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-2g45n,Uid:999c48c5-f3b1-4bfd-91f4-26cafee1eef0,Namespace:calico-system,Attempt:0,} returns sandbox id \"1511e7b346de1c1fd16d6fd1254d962c6d8399af2be101bdcead201de11d4bb4\"" Jan 24 11:51:43.003077 containerd[1647]: time="2026-01-24T11:51:43.000374861Z" level=info msg="CreateContainer within sandbox \"7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 24 11:51:43.027411 containerd[1647]: time="2026-01-24T11:51:43.024717305Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 24 11:51:43.123151 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount81145018.mount: Deactivated successfully. Jan 24 11:51:43.132187 kubelet[2895]: E0124 11:51:43.132088 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:43.139200 containerd[1647]: time="2026-01-24T11:51:43.138240795Z" level=info msg="Container 965f610b12ccefbee16680b7fa98f543e9b1be7e268cbe2af4071812ce285a7d: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:51:43.143076 containerd[1647]: time="2026-01-24T11:51:43.143008590Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:51:43.152849 containerd[1647]: time="2026-01-24T11:51:43.151739226Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 24 11:51:43.153334 kubelet[2895]: E0124 11:51:43.153173 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:51:43.153334 kubelet[2895]: E0124 11:51:43.153244 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:51:43.155721 kubelet[2895]: E0124 11:51:43.155647 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lqn89,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2g45n_calico-system(999c48c5-f3b1-4bfd-91f4-26cafee1eef0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 24 11:51:43.156909 containerd[1647]: time="2026-01-24T11:51:43.156739884Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 24 11:51:43.160098 containerd[1647]: time="2026-01-24T11:51:43.160064865Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 24 11:51:43.183581 containerd[1647]: time="2026-01-24T11:51:43.183255900Z" level=info msg="CreateContainer within sandbox \"7572493539634c72c945ff901878919a9171cc413ffc654447433720838dae43\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"965f610b12ccefbee16680b7fa98f543e9b1be7e268cbe2af4071812ce285a7d\"" Jan 24 11:51:43.191365 containerd[1647]: time="2026-01-24T11:51:43.191221896Z" level=info msg="StartContainer for \"965f610b12ccefbee16680b7fa98f543e9b1be7e268cbe2af4071812ce285a7d\"" Jan 24 11:51:43.223227 containerd[1647]: time="2026-01-24T11:51:43.223177690Z" level=info msg="connecting to shim 965f610b12ccefbee16680b7fa98f543e9b1be7e268cbe2af4071812ce285a7d" address="unix:///run/containerd/s/5526e1b56d9e5954e265a3b822732f6640217f1147903f5c34f48b074101fe6d" protocol=ttrpc version=3 Jan 24 11:51:43.276225 containerd[1647]: time="2026-01-24T11:51:43.275493846Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:51:43.295710 containerd[1647]: time="2026-01-24T11:51:43.295567165Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 24 11:51:43.298709 containerd[1647]: time="2026-01-24T11:51:43.296125466Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 24 11:51:43.298841 kubelet[2895]: E0124 11:51:43.298398 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:51:43.299383 kubelet[2895]: E0124 11:51:43.299242 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:51:43.302870 systemd[1]: Started cri-containerd-965f610b12ccefbee16680b7fa98f543e9b1be7e268cbe2af4071812ce285a7d.scope - libcontainer container 965f610b12ccefbee16680b7fa98f543e9b1be7e268cbe2af4071812ce285a7d. Jan 24 11:51:43.305107 kubelet[2895]: E0124 11:51:43.300037 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lqn89,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2g45n_calico-system(999c48c5-f3b1-4bfd-91f4-26cafee1eef0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 24 11:51:43.306172 kubelet[2895]: E0124 11:51:43.305525 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:51:43.349138 systemd-networkd[1544]: cali34480ef1d92: Link UP Jan 24 11:51:43.369389 systemd-networkd[1544]: cali34480ef1d92: Gained carrier Jan 24 11:51:43.421000 audit: BPF prog-id=229 op=LOAD Jan 24 11:51:43.427000 audit: BPF prog-id=230 op=LOAD Jan 24 11:51:43.427000 audit[5001]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000186238 a2=98 a3=0 items=0 ppid=4879 pid=5001 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:43.427000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3936356636313062313263636566626565313636383062376661393866 Jan 24 11:51:43.428000 audit: BPF prog-id=230 op=UNLOAD Jan 24 11:51:43.428000 audit[5001]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4879 pid=5001 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:43.428000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3936356636313062313263636566626565313636383062376661393866 Jan 24 11:51:43.429000 audit: BPF prog-id=231 op=LOAD Jan 24 11:51:43.429000 audit[5001]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000186488 a2=98 a3=0 items=0 ppid=4879 pid=5001 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:43.429000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3936356636313062313263636566626565313636383062376661393866 Jan 24 11:51:43.429000 audit: BPF prog-id=232 op=LOAD Jan 24 11:51:43.429000 audit[5001]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000186218 a2=98 a3=0 items=0 ppid=4879 pid=5001 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:43.429000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3936356636313062313263636566626565313636383062376661393866 Jan 24 11:51:43.429000 audit: BPF prog-id=232 op=UNLOAD Jan 24 11:51:43.429000 audit[5001]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4879 pid=5001 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:43.429000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3936356636313062313263636566626565313636383062376661393866 Jan 24 11:51:43.429000 audit: BPF prog-id=231 op=UNLOAD Jan 24 11:51:43.429000 audit[5001]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4879 pid=5001 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:43.429000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3936356636313062313263636566626565313636383062376661393866 Jan 24 11:51:43.429000 audit: BPF prog-id=233 op=LOAD Jan 24 11:51:43.429000 audit[5001]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001866e8 a2=98 a3=0 items=0 ppid=4879 pid=5001 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:43.429000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3936356636313062313263636566626565313636383062376661393866 Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:41.934 [INFO][4913] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--666569f655--jfngl-eth0 goldmane-666569f655- calico-system 293f1d4a-795c-44fc-bda8-3c550f56df72 966 0 2026-01-24 11:50:48 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:666569f655 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-666569f655-jfngl eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali34480ef1d92 [] [] }} ContainerID="76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" Namespace="calico-system" Pod="goldmane-666569f655-jfngl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--jfngl-" Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:41.936 [INFO][4913] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" Namespace="calico-system" Pod="goldmane-666569f655-jfngl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--jfngl-eth0" Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:42.982 [INFO][4938] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" HandleID="k8s-pod-network.76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" Workload="localhost-k8s-goldmane--666569f655--jfngl-eth0" Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:42.983 [INFO][4938] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" HandleID="k8s-pod-network.76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" Workload="localhost-k8s-goldmane--666569f655--jfngl-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00034c2a0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-666569f655-jfngl", "timestamp":"2026-01-24 11:51:42.98275802 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:42.983 [INFO][4938] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:42.983 [INFO][4938] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:42.983 [INFO][4938] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:43.037 [INFO][4938] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" host="localhost" Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:43.117 [INFO][4938] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:43.173 [INFO][4938] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:43.192 [INFO][4938] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:43.208 [INFO][4938] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:43.210 [INFO][4938] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" host="localhost" Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:43.222 [INFO][4938] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63 Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:43.244 [INFO][4938] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" host="localhost" Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:43.293 [INFO][4938] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" host="localhost" Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:43.295 [INFO][4938] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" host="localhost" Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:43.296 [INFO][4938] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:51:43.596355 containerd[1647]: 2026-01-24 11:51:43.296 [INFO][4938] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" HandleID="k8s-pod-network.76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" Workload="localhost-k8s-goldmane--666569f655--jfngl-eth0" Jan 24 11:51:43.602151 containerd[1647]: 2026-01-24 11:51:43.320 [INFO][4913] cni-plugin/k8s.go 418: Populated endpoint ContainerID="76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" Namespace="calico-system" Pod="goldmane-666569f655-jfngl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--jfngl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--666569f655--jfngl-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"293f1d4a-795c-44fc-bda8-3c550f56df72", ResourceVersion:"966", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 50, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-666569f655-jfngl", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali34480ef1d92", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:43.602151 containerd[1647]: 2026-01-24 11:51:43.321 [INFO][4913] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" Namespace="calico-system" Pod="goldmane-666569f655-jfngl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--jfngl-eth0" Jan 24 11:51:43.602151 containerd[1647]: 2026-01-24 11:51:43.321 [INFO][4913] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali34480ef1d92 ContainerID="76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" Namespace="calico-system" Pod="goldmane-666569f655-jfngl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--jfngl-eth0" Jan 24 11:51:43.602151 containerd[1647]: 2026-01-24 11:51:43.380 [INFO][4913] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" Namespace="calico-system" Pod="goldmane-666569f655-jfngl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--jfngl-eth0" Jan 24 11:51:43.602151 containerd[1647]: 2026-01-24 11:51:43.381 [INFO][4913] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" Namespace="calico-system" Pod="goldmane-666569f655-jfngl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--jfngl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--666569f655--jfngl-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"293f1d4a-795c-44fc-bda8-3c550f56df72", ResourceVersion:"966", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 50, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63", Pod:"goldmane-666569f655-jfngl", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali34480ef1d92", MAC:"26:47:06:3b:9e:56", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:43.602151 containerd[1647]: 2026-01-24 11:51:43.563 [INFO][4913] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" Namespace="calico-system" Pod="goldmane-666569f655-jfngl" WorkloadEndpoint="localhost-k8s-goldmane--666569f655--jfngl-eth0" Jan 24 11:51:43.679931 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3149811034.mount: Deactivated successfully. Jan 24 11:51:43.704182 containerd[1647]: time="2026-01-24T11:51:43.703794715Z" level=info msg="StartContainer for \"965f610b12ccefbee16680b7fa98f543e9b1be7e268cbe2af4071812ce285a7d\" returns successfully" Jan 24 11:51:43.770000 audit[5055]: NETFILTER_CFG table=filter:132 family=2 entries=62 op=nft_register_chain pid=5055 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:51:43.770000 audit[5055]: SYSCALL arch=c000003e syscall=46 success=yes exit=31596 a0=3 a1=7ffd2d468150 a2=0 a3=7ffd2d46813c items=0 ppid=4422 pid=5055 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:43.770000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:51:43.836680 containerd[1647]: time="2026-01-24T11:51:43.835258118Z" level=info msg="connecting to shim 76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63" address="unix:///run/containerd/s/da31eb1d87513884e2194773fd7b3855e65f06d34f674ab039fbe80513c2a824" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:51:43.851081 kubelet[2895]: E0124 11:51:43.850906 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:43.915119 kubelet[2895]: E0124 11:51:43.912207 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:51:43.937685 systemd-networkd[1544]: cali6c302ca2adb: Link UP Jan 24 11:51:43.943483 systemd-networkd[1544]: cali6c302ca2adb: Gained carrier Jan 24 11:51:43.957806 kubelet[2895]: I0124 11:51:43.953231 2895 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-k9kt6" podStartSLOduration=80.95321093 podStartE2EDuration="1m20.95321093s" podCreationTimestamp="2026-01-24 11:50:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 11:51:43.952529428 +0000 UTC m=+83.474322145" watchObservedRunningTime="2026-01-24 11:51:43.95321093 +0000 UTC m=+83.475003636" Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.252 [INFO][4971] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--f485b4455--qhcml-eth0 calico-apiserver-f485b4455- calico-apiserver c28e9f44-9bc9-4328-bf7a-f4f9d016a87e 964 0 2026-01-24 11:50:40 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:f485b4455 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-f485b4455-qhcml eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali6c302ca2adb [] [] }} ContainerID="5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-qhcml" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--qhcml-" Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.252 [INFO][4971] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-qhcml" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--qhcml-eth0" Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.414 [INFO][5013] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" HandleID="k8s-pod-network.5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" Workload="localhost-k8s-calico--apiserver--f485b4455--qhcml-eth0" Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.415 [INFO][5013] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" HandleID="k8s-pod-network.5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" Workload="localhost-k8s-calico--apiserver--f485b4455--qhcml-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0006ac7c0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-f485b4455-qhcml", "timestamp":"2026-01-24 11:51:43.414948276 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.415 [INFO][5013] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.415 [INFO][5013] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.415 [INFO][5013] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.579 [INFO][5013] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" host="localhost" Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.613 [INFO][5013] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.667 [INFO][5013] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.692 [INFO][5013] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.709 [INFO][5013] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.709 [INFO][5013] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" host="localhost" Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.714 [INFO][5013] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.758 [INFO][5013] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" host="localhost" Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.827 [INFO][5013] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" host="localhost" Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.830 [INFO][5013] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" host="localhost" Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.831 [INFO][5013] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:51:44.058020 containerd[1647]: 2026-01-24 11:51:43.831 [INFO][5013] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" HandleID="k8s-pod-network.5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" Workload="localhost-k8s-calico--apiserver--f485b4455--qhcml-eth0" Jan 24 11:51:44.062789 containerd[1647]: 2026-01-24 11:51:43.851 [INFO][4971] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-qhcml" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--qhcml-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--f485b4455--qhcml-eth0", GenerateName:"calico-apiserver-f485b4455-", Namespace:"calico-apiserver", SelfLink:"", UID:"c28e9f44-9bc9-4328-bf7a-f4f9d016a87e", ResourceVersion:"964", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 50, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"f485b4455", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-f485b4455-qhcml", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali6c302ca2adb", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:44.062789 containerd[1647]: 2026-01-24 11:51:43.852 [INFO][4971] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-qhcml" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--qhcml-eth0" Jan 24 11:51:44.062789 containerd[1647]: 2026-01-24 11:51:43.852 [INFO][4971] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6c302ca2adb ContainerID="5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-qhcml" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--qhcml-eth0" Jan 24 11:51:44.062789 containerd[1647]: 2026-01-24 11:51:43.939 [INFO][4971] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-qhcml" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--qhcml-eth0" Jan 24 11:51:44.062789 containerd[1647]: 2026-01-24 11:51:43.971 [INFO][4971] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-qhcml" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--qhcml-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--f485b4455--qhcml-eth0", GenerateName:"calico-apiserver-f485b4455-", Namespace:"calico-apiserver", SelfLink:"", UID:"c28e9f44-9bc9-4328-bf7a-f4f9d016a87e", ResourceVersion:"964", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 50, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"f485b4455", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e", Pod:"calico-apiserver-f485b4455-qhcml", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali6c302ca2adb", MAC:"ae:dc:28:6b:e6:c5", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:44.062789 containerd[1647]: 2026-01-24 11:51:44.023 [INFO][4971] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-qhcml" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--qhcml-eth0" Jan 24 11:51:44.063000 audit[5087]: NETFILTER_CFG table=filter:133 family=2 entries=20 op=nft_register_rule pid=5087 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:44.063000 audit[5087]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffe9fbe6670 a2=0 a3=7ffe9fbe665c items=0 ppid=3040 pid=5087 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.063000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:44.077000 audit[5087]: NETFILTER_CFG table=nat:134 family=2 entries=14 op=nft_register_rule pid=5087 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:44.077000 audit[5087]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffe9fbe6670 a2=0 a3=0 items=0 ppid=3040 pid=5087 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.077000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:44.089335 systemd[1]: Started cri-containerd-76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63.scope - libcontainer container 76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63. Jan 24 11:51:44.133864 containerd[1647]: time="2026-01-24T11:51:44.133819254Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f485b4455-nc2lh,Uid:d47a09e2-52ab-4e49-942e-05f2cedb67ae,Namespace:calico-apiserver,Attempt:0,}" Jan 24 11:51:44.195000 audit[5114]: NETFILTER_CFG table=filter:135 family=2 entries=62 op=nft_register_chain pid=5114 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:51:44.195000 audit[5114]: SYSCALL arch=c000003e syscall=46 success=yes exit=31756 a0=3 a1=7fff6fa685c0 a2=0 a3=7fff6fa685ac items=0 ppid=4422 pid=5114 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.195000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:51:44.334383 systemd-networkd[1544]: calid8a54428aae: Link UP Jan 24 11:51:44.354251 systemd-networkd[1544]: calid8a54428aae: Gained carrier Jan 24 11:51:44.370000 audit: BPF prog-id=234 op=LOAD Jan 24 11:51:44.374000 audit: BPF prog-id=235 op=LOAD Jan 24 11:51:44.374000 audit[5084]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=5065 pid=5084 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.374000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736343033623035663136326339373237623130613765626239313939 Jan 24 11:51:44.375000 audit: BPF prog-id=235 op=UNLOAD Jan 24 11:51:44.375000 audit[5084]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5065 pid=5084 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.375000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736343033623035663136326339373237623130613765626239313939 Jan 24 11:51:44.378000 audit: BPF prog-id=236 op=LOAD Jan 24 11:51:44.378000 audit[5084]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=5065 pid=5084 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.378000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736343033623035663136326339373237623130613765626239313939 Jan 24 11:51:44.379000 audit: BPF prog-id=237 op=LOAD Jan 24 11:51:44.379000 audit[5084]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=5065 pid=5084 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.379000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736343033623035663136326339373237623130613765626239313939 Jan 24 11:51:44.380000 audit: BPF prog-id=237 op=UNLOAD Jan 24 11:51:44.380000 audit[5084]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5065 pid=5084 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.380000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736343033623035663136326339373237623130613765626239313939 Jan 24 11:51:44.381000 audit: BPF prog-id=236 op=UNLOAD Jan 24 11:51:44.381000 audit[5084]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5065 pid=5084 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.381000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736343033623035663136326339373237623130613765626239313939 Jan 24 11:51:44.383000 audit: BPF prog-id=238 op=LOAD Jan 24 11:51:44.383000 audit[5084]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=5065 pid=5084 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.383000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3736343033623035663136326339373237623130613765626239313939 Jan 24 11:51:44.395111 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:51:44.396065 containerd[1647]: time="2026-01-24T11:51:44.395967189Z" level=info msg="connecting to shim 5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e" address="unix:///run/containerd/s/8789bd821a64854fcde1d5845bf14ca7c9a62bb137f5aee44d190a5bfa213fe1" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:43.297 [INFO][4962] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--674b8bbfcf--dnscb-eth0 coredns-674b8bbfcf- kube-system 7a087061-e7ac-47a8-a3c4-b753efe3e78f 965 0 2026-01-24 11:50:23 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-674b8bbfcf-dnscb eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calid8a54428aae [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" Namespace="kube-system" Pod="coredns-674b8bbfcf-dnscb" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--dnscb-" Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:43.297 [INFO][4962] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" Namespace="kube-system" Pod="coredns-674b8bbfcf-dnscb" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--dnscb-eth0" Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:43.712 [INFO][5029] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" HandleID="k8s-pod-network.c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" Workload="localhost-k8s-coredns--674b8bbfcf--dnscb-eth0" Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:43.712 [INFO][5029] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" HandleID="k8s-pod-network.c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" Workload="localhost-k8s-coredns--674b8bbfcf--dnscb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000191230), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-674b8bbfcf-dnscb", "timestamp":"2026-01-24 11:51:43.712578933 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:43.715 [INFO][5029] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:43.837 [INFO][5029] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:43.838 [INFO][5029] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:43.941 [INFO][5029] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" host="localhost" Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:43.999 [INFO][5029] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:44.052 [INFO][5029] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:44.060 [INFO][5029] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:44.080 [INFO][5029] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:44.080 [INFO][5029] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" host="localhost" Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:44.091 [INFO][5029] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448 Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:44.145 [INFO][5029] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" host="localhost" Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:44.170 [INFO][5029] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" host="localhost" Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:44.170 [INFO][5029] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" host="localhost" Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:44.170 [INFO][5029] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:51:44.460024 containerd[1647]: 2026-01-24 11:51:44.170 [INFO][5029] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" HandleID="k8s-pod-network.c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" Workload="localhost-k8s-coredns--674b8bbfcf--dnscb-eth0" Jan 24 11:51:44.464053 containerd[1647]: 2026-01-24 11:51:44.199 [INFO][4962] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" Namespace="kube-system" Pod="coredns-674b8bbfcf-dnscb" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--dnscb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--dnscb-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"7a087061-e7ac-47a8-a3c4-b753efe3e78f", ResourceVersion:"965", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 50, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-674b8bbfcf-dnscb", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calid8a54428aae", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:44.464053 containerd[1647]: 2026-01-24 11:51:44.202 [INFO][4962] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" Namespace="kube-system" Pod="coredns-674b8bbfcf-dnscb" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--dnscb-eth0" Jan 24 11:51:44.464053 containerd[1647]: 2026-01-24 11:51:44.202 [INFO][4962] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid8a54428aae ContainerID="c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" Namespace="kube-system" Pod="coredns-674b8bbfcf-dnscb" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--dnscb-eth0" Jan 24 11:51:44.464053 containerd[1647]: 2026-01-24 11:51:44.361 [INFO][4962] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" Namespace="kube-system" Pod="coredns-674b8bbfcf-dnscb" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--dnscb-eth0" Jan 24 11:51:44.464053 containerd[1647]: 2026-01-24 11:51:44.363 [INFO][4962] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" Namespace="kube-system" Pod="coredns-674b8bbfcf-dnscb" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--dnscb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--674b8bbfcf--dnscb-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"7a087061-e7ac-47a8-a3c4-b753efe3e78f", ResourceVersion:"965", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 50, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448", Pod:"coredns-674b8bbfcf-dnscb", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calid8a54428aae", MAC:"de:eb:02:90:39:4f", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:44.464053 containerd[1647]: 2026-01-24 11:51:44.419 [INFO][4962] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" Namespace="kube-system" Pod="coredns-674b8bbfcf-dnscb" WorkloadEndpoint="localhost-k8s-coredns--674b8bbfcf--dnscb-eth0" Jan 24 11:51:44.587245 systemd[1]: Started cri-containerd-5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e.scope - libcontainer container 5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e. Jan 24 11:51:44.619164 containerd[1647]: time="2026-01-24T11:51:44.619048524Z" level=info msg="connecting to shim c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448" address="unix:///run/containerd/s/0528620bce623c39d8c647e5bf3aac9aa536a3c1b2b653783327b0783baa9151" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:51:44.696000 audit: BPF prog-id=239 op=LOAD Jan 24 11:51:44.698000 audit: BPF prog-id=240 op=LOAD Jan 24 11:51:44.698000 audit[5152]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=5139 pid=5152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.698000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3535393763666433363264663637373162356363623262333030323739 Jan 24 11:51:44.730000 audit: BPF prog-id=240 op=UNLOAD Jan 24 11:51:44.730000 audit[5152]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=5139 pid=5152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.730000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3535393763666433363264663637373162356363623262333030323739 Jan 24 11:51:44.747000 audit: BPF prog-id=241 op=LOAD Jan 24 11:51:44.747000 audit[5152]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=5139 pid=5152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.747000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3535393763666433363264663637373162356363623262333030323739 Jan 24 11:51:44.753000 audit: BPF prog-id=242 op=LOAD Jan 24 11:51:44.753000 audit[5152]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=5139 pid=5152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.753000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3535393763666433363264663637373162356363623262333030323739 Jan 24 11:51:44.758000 audit: BPF prog-id=242 op=UNLOAD Jan 24 11:51:44.758000 audit[5152]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=5139 pid=5152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.758000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3535393763666433363264663637373162356363623262333030323739 Jan 24 11:51:44.763000 audit: BPF prog-id=241 op=UNLOAD Jan 24 11:51:44.763000 audit[5152]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=5139 pid=5152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.763000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3535393763666433363264663637373162356363623262333030323739 Jan 24 11:51:44.763000 audit: BPF prog-id=243 op=LOAD Jan 24 11:51:44.763000 audit[5152]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=5139 pid=5152 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.763000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3535393763666433363264663637373162356363623262333030323739 Jan 24 11:51:44.785114 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:51:44.806510 containerd[1647]: time="2026-01-24T11:51:44.806468671Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-jfngl,Uid:293f1d4a-795c-44fc-bda8-3c550f56df72,Namespace:calico-system,Attempt:0,} returns sandbox id \"76403b05f162c9727b10a7ebb9199212112cb879ff1ea7c37c21e61d23718d63\"" Jan 24 11:51:44.857340 containerd[1647]: time="2026-01-24T11:51:44.857205278Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 24 11:51:44.860000 audit[5215]: NETFILTER_CFG table=filter:136 family=2 entries=48 op=nft_register_chain pid=5215 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:51:44.860000 audit[5215]: SYSCALL arch=c000003e syscall=46 success=yes exit=22704 a0=3 a1=7fff90d3e540 a2=0 a3=7fff90d3e52c items=0 ppid=4422 pid=5215 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:44.860000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:51:44.885244 kubelet[2895]: E0124 11:51:44.884704 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:44.887973 kubelet[2895]: E0124 11:51:44.887927 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:51:44.948703 systemd[1]: Started cri-containerd-c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448.scope - libcontainer container c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448. Jan 24 11:51:44.961049 containerd[1647]: time="2026-01-24T11:51:44.960957677Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:51:44.982084 containerd[1647]: time="2026-01-24T11:51:44.981853680Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 24 11:51:44.982836 containerd[1647]: time="2026-01-24T11:51:44.982413675Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 24 11:51:44.985438 kubelet[2895]: E0124 11:51:44.985393 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:51:44.987841 kubelet[2895]: E0124 11:51:44.987034 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:51:44.989952 kubelet[2895]: E0124 11:51:44.989258 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k2g54,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-jfngl_calico-system(293f1d4a-795c-44fc-bda8-3c550f56df72): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 24 11:51:44.991548 kubelet[2895]: E0124 11:51:44.991513 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:51:45.060000 audit: BPF prog-id=244 op=LOAD Jan 24 11:51:45.061000 audit: BPF prog-id=245 op=LOAD Jan 24 11:51:45.061000 audit[5207]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=5180 pid=5207 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.061000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6334373831646366366138373032353230386233313165626364633266 Jan 24 11:51:45.061000 audit: BPF prog-id=245 op=UNLOAD Jan 24 11:51:45.061000 audit[5207]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5180 pid=5207 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.061000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6334373831646366366138373032353230386233313165626364633266 Jan 24 11:51:45.061000 audit: BPF prog-id=246 op=LOAD Jan 24 11:51:45.061000 audit[5207]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=5180 pid=5207 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.061000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6334373831646366366138373032353230386233313165626364633266 Jan 24 11:51:45.061000 audit: BPF prog-id=247 op=LOAD Jan 24 11:51:45.061000 audit[5207]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=5180 pid=5207 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.061000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6334373831646366366138373032353230386233313165626364633266 Jan 24 11:51:45.061000 audit: BPF prog-id=247 op=UNLOAD Jan 24 11:51:45.061000 audit[5207]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5180 pid=5207 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.061000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6334373831646366366138373032353230386233313165626364633266 Jan 24 11:51:45.061000 audit: BPF prog-id=246 op=UNLOAD Jan 24 11:51:45.061000 audit[5207]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5180 pid=5207 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.061000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6334373831646366366138373032353230386233313165626364633266 Jan 24 11:51:45.061000 audit: BPF prog-id=248 op=LOAD Jan 24 11:51:45.061000 audit[5207]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=5180 pid=5207 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.061000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6334373831646366366138373032353230386233313165626364633266 Jan 24 11:51:45.067418 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:51:45.074000 audit[5245]: NETFILTER_CFG table=filter:137 family=2 entries=17 op=nft_register_rule pid=5245 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:45.074000 audit[5245]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffd570d8950 a2=0 a3=7ffd570d893c items=0 ppid=3040 pid=5245 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.074000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:45.083000 audit[5245]: NETFILTER_CFG table=nat:138 family=2 entries=35 op=nft_register_chain pid=5245 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:45.083000 audit[5245]: SYSCALL arch=c000003e syscall=46 success=yes exit=14196 a0=3 a1=7ffd570d8950 a2=0 a3=7ffd570d893c items=0 ppid=3040 pid=5245 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.083000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:45.089808 containerd[1647]: time="2026-01-24T11:51:45.089579489Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f485b4455-qhcml,Uid:c28e9f44-9bc9-4328-bf7a-f4f9d016a87e,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"5597cfd362df6771b5ccb2b3002797f9d5014fca3dfa03e4c6eedfbcbbfd8e6e\"" Jan 24 11:51:45.102193 containerd[1647]: time="2026-01-24T11:51:45.098417040Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:51:45.130659 kubelet[2895]: E0124 11:51:45.130010 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:45.258746 containerd[1647]: time="2026-01-24T11:51:45.258164548Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-dnscb,Uid:7a087061-e7ac-47a8-a3c4-b753efe3e78f,Namespace:kube-system,Attempt:0,} returns sandbox id \"c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448\"" Jan 24 11:51:45.270085 kubelet[2895]: E0124 11:51:45.269108 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:45.280058 containerd[1647]: time="2026-01-24T11:51:45.279878770Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:51:45.285387 containerd[1647]: time="2026-01-24T11:51:45.285168077Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:51:45.285845 containerd[1647]: time="2026-01-24T11:51:45.285743210Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:51:45.286548 kubelet[2895]: E0124 11:51:45.286505 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:51:45.286822 kubelet[2895]: E0124 11:51:45.286793 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:51:45.287784 kubelet[2895]: E0124 11:51:45.287495 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l682s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-f485b4455-qhcml_calico-apiserver(c28e9f44-9bc9-4328-bf7a-f4f9d016a87e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:51:45.289246 containerd[1647]: time="2026-01-24T11:51:45.289022957Z" level=info msg="CreateContainer within sandbox \"c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 24 11:51:45.290212 kubelet[2895]: E0124 11:51:45.290148 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:51:45.310179 systemd-networkd[1544]: cali34480ef1d92: Gained IPv6LL Jan 24 11:51:45.392651 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3552065862.mount: Deactivated successfully. Jan 24 11:51:45.407656 containerd[1647]: time="2026-01-24T11:51:45.407471278Z" level=info msg="Container 70a7cab8ca93d72f8e85d9a645defe8669c856530c28f4d4b90282ab3b1495ee: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:51:45.422646 systemd-networkd[1544]: cali485b9cfd6a6: Link UP Jan 24 11:51:45.426547 systemd-networkd[1544]: cali485b9cfd6a6: Gained carrier Jan 24 11:51:45.480715 containerd[1647]: time="2026-01-24T11:51:45.480643662Z" level=info msg="CreateContainer within sandbox \"c4781dcf6a87025208b311ebcdc2fffdf38c09d5c431d2ceb6a6e700b4537448\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"70a7cab8ca93d72f8e85d9a645defe8669c856530c28f4d4b90282ab3b1495ee\"" Jan 24 11:51:45.492739 containerd[1647]: time="2026-01-24T11:51:45.487761932Z" level=info msg="StartContainer for \"70a7cab8ca93d72f8e85d9a645defe8669c856530c28f4d4b90282ab3b1495ee\"" Jan 24 11:51:45.502085 systemd-networkd[1544]: cali6c302ca2adb: Gained IPv6LL Jan 24 11:51:45.553043 containerd[1647]: time="2026-01-24T11:51:45.552925108Z" level=info msg="connecting to shim 70a7cab8ca93d72f8e85d9a645defe8669c856530c28f4d4b90282ab3b1495ee" address="unix:///run/containerd/s/0528620bce623c39d8c647e5bf3aac9aa536a3c1b2b653783327b0783baa9151" protocol=ttrpc version=3 Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:44.807 [INFO][5118] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--f485b4455--nc2lh-eth0 calico-apiserver-f485b4455- calico-apiserver d47a09e2-52ab-4e49-942e-05f2cedb67ae 960 0 2026-01-24 11:50:40 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:f485b4455 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-f485b4455-nc2lh eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali485b9cfd6a6 [] [] }} ContainerID="59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-nc2lh" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--nc2lh-" Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:44.832 [INFO][5118] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-nc2lh" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--nc2lh-eth0" Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:44.991 [INFO][5218] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" HandleID="k8s-pod-network.59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" Workload="localhost-k8s-calico--apiserver--f485b4455--nc2lh-eth0" Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:44.992 [INFO][5218] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" HandleID="k8s-pod-network.59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" Workload="localhost-k8s-calico--apiserver--f485b4455--nc2lh-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003253d0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-f485b4455-nc2lh", "timestamp":"2026-01-24 11:51:44.991721667 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:44.992 [INFO][5218] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:44.992 [INFO][5218] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:44.992 [INFO][5218] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:45.089 [INFO][5218] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" host="localhost" Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:45.130 [INFO][5218] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:45.158 [INFO][5218] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:45.194 [INFO][5218] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:45.280 [INFO][5218] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:45.282 [INFO][5218] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" host="localhost" Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:45.295 [INFO][5218] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:45.332 [INFO][5218] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" host="localhost" Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:45.388 [INFO][5218] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" host="localhost" Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:45.388 [INFO][5218] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" host="localhost" Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:45.388 [INFO][5218] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 24 11:51:45.563089 containerd[1647]: 2026-01-24 11:51:45.388 [INFO][5218] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" HandleID="k8s-pod-network.59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" Workload="localhost-k8s-calico--apiserver--f485b4455--nc2lh-eth0" Jan 24 11:51:45.566363 containerd[1647]: 2026-01-24 11:51:45.399 [INFO][5118] cni-plugin/k8s.go 418: Populated endpoint ContainerID="59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-nc2lh" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--nc2lh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--f485b4455--nc2lh-eth0", GenerateName:"calico-apiserver-f485b4455-", Namespace:"calico-apiserver", SelfLink:"", UID:"d47a09e2-52ab-4e49-942e-05f2cedb67ae", ResourceVersion:"960", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 50, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"f485b4455", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-f485b4455-nc2lh", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali485b9cfd6a6", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:45.566363 containerd[1647]: 2026-01-24 11:51:45.400 [INFO][5118] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-nc2lh" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--nc2lh-eth0" Jan 24 11:51:45.566363 containerd[1647]: 2026-01-24 11:51:45.400 [INFO][5118] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali485b9cfd6a6 ContainerID="59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-nc2lh" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--nc2lh-eth0" Jan 24 11:51:45.566363 containerd[1647]: 2026-01-24 11:51:45.432 [INFO][5118] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-nc2lh" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--nc2lh-eth0" Jan 24 11:51:45.566363 containerd[1647]: 2026-01-24 11:51:45.441 [INFO][5118] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-nc2lh" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--nc2lh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--f485b4455--nc2lh-eth0", GenerateName:"calico-apiserver-f485b4455-", Namespace:"calico-apiserver", SelfLink:"", UID:"d47a09e2-52ab-4e49-942e-05f2cedb67ae", ResourceVersion:"960", Generation:0, CreationTimestamp:time.Date(2026, time.January, 24, 11, 50, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"f485b4455", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea", Pod:"calico-apiserver-f485b4455-nc2lh", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali485b9cfd6a6", MAC:"a6:62:bf:cf:68:87", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 24 11:51:45.566363 containerd[1647]: 2026-01-24 11:51:45.553 [INFO][5118] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" Namespace="calico-apiserver" Pod="calico-apiserver-f485b4455-nc2lh" WorkloadEndpoint="localhost-k8s-calico--apiserver--f485b4455--nc2lh-eth0" Jan 24 11:51:45.616000 audit[5272]: NETFILTER_CFG table=filter:139 family=2 entries=57 op=nft_register_chain pid=5272 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 24 11:51:45.616000 audit[5272]: SYSCALL arch=c000003e syscall=46 success=yes exit=27812 a0=3 a1=7fffd153f6a0 a2=0 a3=7fffd153f68c items=0 ppid=4422 pid=5272 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.616000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 24 11:51:45.648380 systemd[1]: Started cri-containerd-70a7cab8ca93d72f8e85d9a645defe8669c856530c28f4d4b90282ab3b1495ee.scope - libcontainer container 70a7cab8ca93d72f8e85d9a645defe8669c856530c28f4d4b90282ab3b1495ee. Jan 24 11:51:45.692054 containerd[1647]: time="2026-01-24T11:51:45.691991624Z" level=info msg="connecting to shim 59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea" address="unix:///run/containerd/s/5535fd79f2de96465e830e8277a2bab0d54ff4e5e32db6160617bb24307350cb" namespace=k8s.io protocol=ttrpc version=3 Jan 24 11:51:45.774000 audit: BPF prog-id=249 op=LOAD Jan 24 11:51:45.777000 audit: BPF prog-id=250 op=LOAD Jan 24 11:51:45.777000 audit[5256]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000220238 a2=98 a3=0 items=0 ppid=5180 pid=5256 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.777000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3730613763616238636139336437326638653835643961363435646566 Jan 24 11:51:45.777000 audit: BPF prog-id=250 op=UNLOAD Jan 24 11:51:45.777000 audit[5256]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5180 pid=5256 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.777000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3730613763616238636139336437326638653835643961363435646566 Jan 24 11:51:45.777000 audit: BPF prog-id=251 op=LOAD Jan 24 11:51:45.777000 audit[5256]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000220488 a2=98 a3=0 items=0 ppid=5180 pid=5256 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.777000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3730613763616238636139336437326638653835643961363435646566 Jan 24 11:51:45.777000 audit: BPF prog-id=252 op=LOAD Jan 24 11:51:45.777000 audit[5256]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000220218 a2=98 a3=0 items=0 ppid=5180 pid=5256 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.777000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3730613763616238636139336437326638653835643961363435646566 Jan 24 11:51:45.778000 audit: BPF prog-id=252 op=UNLOAD Jan 24 11:51:45.778000 audit[5256]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5180 pid=5256 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.778000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3730613763616238636139336437326638653835643961363435646566 Jan 24 11:51:45.778000 audit: BPF prog-id=251 op=UNLOAD Jan 24 11:51:45.778000 audit[5256]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5180 pid=5256 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.778000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3730613763616238636139336437326638653835643961363435646566 Jan 24 11:51:45.778000 audit: BPF prog-id=253 op=LOAD Jan 24 11:51:45.778000 audit[5256]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0002206e8 a2=98 a3=0 items=0 ppid=5180 pid=5256 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.778000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3730613763616238636139336437326638653835643961363435646566 Jan 24 11:51:45.808540 systemd[1]: Started cri-containerd-59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea.scope - libcontainer container 59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea. Jan 24 11:51:45.880000 audit: BPF prog-id=254 op=LOAD Jan 24 11:51:45.882000 audit: BPF prog-id=255 op=LOAD Jan 24 11:51:45.882000 audit[5303]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=5283 pid=5303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.882000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3539663535646162636333636366666439376331346630393361643739 Jan 24 11:51:45.882000 audit: BPF prog-id=255 op=UNLOAD Jan 24 11:51:45.882000 audit[5303]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5283 pid=5303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.882000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3539663535646162636333636366666439376331346630393361643739 Jan 24 11:51:45.882000 audit: BPF prog-id=256 op=LOAD Jan 24 11:51:45.882000 audit[5303]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=5283 pid=5303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.882000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3539663535646162636333636366666439376331346630393361643739 Jan 24 11:51:45.882000 audit: BPF prog-id=257 op=LOAD Jan 24 11:51:45.882000 audit[5303]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=5283 pid=5303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.882000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3539663535646162636333636366666439376331346630393361643739 Jan 24 11:51:45.883000 audit: BPF prog-id=257 op=UNLOAD Jan 24 11:51:45.883000 audit[5303]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=5283 pid=5303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.883000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3539663535646162636333636366666439376331346630393361643739 Jan 24 11:51:45.883000 audit: BPF prog-id=256 op=UNLOAD Jan 24 11:51:45.883000 audit[5303]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=5283 pid=5303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.883000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3539663535646162636333636366666439376331346630393361643739 Jan 24 11:51:45.883000 audit: BPF prog-id=258 op=LOAD Jan 24 11:51:45.883000 audit[5303]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=5283 pid=5303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:45.883000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3539663535646162636333636366666439376331346630393361643739 Jan 24 11:51:45.889648 systemd-resolved[1287]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 24 11:51:45.899686 containerd[1647]: time="2026-01-24T11:51:45.899641400Z" level=info msg="StartContainer for \"70a7cab8ca93d72f8e85d9a645defe8669c856530c28f4d4b90282ab3b1495ee\" returns successfully" Jan 24 11:51:45.978240 kubelet[2895]: E0124 11:51:45.972539 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:45.980022 kubelet[2895]: E0124 11:51:45.979949 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:51:45.988969 kubelet[2895]: E0124 11:51:45.988845 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:51:46.146483 containerd[1647]: time="2026-01-24T11:51:46.146385438Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-f485b4455-nc2lh,Uid:d47a09e2-52ab-4e49-942e-05f2cedb67ae,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"59f55dabcc3ccffd97c14f093ad790575344343489cb6ffb86011458470c96ea\"" Jan 24 11:51:46.194695 containerd[1647]: time="2026-01-24T11:51:46.192167225Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:51:46.279000 audit[5348]: NETFILTER_CFG table=filter:140 family=2 entries=14 op=nft_register_rule pid=5348 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:46.279000 audit[5348]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffc1c32a440 a2=0 a3=7ffc1c32a42c items=0 ppid=3040 pid=5348 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:46.279000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:46.300000 audit[5348]: NETFILTER_CFG table=nat:141 family=2 entries=20 op=nft_register_rule pid=5348 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:46.300000 audit[5348]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffc1c32a440 a2=0 a3=7ffc1c32a42c items=0 ppid=3040 pid=5348 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:46.300000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:46.376056 kubelet[2895]: E0124 11:51:46.364432 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:51:46.376056 kubelet[2895]: E0124 11:51:46.365515 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:51:46.376056 kubelet[2895]: E0124 11:51:46.368095 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r4s6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-f485b4455-nc2lh_calico-apiserver(d47a09e2-52ab-4e49-942e-05f2cedb67ae): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:51:46.376056 kubelet[2895]: E0124 11:51:46.371463 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:51:46.376552 containerd[1647]: time="2026-01-24T11:51:46.325946684Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:51:46.376552 containerd[1647]: time="2026-01-24T11:51:46.358593182Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:51:46.376552 containerd[1647]: time="2026-01-24T11:51:46.359469377Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:51:46.386000 audit[5350]: NETFILTER_CFG table=filter:142 family=2 entries=14 op=nft_register_rule pid=5350 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:46.386000 audit[5350]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7fff7eeeff60 a2=0 a3=7fff7eeeff4c items=0 ppid=3040 pid=5350 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:46.386000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:46.396243 systemd-networkd[1544]: calid8a54428aae: Gained IPv6LL Jan 24 11:51:46.398000 audit[5350]: NETFILTER_CFG table=nat:143 family=2 entries=20 op=nft_register_rule pid=5350 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:46.398000 audit[5350]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7fff7eeeff60 a2=0 a3=7fff7eeeff4c items=0 ppid=3040 pid=5350 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:46.398000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:46.471439 systemd-networkd[1544]: cali485b9cfd6a6: Gained IPv6LL Jan 24 11:51:46.988820 kubelet[2895]: E0124 11:51:46.982371 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:51:46.988820 kubelet[2895]: E0124 11:51:46.984103 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:51:46.988820 kubelet[2895]: E0124 11:51:46.986097 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:47.122545 kubelet[2895]: I0124 11:51:47.121892 2895 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-dnscb" podStartSLOduration=84.121870014 podStartE2EDuration="1m24.121870014s" podCreationTimestamp="2026-01-24 11:50:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 11:51:47.118142129 +0000 UTC m=+86.639934845" watchObservedRunningTime="2026-01-24 11:51:47.121870014 +0000 UTC m=+86.643662721" Jan 24 11:51:47.158000 audit[5352]: NETFILTER_CFG table=filter:144 family=2 entries=14 op=nft_register_rule pid=5352 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:47.166836 kernel: kauditd_printk_skb: 202 callbacks suppressed Jan 24 11:51:47.167021 kernel: audit: type=1325 audit(1769255507.158:747): table=filter:144 family=2 entries=14 op=nft_register_rule pid=5352 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:47.158000 audit[5352]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffcc32e1ec0 a2=0 a3=7ffcc32e1eac items=0 ppid=3040 pid=5352 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:47.199681 kernel: audit: type=1300 audit(1769255507.158:747): arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffcc32e1ec0 a2=0 a3=7ffcc32e1eac items=0 ppid=3040 pid=5352 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:47.200169 kernel: audit: type=1327 audit(1769255507.158:747): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:47.158000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:47.228000 audit[5352]: NETFILTER_CFG table=nat:145 family=2 entries=56 op=nft_register_chain pid=5352 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:47.228000 audit[5352]: SYSCALL arch=c000003e syscall=46 success=yes exit=19860 a0=3 a1=7ffcc32e1ec0 a2=0 a3=7ffcc32e1eac items=0 ppid=3040 pid=5352 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:47.254860 kernel: audit: type=1325 audit(1769255507.228:748): table=nat:145 family=2 entries=56 op=nft_register_chain pid=5352 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:51:47.254989 kernel: audit: type=1300 audit(1769255507.228:748): arch=c000003e syscall=46 success=yes exit=19860 a0=3 a1=7ffcc32e1ec0 a2=0 a3=7ffcc32e1eac items=0 ppid=3040 pid=5352 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:47.255053 kernel: audit: type=1327 audit(1769255507.228:748): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:47.228000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:51:47.988430 kubelet[2895]: E0124 11:51:47.987944 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:51:47.989569 kubelet[2895]: E0124 11:51:47.988860 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:48.131154 kubelet[2895]: E0124 11:51:48.131022 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:48.989076 kubelet[2895]: E0124 11:51:48.988882 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:49.562514 systemd[1]: Started sshd@7-10.0.0.67:22-10.0.0.1:42144.service - OpenSSH per-connection server daemon (10.0.0.1:42144). Jan 24 11:51:49.562000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.67:22-10.0.0.1:42144 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:51:49.578572 kernel: audit: type=1130 audit(1769255509.562:749): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.67:22-10.0.0.1:42144 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:51:49.900407 sshd[5364]: Accepted publickey for core from 10.0.0.1 port 42144 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:51:49.899000 audit[5364]: USER_ACCT pid=5364 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:49.913508 sshd-session[5364]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:51:49.916647 kernel: audit: type=1101 audit(1769255509.899:750): pid=5364 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:49.907000 audit[5364]: CRED_ACQ pid=5364 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:49.936814 systemd-logind[1621]: New session 9 of user core. Jan 24 11:51:49.939934 kernel: audit: type=1103 audit(1769255509.907:751): pid=5364 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:49.940004 kernel: audit: type=1006 audit(1769255509.907:752): pid=5364 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=9 res=1 Jan 24 11:51:49.907000 audit[5364]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcb36a2690 a2=3 a3=0 items=0 ppid=1 pid=5364 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=9 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:49.907000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:51:49.955714 systemd[1]: Started session-9.scope - Session 9 of User core. Jan 24 11:51:49.978000 audit[5364]: USER_START pid=5364 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:49.986000 audit[5368]: CRED_ACQ pid=5368 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:50.271668 sshd[5368]: Connection closed by 10.0.0.1 port 42144 Jan 24 11:51:50.272645 sshd-session[5364]: pam_unix(sshd:session): session closed for user core Jan 24 11:51:50.275000 audit[5364]: USER_END pid=5364 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:50.276000 audit[5364]: CRED_DISP pid=5364 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:50.282782 systemd[1]: sshd@7-10.0.0.67:22-10.0.0.1:42144.service: Deactivated successfully. Jan 24 11:51:50.282000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.67:22-10.0.0.1:42144 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:51:50.286048 systemd[1]: session-9.scope: Deactivated successfully. Jan 24 11:51:50.291168 systemd-logind[1621]: Session 9 logged out. Waiting for processes to exit. Jan 24 11:51:50.302061 systemd-logind[1621]: Removed session 9. Jan 24 11:51:52.135554 containerd[1647]: time="2026-01-24T11:51:52.135364107Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 24 11:51:52.246307 containerd[1647]: time="2026-01-24T11:51:52.246043920Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:51:52.258073 containerd[1647]: time="2026-01-24T11:51:52.257161693Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 24 11:51:52.258073 containerd[1647]: time="2026-01-24T11:51:52.257389928Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 24 11:51:52.258381 kubelet[2895]: E0124 11:51:52.257566 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:51:52.258381 kubelet[2895]: E0124 11:51:52.257681 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:51:52.258381 kubelet[2895]: E0124 11:51:52.257848 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:d91f0bbcac534ce79a38827d5dd29c41,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5n6cl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c99cdc8ff-zh5gl_calico-system(5cdab4ea-dcce-4467-b450-26abe2130c8d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 24 11:51:52.262547 containerd[1647]: time="2026-01-24T11:51:52.262430032Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 24 11:51:52.359829 containerd[1647]: time="2026-01-24T11:51:52.359607544Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:51:52.397754 containerd[1647]: time="2026-01-24T11:51:52.397440456Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 24 11:51:52.397754 containerd[1647]: time="2026-01-24T11:51:52.397605083Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 24 11:51:52.399249 kubelet[2895]: E0124 11:51:52.398344 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:51:52.399249 kubelet[2895]: E0124 11:51:52.398521 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:51:52.403769 kubelet[2895]: E0124 11:51:52.401154 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5n6cl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c99cdc8ff-zh5gl_calico-system(5cdab4ea-dcce-4467-b450-26abe2130c8d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 24 11:51:52.404326 kubelet[2895]: E0124 11:51:52.404084 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:51:53.148776 containerd[1647]: time="2026-01-24T11:51:53.147377043Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 24 11:51:53.264046 containerd[1647]: time="2026-01-24T11:51:53.263772873Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:51:53.279819 containerd[1647]: time="2026-01-24T11:51:53.279548719Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 24 11:51:53.279819 containerd[1647]: time="2026-01-24T11:51:53.279709402Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 24 11:51:53.283776 kubelet[2895]: E0124 11:51:53.282484 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:51:53.283776 kubelet[2895]: E0124 11:51:53.282553 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:51:53.284461 kubelet[2895]: E0124 11:51:53.284092 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p5kgf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-66955b7fc7-xmscg_calico-system(41fe2d74-dafe-45a2-82d2-5fd56447e59d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 24 11:51:53.285948 kubelet[2895]: E0124 11:51:53.285862 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:51:55.139390 kubelet[2895]: E0124 11:51:55.133121 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:51:55.303702 systemd[1]: Started sshd@8-10.0.0.67:22-10.0.0.1:57918.service - OpenSSH per-connection server daemon (10.0.0.1:57918). Jan 24 11:51:55.322419 kernel: kauditd_printk_skb: 7 callbacks suppressed Jan 24 11:51:55.322732 kernel: audit: type=1130 audit(1769255515.303:758): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.67:22-10.0.0.1:57918 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:51:55.303000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.67:22-10.0.0.1:57918 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:51:55.459834 sshd[5384]: Accepted publickey for core from 10.0.0.1 port 57918 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:51:55.452000 audit[5384]: USER_ACCT pid=5384 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:55.485755 kernel: audit: type=1101 audit(1769255515.452:759): pid=5384 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:55.488846 sshd-session[5384]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:51:55.484000 audit[5384]: CRED_ACQ pid=5384 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:55.539006 kernel: audit: type=1103 audit(1769255515.484:760): pid=5384 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:55.539220 kernel: audit: type=1006 audit(1769255515.484:761): pid=5384 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=10 res=1 Jan 24 11:51:55.539351 kernel: audit: type=1300 audit(1769255515.484:761): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe5b26db80 a2=3 a3=0 items=0 ppid=1 pid=5384 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:55.484000 audit[5384]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe5b26db80 a2=3 a3=0 items=0 ppid=1 pid=5384 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:51:55.538367 systemd-logind[1621]: New session 10 of user core. Jan 24 11:51:55.484000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:51:55.561389 kernel: audit: type=1327 audit(1769255515.484:761): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:51:55.569133 systemd[1]: Started session-10.scope - Session 10 of User core. Jan 24 11:51:55.587000 audit[5384]: USER_START pid=5384 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:55.631890 kernel: audit: type=1105 audit(1769255515.587:762): pid=5384 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:55.632042 kernel: audit: type=1103 audit(1769255515.595:763): pid=5390 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:55.595000 audit[5390]: CRED_ACQ pid=5390 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:55.864373 sshd[5390]: Connection closed by 10.0.0.1 port 57918 Jan 24 11:51:55.871584 sshd-session[5384]: pam_unix(sshd:session): session closed for user core Jan 24 11:51:55.874000 audit[5384]: USER_END pid=5384 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:55.892018 systemd[1]: sshd@8-10.0.0.67:22-10.0.0.1:57918.service: Deactivated successfully. Jan 24 11:51:55.897084 kernel: audit: type=1106 audit(1769255515.874:764): pid=5384 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:55.897212 kernel: audit: type=1104 audit(1769255515.877:765): pid=5384 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:55.877000 audit[5384]: CRED_DISP pid=5384 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:51:55.909558 systemd[1]: session-10.scope: Deactivated successfully. Jan 24 11:51:55.891000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.67:22-10.0.0.1:57918 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:51:55.923344 systemd-logind[1621]: Session 10 logged out. Waiting for processes to exit. Jan 24 11:51:55.930139 systemd-logind[1621]: Removed session 10. Jan 24 11:51:59.142195 containerd[1647]: time="2026-01-24T11:51:59.141974141Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 24 11:51:59.285996 containerd[1647]: time="2026-01-24T11:51:59.285937226Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:51:59.302040 containerd[1647]: time="2026-01-24T11:51:59.301536891Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 24 11:51:59.302040 containerd[1647]: time="2026-01-24T11:51:59.301993503Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 24 11:51:59.305130 kubelet[2895]: E0124 11:51:59.304371 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:51:59.305130 kubelet[2895]: E0124 11:51:59.304438 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:51:59.305130 kubelet[2895]: E0124 11:51:59.304591 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lqn89,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2g45n_calico-system(999c48c5-f3b1-4bfd-91f4-26cafee1eef0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 24 11:51:59.327978 containerd[1647]: time="2026-01-24T11:51:59.327491375Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 24 11:51:59.445555 containerd[1647]: time="2026-01-24T11:51:59.444993332Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:51:59.453585 containerd[1647]: time="2026-01-24T11:51:59.451887625Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 24 11:51:59.453585 containerd[1647]: time="2026-01-24T11:51:59.452037675Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 24 11:51:59.453806 kubelet[2895]: E0124 11:51:59.452531 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:51:59.453806 kubelet[2895]: E0124 11:51:59.452607 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:51:59.453806 kubelet[2895]: E0124 11:51:59.452814 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lqn89,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2g45n_calico-system(999c48c5-f3b1-4bfd-91f4-26cafee1eef0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 24 11:51:59.462389 kubelet[2895]: E0124 11:51:59.459893 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:52:00.139077 containerd[1647]: time="2026-01-24T11:52:00.138517730Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:52:00.205346 containerd[1647]: time="2026-01-24T11:52:00.205042206Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:52:00.212790 containerd[1647]: time="2026-01-24T11:52:00.212602762Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:52:00.212790 containerd[1647]: time="2026-01-24T11:52:00.212784061Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:52:00.213052 kubelet[2895]: E0124 11:52:00.212959 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:52:00.213052 kubelet[2895]: E0124 11:52:00.213010 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:52:00.213420 kubelet[2895]: E0124 11:52:00.213168 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l682s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-f485b4455-qhcml_calico-apiserver(c28e9f44-9bc9-4328-bf7a-f4f9d016a87e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:52:00.215453 kubelet[2895]: E0124 11:52:00.214890 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:52:15.193686 systemd[1]: Started sshd@9-10.0.0.67:22-10.0.0.1:57932.service - OpenSSH per-connection server daemon (10.0.0.1:57932). Jan 24 11:52:15.253372 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:52:15.267972 kernel: audit: type=1130 audit(1769255535.192:767): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.67:22-10.0.0.1:57932 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:15.192000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.67:22-10.0.0.1:57932 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:15.613000 audit: BPF prog-id=259 op=LOAD Jan 24 11:52:15.600144 systemd[1]: cri-containerd-b0c4c98e002c741713e61e61888663dbd775dc4bae23fe7d0b86e98503190f63.scope: Deactivated successfully. Jan 24 11:52:15.601042 systemd[1]: cri-containerd-b0c4c98e002c741713e61e61888663dbd775dc4bae23fe7d0b86e98503190f63.scope: Consumed 11.104s CPU time, 71.2M memory peak, 10.9M read from disk. Jan 24 11:52:15.620383 kernel: audit: type=1334 audit(1769255535.613:768): prog-id=259 op=LOAD Jan 24 11:52:15.615000 audit: BPF prog-id=96 op=UNLOAD Jan 24 11:52:15.615000 audit: BPF prog-id=100 op=UNLOAD Jan 24 11:52:15.629380 kernel: audit: type=1334 audit(1769255535.615:769): prog-id=96 op=UNLOAD Jan 24 11:52:15.629505 kernel: audit: type=1334 audit(1769255535.615:770): prog-id=100 op=UNLOAD Jan 24 11:52:15.615000 audit: BPF prog-id=81 op=UNLOAD Jan 24 11:52:15.651531 kernel: audit: type=1334 audit(1769255535.615:771): prog-id=81 op=UNLOAD Jan 24 11:52:15.690000 audit: BPF prog-id=260 op=LOAD Jan 24 11:52:15.703448 kernel: audit: type=1334 audit(1769255535.690:772): prog-id=260 op=LOAD Jan 24 11:52:15.703565 kernel: audit: type=1334 audit(1769255535.690:773): prog-id=83 op=UNLOAD Jan 24 11:52:15.690000 audit: BPF prog-id=83 op=UNLOAD Jan 24 11:52:15.688694 systemd[1]: cri-containerd-51df99e807c575426a0b1b0edb58bf4d25dbd42f2bb9cd6e380438339a94e0ef.scope: Deactivated successfully. Jan 24 11:52:15.713000 audit: BPF prog-id=101 op=UNLOAD Jan 24 11:52:15.689936 systemd[1]: cri-containerd-51df99e807c575426a0b1b0edb58bf4d25dbd42f2bb9cd6e380438339a94e0ef.scope: Consumed 6.883s CPU time, 27.2M memory peak, 5.5M read from disk. Jan 24 11:52:15.724209 kernel: audit: type=1334 audit(1769255535.713:774): prog-id=101 op=UNLOAD Jan 24 11:52:15.724415 kernel: audit: type=1334 audit(1769255535.713:775): prog-id=105 op=UNLOAD Jan 24 11:52:15.713000 audit: BPF prog-id=105 op=UNLOAD Jan 24 11:52:15.834684 containerd[1647]: time="2026-01-24T11:52:15.828328089Z" level=info msg="received container exit event container_id:\"51df99e807c575426a0b1b0edb58bf4d25dbd42f2bb9cd6e380438339a94e0ef\" id:\"51df99e807c575426a0b1b0edb58bf4d25dbd42f2bb9cd6e380438339a94e0ef\" pid:2726 exit_status:1 exited_at:{seconds:1769255535 nanos:723553701}" Jan 24 11:52:15.855343 containerd[1647]: time="2026-01-24T11:52:15.854990317Z" level=info msg="received container exit event container_id:\"b0c4c98e002c741713e61e61888663dbd775dc4bae23fe7d0b86e98503190f63\" id:\"b0c4c98e002c741713e61e61888663dbd775dc4bae23fe7d0b86e98503190f63\" pid:2703 exit_status:1 exited_at:{seconds:1769255535 nanos:678787764}" Jan 24 11:52:15.880000 audit[5412]: USER_ACCT pid=5412 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:15.886976 sshd[5412]: Accepted publickey for core from 10.0.0.1 port 57932 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:52:15.913687 sshd-session[5412]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:52:15.923563 kernel: audit: type=1101 audit(1769255535.880:776): pid=5412 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:15.887000 audit[5412]: CRED_ACQ pid=5412 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:15.893000 audit[5412]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd6efb6150 a2=3 a3=0 items=0 ppid=1 pid=5412 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=11 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:15.893000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:16.017618 kubelet[2895]: E0124 11:52:16.016142 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:16.029579 systemd-logind[1621]: New session 11 of user core. Jan 24 11:52:16.053676 systemd[1]: Started session-11.scope - Session 11 of User core. Jan 24 11:52:16.128246 kubelet[2895]: E0124 11:52:16.118990 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:52:16.129153 containerd[1647]: time="2026-01-24T11:52:16.128941421Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 24 11:52:16.149671 kubelet[2895]: E0124 11:52:16.139488 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:52:16.192000 audit[5412]: USER_START pid=5412 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:16.211000 audit[5426]: CRED_ACQ pid=5426 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:16.215766 kubelet[2895]: E0124 11:52:16.213191 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:16.585070 containerd[1647]: time="2026-01-24T11:52:16.551513261Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:52:16.611062 containerd[1647]: time="2026-01-24T11:52:16.610548210Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 24 11:52:16.613585 containerd[1647]: time="2026-01-24T11:52:16.613488114Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 24 11:52:16.616980 kubelet[2895]: E0124 11:52:16.614766 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:52:16.616980 kubelet[2895]: E0124 11:52:16.614937 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:52:16.616980 kubelet[2895]: E0124 11:52:16.616009 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k2g54,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-jfngl_calico-system(293f1d4a-795c-44fc-bda8-3c550f56df72): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 24 11:52:16.622257 containerd[1647]: time="2026-01-24T11:52:16.622218353Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 24 11:52:16.622757 kubelet[2895]: E0124 11:52:16.622673 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:52:16.656492 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b0c4c98e002c741713e61e61888663dbd775dc4bae23fe7d0b86e98503190f63-rootfs.mount: Deactivated successfully. Jan 24 11:52:16.747326 containerd[1647]: time="2026-01-24T11:52:16.745414488Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:52:16.750367 containerd[1647]: time="2026-01-24T11:52:16.748530902Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 24 11:52:16.804489 containerd[1647]: time="2026-01-24T11:52:16.801016981Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 24 11:52:16.804655 kubelet[2895]: E0124 11:52:16.802219 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:52:16.815977 kubelet[2895]: E0124 11:52:16.806615 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:52:16.815977 kubelet[2895]: E0124 11:52:16.807159 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:d91f0bbcac534ce79a38827d5dd29c41,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5n6cl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c99cdc8ff-zh5gl_calico-system(5cdab4ea-dcce-4467-b450-26abe2130c8d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 24 11:52:16.816386 containerd[1647]: time="2026-01-24T11:52:16.807579276Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:52:16.862493 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-51df99e807c575426a0b1b0edb58bf4d25dbd42f2bb9cd6e380438339a94e0ef-rootfs.mount: Deactivated successfully. Jan 24 11:52:16.887115 kubelet[2895]: I0124 11:52:16.887051 2895 scope.go:117] "RemoveContainer" containerID="b0c4c98e002c741713e61e61888663dbd775dc4bae23fe7d0b86e98503190f63" Jan 24 11:52:16.887466 kubelet[2895]: E0124 11:52:16.887185 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:16.956959 containerd[1647]: time="2026-01-24T11:52:16.956631813Z" level=info msg="CreateContainer within sandbox \"4a95108dd84d5620bbce07e08082acff25e837447724f0b2cb1ea1fc3e7ddde6\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Jan 24 11:52:16.967680 containerd[1647]: time="2026-01-24T11:52:16.967625488Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:52:16.984089 containerd[1647]: time="2026-01-24T11:52:16.979696822Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:52:16.984610 containerd[1647]: time="2026-01-24T11:52:16.979764658Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:52:16.989858 kubelet[2895]: E0124 11:52:16.989439 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:52:16.990650 kubelet[2895]: E0124 11:52:16.990610 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:52:17.011165 kubelet[2895]: E0124 11:52:16.998970 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r4s6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-f485b4455-nc2lh_calico-apiserver(d47a09e2-52ab-4e49-942e-05f2cedb67ae): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:52:17.011165 kubelet[2895]: E0124 11:52:17.010999 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:52:17.014939 containerd[1647]: time="2026-01-24T11:52:16.999514048Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 24 11:52:17.090102 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3752851210.mount: Deactivated successfully. Jan 24 11:52:17.096437 containerd[1647]: time="2026-01-24T11:52:17.090699863Z" level=info msg="Container 79e0c9658df79750886e6ff8534b5c09bf33181a6bc0e99289aac7e02ea05193: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:52:17.120902 containerd[1647]: time="2026-01-24T11:52:17.119563468Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:52:17.133467 containerd[1647]: time="2026-01-24T11:52:17.133357646Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 24 11:52:17.133637 containerd[1647]: time="2026-01-24T11:52:17.133553652Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 24 11:52:17.139380 kubelet[2895]: E0124 11:52:17.136957 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:52:17.139380 kubelet[2895]: E0124 11:52:17.137020 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:52:17.139380 kubelet[2895]: E0124 11:52:17.138041 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p5kgf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-66955b7fc7-xmscg_calico-system(41fe2d74-dafe-45a2-82d2-5fd56447e59d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 24 11:52:17.149131 kubelet[2895]: E0124 11:52:17.147207 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:52:17.149246 containerd[1647]: time="2026-01-24T11:52:17.148612397Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 24 11:52:17.157647 sshd[5426]: Connection closed by 10.0.0.1 port 57932 Jan 24 11:52:17.159526 sshd-session[5412]: pam_unix(sshd:session): session closed for user core Jan 24 11:52:17.172000 audit[5412]: USER_END pid=5412 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:17.173000 audit[5412]: CRED_DISP pid=5412 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:17.185030 systemd-logind[1621]: Session 11 logged out. Waiting for processes to exit. Jan 24 11:52:17.191873 systemd[1]: sshd@9-10.0.0.67:22-10.0.0.1:57932.service: Deactivated successfully. Jan 24 11:52:17.191000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.67:22-10.0.0.1:57932 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:17.199221 systemd[1]: session-11.scope: Deactivated successfully. Jan 24 11:52:17.212210 systemd-logind[1621]: Removed session 11. Jan 24 11:52:17.260157 containerd[1647]: time="2026-01-24T11:52:17.260091374Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:52:17.264378 containerd[1647]: time="2026-01-24T11:52:17.264336101Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 24 11:52:17.264823 containerd[1647]: time="2026-01-24T11:52:17.264533419Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 24 11:52:17.273751 containerd[1647]: time="2026-01-24T11:52:17.273172024Z" level=info msg="CreateContainer within sandbox \"4a95108dd84d5620bbce07e08082acff25e837447724f0b2cb1ea1fc3e7ddde6\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"79e0c9658df79750886e6ff8534b5c09bf33181a6bc0e99289aac7e02ea05193\"" Jan 24 11:52:17.274648 kubelet[2895]: E0124 11:52:17.273629 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:52:17.274648 kubelet[2895]: E0124 11:52:17.273681 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:52:17.275597 containerd[1647]: time="2026-01-24T11:52:17.274684637Z" level=info msg="StartContainer for \"79e0c9658df79750886e6ff8534b5c09bf33181a6bc0e99289aac7e02ea05193\"" Jan 24 11:52:17.276093 kubelet[2895]: E0124 11:52:17.275611 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5n6cl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c99cdc8ff-zh5gl_calico-system(5cdab4ea-dcce-4467-b450-26abe2130c8d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 24 11:52:17.281016 kubelet[2895]: E0124 11:52:17.280833 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:52:17.295808 containerd[1647]: time="2026-01-24T11:52:17.295504799Z" level=info msg="connecting to shim 79e0c9658df79750886e6ff8534b5c09bf33181a6bc0e99289aac7e02ea05193" address="unix:///run/containerd/s/e778a2d245ed037f51735cce3d2b99c0d59843448210aad0488703e6956a3587" protocol=ttrpc version=3 Jan 24 11:52:17.416070 systemd[1]: Started cri-containerd-79e0c9658df79750886e6ff8534b5c09bf33181a6bc0e99289aac7e02ea05193.scope - libcontainer container 79e0c9658df79750886e6ff8534b5c09bf33181a6bc0e99289aac7e02ea05193. Jan 24 11:52:17.543000 audit: BPF prog-id=261 op=LOAD Jan 24 11:52:17.551000 audit: BPF prog-id=262 op=LOAD Jan 24 11:52:17.551000 audit[5480]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=2571 pid=5480 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:17.551000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3739653063393635386466373937353038383665366666383533346235 Jan 24 11:52:17.551000 audit: BPF prog-id=262 op=UNLOAD Jan 24 11:52:17.551000 audit[5480]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2571 pid=5480 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:17.551000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3739653063393635386466373937353038383665366666383533346235 Jan 24 11:52:17.551000 audit: BPF prog-id=263 op=LOAD Jan 24 11:52:17.551000 audit[5480]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=2571 pid=5480 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:17.551000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3739653063393635386466373937353038383665366666383533346235 Jan 24 11:52:17.551000 audit: BPF prog-id=264 op=LOAD Jan 24 11:52:17.551000 audit[5480]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=2571 pid=5480 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:17.551000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3739653063393635386466373937353038383665366666383533346235 Jan 24 11:52:17.551000 audit: BPF prog-id=264 op=UNLOAD Jan 24 11:52:17.551000 audit[5480]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2571 pid=5480 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:17.551000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3739653063393635386466373937353038383665366666383533346235 Jan 24 11:52:17.551000 audit: BPF prog-id=263 op=UNLOAD Jan 24 11:52:17.551000 audit[5480]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2571 pid=5480 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:17.551000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3739653063393635386466373937353038383665366666383533346235 Jan 24 11:52:17.551000 audit: BPF prog-id=265 op=LOAD Jan 24 11:52:17.551000 audit[5480]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=2571 pid=5480 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:17.551000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3739653063393635386466373937353038383665366666383533346235 Jan 24 11:52:17.581215 kubelet[2895]: E0124 11:52:17.581172 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:17.793568 containerd[1647]: time="2026-01-24T11:52:17.786385802Z" level=info msg="StartContainer for \"79e0c9658df79750886e6ff8534b5c09bf33181a6bc0e99289aac7e02ea05193\" returns successfully" Jan 24 11:52:17.911908 kubelet[2895]: I0124 11:52:17.911226 2895 scope.go:117] "RemoveContainer" containerID="51df99e807c575426a0b1b0edb58bf4d25dbd42f2bb9cd6e380438339a94e0ef" Jan 24 11:52:17.911908 kubelet[2895]: E0124 11:52:17.911381 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:17.928868 containerd[1647]: time="2026-01-24T11:52:17.928815165Z" level=info msg="CreateContainer within sandbox \"04a2d0581942060efb504e6b881240a2eb48d2748a75d386f525d7b4a3d239d3\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Jan 24 11:52:17.937212 kubelet[2895]: E0124 11:52:17.937030 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:18.026795 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount281743188.mount: Deactivated successfully. Jan 24 11:52:18.028965 containerd[1647]: time="2026-01-24T11:52:18.028145062Z" level=info msg="Container aa2867b8a0e896b6e08b53743281633d5315a1340cb4a0f81ba946682b182035: CDI devices from CRI Config.CDIDevices: []" Jan 24 11:52:18.079822 containerd[1647]: time="2026-01-24T11:52:18.078608230Z" level=info msg="CreateContainer within sandbox \"04a2d0581942060efb504e6b881240a2eb48d2748a75d386f525d7b4a3d239d3\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"aa2867b8a0e896b6e08b53743281633d5315a1340cb4a0f81ba946682b182035\"" Jan 24 11:52:18.089222 containerd[1647]: time="2026-01-24T11:52:18.085402174Z" level=info msg="StartContainer for \"aa2867b8a0e896b6e08b53743281633d5315a1340cb4a0f81ba946682b182035\"" Jan 24 11:52:18.110515 containerd[1647]: time="2026-01-24T11:52:18.098954064Z" level=info msg="connecting to shim aa2867b8a0e896b6e08b53743281633d5315a1340cb4a0f81ba946682b182035" address="unix:///run/containerd/s/b10f8ef1c7667849f1dc396b7ff875f13fc04875b472f1efaf696ed9b8913a98" protocol=ttrpc version=3 Jan 24 11:52:18.190545 systemd[1]: Started cri-containerd-aa2867b8a0e896b6e08b53743281633d5315a1340cb4a0f81ba946682b182035.scope - libcontainer container aa2867b8a0e896b6e08b53743281633d5315a1340cb4a0f81ba946682b182035. Jan 24 11:52:18.279000 audit: BPF prog-id=266 op=LOAD Jan 24 11:52:18.280000 audit: BPF prog-id=267 op=LOAD Jan 24 11:52:18.280000 audit[5510]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=2580 pid=5510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:18.280000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6161323836376238613065383936623665303862353337343332383136 Jan 24 11:52:18.280000 audit: BPF prog-id=267 op=UNLOAD Jan 24 11:52:18.280000 audit[5510]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2580 pid=5510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:18.280000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6161323836376238613065383936623665303862353337343332383136 Jan 24 11:52:18.280000 audit: BPF prog-id=268 op=LOAD Jan 24 11:52:18.280000 audit[5510]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=2580 pid=5510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:18.280000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6161323836376238613065383936623665303862353337343332383136 Jan 24 11:52:18.282000 audit: BPF prog-id=269 op=LOAD Jan 24 11:52:18.282000 audit[5510]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=2580 pid=5510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:18.282000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6161323836376238613065383936623665303862353337343332383136 Jan 24 11:52:18.283000 audit: BPF prog-id=269 op=UNLOAD Jan 24 11:52:18.283000 audit[5510]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2580 pid=5510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:18.283000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6161323836376238613065383936623665303862353337343332383136 Jan 24 11:52:18.283000 audit: BPF prog-id=268 op=UNLOAD Jan 24 11:52:18.283000 audit[5510]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2580 pid=5510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:18.283000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6161323836376238613065383936623665303862353337343332383136 Jan 24 11:52:18.283000 audit: BPF prog-id=270 op=LOAD Jan 24 11:52:18.283000 audit[5510]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=2580 pid=5510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:18.283000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6161323836376238613065383936623665303862353337343332383136 Jan 24 11:52:18.472599 containerd[1647]: time="2026-01-24T11:52:18.472465431Z" level=info msg="StartContainer for \"aa2867b8a0e896b6e08b53743281633d5315a1340cb4a0f81ba946682b182035\" returns successfully" Jan 24 11:52:18.950251 kubelet[2895]: E0124 11:52:18.950202 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:19.959430 kubelet[2895]: E0124 11:52:19.957887 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:20.973425 kubelet[2895]: E0124 11:52:20.973382 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:21.253677 kubelet[2895]: E0124 11:52:21.253116 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:22.189922 systemd[1]: Started sshd@10-10.0.0.67:22-10.0.0.1:47928.service - OpenSSH per-connection server daemon (10.0.0.1:47928). Jan 24 11:52:22.203586 kernel: kauditd_printk_skb: 53 callbacks suppressed Jan 24 11:52:22.203802 kernel: audit: type=1130 audit(1769255542.189:800): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.67:22-10.0.0.1:47928 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:22.189000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.67:22-10.0.0.1:47928 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:22.399355 sshd[5559]: Accepted publickey for core from 10.0.0.1 port 47928 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:52:22.398000 audit[5559]: USER_ACCT pid=5559 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:22.402881 sshd-session[5559]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:52:22.422205 kernel: audit: type=1101 audit(1769255542.398:801): pid=5559 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:22.422353 kernel: audit: type=1103 audit(1769255542.398:802): pid=5559 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:22.398000 audit[5559]: CRED_ACQ pid=5559 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:22.422632 systemd-logind[1621]: New session 12 of user core. Jan 24 11:52:22.398000 audit[5559]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcedd3b720 a2=3 a3=0 items=0 ppid=1 pid=5559 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=12 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:22.441392 kernel: audit: type=1006 audit(1769255542.398:803): pid=5559 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=12 res=1 Jan 24 11:52:22.441457 kernel: audit: type=1300 audit(1769255542.398:803): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcedd3b720 a2=3 a3=0 items=0 ppid=1 pid=5559 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=12 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:22.452162 kernel: audit: type=1327 audit(1769255542.398:803): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:22.398000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:22.446846 systemd[1]: Started session-12.scope - Session 12 of User core. Jan 24 11:52:22.456000 audit[5559]: USER_START pid=5559 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:22.497171 kernel: audit: type=1105 audit(1769255542.456:804): pid=5559 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:22.497371 kernel: audit: type=1103 audit(1769255542.488:805): pid=5563 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:22.488000 audit[5563]: CRED_ACQ pid=5563 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:23.689376 sshd[5563]: Connection closed by 10.0.0.1 port 47928 Jan 24 11:52:23.691869 sshd-session[5559]: pam_unix(sshd:session): session closed for user core Jan 24 11:52:23.707000 audit[5559]: USER_END pid=5559 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:23.737612 systemd[1]: sshd@10-10.0.0.67:22-10.0.0.1:47928.service: Deactivated successfully. Jan 24 11:52:23.707000 audit[5559]: CRED_DISP pid=5559 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:23.752328 kernel: audit: type=1106 audit(1769255543.707:806): pid=5559 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:23.752465 kernel: audit: type=1104 audit(1769255543.707:807): pid=5559 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:23.735000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.67:22-10.0.0.1:47928 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:23.756440 systemd[1]: session-12.scope: Deactivated successfully. Jan 24 11:52:23.760184 systemd-logind[1621]: Session 12 logged out. Waiting for processes to exit. Jan 24 11:52:23.768090 systemd-logind[1621]: Removed session 12. Jan 24 11:52:29.123430 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:52:29.126040 kernel: audit: type=1130 audit(1769255549.078:809): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.67:22-10.0.0.1:38024 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:29.078000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.67:22-10.0.0.1:38024 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:29.078529 systemd[1]: Started sshd@11-10.0.0.67:22-10.0.0.1:38024.service - OpenSSH per-connection server daemon (10.0.0.1:38024). Jan 24 11:52:29.323118 kubelet[2895]: E0124 11:52:29.322317 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:52:29.335465 kubelet[2895]: E0124 11:52:29.335369 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:52:30.390678 containerd[1647]: time="2026-01-24T11:52:30.369127745Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 24 11:52:30.545109 kubelet[2895]: E0124 11:52:30.529476 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:52:30.977685 sshd[5581]: Accepted publickey for core from 10.0.0.1 port 38024 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:52:30.964000 audit[5581]: USER_ACCT pid=5581 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:30.990460 sshd-session[5581]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:52:31.007091 kubelet[2895]: E0124 11:52:31.006695 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:31.034446 kernel: audit: type=1101 audit(1769255550.964:810): pid=5581 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:31.034593 kernel: audit: type=1103 audit(1769255550.976:811): pid=5581 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:30.976000 audit[5581]: CRED_ACQ pid=5581 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:31.065795 kernel: audit: type=1006 audit(1769255550.976:812): pid=5581 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=13 res=1 Jan 24 11:52:31.065887 systemd-logind[1621]: New session 13 of user core. Jan 24 11:52:30.976000 audit[5581]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcfc741e10 a2=3 a3=0 items=0 ppid=1 pid=5581 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=13 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:31.108603 containerd[1647]: time="2026-01-24T11:52:31.106540424Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:52:31.124061 kernel: audit: type=1300 audit(1769255550.976:812): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcfc741e10 a2=3 a3=0 items=0 ppid=1 pid=5581 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=13 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:31.124258 kernel: audit: type=1327 audit(1769255550.976:812): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:30.976000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:31.124536 containerd[1647]: time="2026-01-24T11:52:31.124396278Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 24 11:52:31.125197 kubelet[2895]: E0124 11:52:31.125137 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:52:31.125432 kubelet[2895]: E0124 11:52:31.125398 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:52:31.126101 containerd[1647]: time="2026-01-24T11:52:31.126009678Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 24 11:52:31.128620 kubelet[2895]: E0124 11:52:31.127380 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lqn89,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2g45n_calico-system(999c48c5-f3b1-4bfd-91f4-26cafee1eef0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 24 11:52:31.136246 containerd[1647]: time="2026-01-24T11:52:31.132814445Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:52:31.143704 kubelet[2895]: E0124 11:52:31.143657 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:52:31.147943 systemd[1]: Started session-13.scope - Session 13 of User core. Jan 24 11:52:31.154000 audit[5581]: USER_START pid=5581 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:31.186503 kernel: audit: type=1105 audit(1769255551.154:813): pid=5581 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:31.170000 audit[5585]: CRED_ACQ pid=5585 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:31.234112 kernel: audit: type=1103 audit(1769255551.170:814): pid=5585 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:31.265850 containerd[1647]: time="2026-01-24T11:52:31.265754062Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:52:31.270601 containerd[1647]: time="2026-01-24T11:52:31.268632406Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:52:31.270601 containerd[1647]: time="2026-01-24T11:52:31.268808455Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:52:31.270601 containerd[1647]: time="2026-01-24T11:52:31.269860162Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 24 11:52:31.270927 kubelet[2895]: E0124 11:52:31.269123 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:52:31.270927 kubelet[2895]: E0124 11:52:31.269179 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:52:31.270927 kubelet[2895]: E0124 11:52:31.269479 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l682s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-f485b4455-qhcml_calico-apiserver(c28e9f44-9bc9-4328-bf7a-f4f9d016a87e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:52:31.272231 kubelet[2895]: E0124 11:52:31.272123 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:52:31.334699 kubelet[2895]: E0124 11:52:31.334437 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:31.391141 kubelet[2895]: E0124 11:52:31.391104 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:31.424447 containerd[1647]: time="2026-01-24T11:52:31.424238814Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:52:31.433661 containerd[1647]: time="2026-01-24T11:52:31.433537764Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 24 11:52:31.433894 containerd[1647]: time="2026-01-24T11:52:31.433606833Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 24 11:52:31.434461 kubelet[2895]: E0124 11:52:31.434383 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:52:31.434461 kubelet[2895]: E0124 11:52:31.434432 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:52:31.434966 kubelet[2895]: E0124 11:52:31.434903 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lqn89,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2g45n_calico-system(999c48c5-f3b1-4bfd-91f4-26cafee1eef0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 24 11:52:31.436637 kubelet[2895]: E0124 11:52:31.436603 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:52:31.640019 sshd[5585]: Connection closed by 10.0.0.1 port 38024 Jan 24 11:52:31.641216 sshd-session[5581]: pam_unix(sshd:session): session closed for user core Jan 24 11:52:31.655000 audit[5581]: USER_END pid=5581 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:31.665733 systemd[1]: sshd@11-10.0.0.67:22-10.0.0.1:38024.service: Deactivated successfully. Jan 24 11:52:31.673536 systemd[1]: session-13.scope: Deactivated successfully. Jan 24 11:52:31.692695 systemd-logind[1621]: Session 13 logged out. Waiting for processes to exit. Jan 24 11:52:31.705367 systemd-logind[1621]: Removed session 13. Jan 24 11:52:31.655000 audit[5581]: CRED_DISP pid=5581 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:31.729130 kernel: audit: type=1106 audit(1769255551.655:815): pid=5581 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:31.729371 kernel: audit: type=1104 audit(1769255551.655:816): pid=5581 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:31.661000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.67:22-10.0.0.1:38024 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:36.688624 systemd[1]: Started sshd@12-10.0.0.67:22-10.0.0.1:45006.service - OpenSSH per-connection server daemon (10.0.0.1:45006). Jan 24 11:52:36.706922 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:52:36.707054 kernel: audit: type=1130 audit(1769255556.686:818): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.67:22-10.0.0.1:45006 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:36.686000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.67:22-10.0.0.1:45006 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:36.906000 audit[5627]: USER_ACCT pid=5627 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:36.909679 sshd[5627]: Accepted publickey for core from 10.0.0.1 port 45006 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:52:36.915074 sshd-session[5627]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:52:36.938636 kernel: audit: type=1101 audit(1769255556.906:819): pid=5627 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:36.938834 kernel: audit: type=1103 audit(1769255556.907:820): pid=5627 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:36.907000 audit[5627]: CRED_ACQ pid=5627 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:36.948152 systemd-logind[1621]: New session 14 of user core. Jan 24 11:52:36.907000 audit[5627]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd01044590 a2=3 a3=0 items=0 ppid=1 pid=5627 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:36.972877 kernel: audit: type=1006 audit(1769255556.907:821): pid=5627 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=14 res=1 Jan 24 11:52:36.972953 kernel: audit: type=1300 audit(1769255556.907:821): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd01044590 a2=3 a3=0 items=0 ppid=1 pid=5627 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:36.977336 kernel: audit: type=1327 audit(1769255556.907:821): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:36.907000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:36.995884 systemd[1]: Started session-14.scope - Session 14 of User core. Jan 24 11:52:37.022000 audit[5627]: USER_START pid=5627 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:37.049423 kernel: audit: type=1105 audit(1769255557.022:822): pid=5627 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:37.030000 audit[5631]: CRED_ACQ pid=5631 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:37.071370 kernel: audit: type=1103 audit(1769255557.030:823): pid=5631 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:37.382057 sshd[5631]: Connection closed by 10.0.0.1 port 45006 Jan 24 11:52:37.385889 sshd-session[5627]: pam_unix(sshd:session): session closed for user core Jan 24 11:52:37.398000 audit[5627]: USER_END pid=5627 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:37.411394 systemd[1]: sshd@12-10.0.0.67:22-10.0.0.1:45006.service: Deactivated successfully. Jan 24 11:52:37.422484 systemd[1]: session-14.scope: Deactivated successfully. Jan 24 11:52:37.429402 systemd-logind[1621]: Session 14 logged out. Waiting for processes to exit. Jan 24 11:52:37.447542 systemd-logind[1621]: Removed session 14. Jan 24 11:52:37.452000 kernel: audit: type=1106 audit(1769255557.398:824): pid=5627 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:37.452109 kernel: audit: type=1104 audit(1769255557.398:825): pid=5627 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:37.398000 audit[5627]: CRED_DISP pid=5627 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:37.411000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.67:22-10.0.0.1:45006 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:41.143210 kubelet[2895]: E0124 11:52:41.142935 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:52:42.143057 containerd[1647]: time="2026-01-24T11:52:42.141924992Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:52:42.227009 containerd[1647]: time="2026-01-24T11:52:42.223786757Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:52:42.236943 containerd[1647]: time="2026-01-24T11:52:42.236169563Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:52:42.236943 containerd[1647]: time="2026-01-24T11:52:42.236416294Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:52:42.237950 kubelet[2895]: E0124 11:52:42.237330 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:52:42.237950 kubelet[2895]: E0124 11:52:42.237548 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:52:42.240451 kubelet[2895]: E0124 11:52:42.239551 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r4s6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-f485b4455-nc2lh_calico-apiserver(d47a09e2-52ab-4e49-942e-05f2cedb67ae): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:52:42.241661 kubelet[2895]: E0124 11:52:42.240981 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:52:42.426404 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:52:42.426538 kernel: audit: type=1130 audit(1769255562.416:827): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.67:22-10.0.0.1:59586 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:42.416000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.67:22-10.0.0.1:59586 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:42.418883 systemd[1]: Started sshd@13-10.0.0.67:22-10.0.0.1:59586.service - OpenSSH per-connection server daemon (10.0.0.1:59586). Jan 24 11:52:42.780000 audit[5653]: USER_ACCT pid=5653 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:42.791158 sshd[5653]: Accepted publickey for core from 10.0.0.1 port 59586 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:52:42.790607 sshd-session[5653]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:52:42.820785 kernel: audit: type=1101 audit(1769255562.780:828): pid=5653 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:42.783000 audit[5653]: CRED_ACQ pid=5653 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:42.826508 systemd-logind[1621]: New session 15 of user core. Jan 24 11:52:42.849783 kernel: audit: type=1103 audit(1769255562.783:829): pid=5653 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:42.850132 kernel: audit: type=1006 audit(1769255562.783:830): pid=5653 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=15 res=1 Jan 24 11:52:42.850220 kernel: audit: type=1300 audit(1769255562.783:830): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc0e1f4ec0 a2=3 a3=0 items=0 ppid=1 pid=5653 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:42.783000 audit[5653]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc0e1f4ec0 a2=3 a3=0 items=0 ppid=1 pid=5653 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:42.783000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:42.869339 systemd[1]: Started session-15.scope - Session 15 of User core. Jan 24 11:52:42.869539 kernel: audit: type=1327 audit(1769255562.783:830): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:42.887000 audit[5653]: USER_START pid=5653 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:42.897000 audit[5657]: CRED_ACQ pid=5657 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:42.951563 kernel: audit: type=1105 audit(1769255562.887:831): pid=5653 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:42.951712 kernel: audit: type=1103 audit(1769255562.897:832): pid=5657 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:43.142987 kubelet[2895]: E0124 11:52:43.140931 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:52:43.529103 sshd[5657]: Connection closed by 10.0.0.1 port 59586 Jan 24 11:52:43.528056 sshd-session[5653]: pam_unix(sshd:session): session closed for user core Jan 24 11:52:43.540000 audit[5653]: USER_END pid=5653 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:43.558241 systemd[1]: sshd@13-10.0.0.67:22-10.0.0.1:59586.service: Deactivated successfully. Jan 24 11:52:43.540000 audit[5653]: CRED_DISP pid=5653 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:43.570992 systemd[1]: session-15.scope: Deactivated successfully. Jan 24 11:52:43.577116 systemd-logind[1621]: Session 15 logged out. Waiting for processes to exit. Jan 24 11:52:43.596081 kernel: audit: type=1106 audit(1769255563.540:833): pid=5653 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:43.596334 kernel: audit: type=1104 audit(1769255563.540:834): pid=5653 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:43.556000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.67:22-10.0.0.1:59586 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:43.598249 systemd-logind[1621]: Removed session 15. Jan 24 11:52:43.619000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.67:22-10.0.0.1:59600 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:43.622390 systemd[1]: Started sshd@14-10.0.0.67:22-10.0.0.1:59600.service - OpenSSH per-connection server daemon (10.0.0.1:59600). Jan 24 11:52:43.948000 audit[5678]: USER_ACCT pid=5678 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:43.959000 audit[5678]: CRED_ACQ pid=5678 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:43.959000 audit[5678]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd03869a70 a2=3 a3=0 items=0 ppid=1 pid=5678 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:43.959000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:43.964654 sshd[5678]: Accepted publickey for core from 10.0.0.1 port 59600 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:52:43.962781 sshd-session[5678]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:52:43.995414 systemd-logind[1621]: New session 16 of user core. Jan 24 11:52:44.017775 systemd[1]: Started session-16.scope - Session 16 of User core. Jan 24 11:52:44.028000 audit[5678]: USER_START pid=5678 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:44.035000 audit[5682]: CRED_ACQ pid=5682 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:44.144526 kubelet[2895]: E0124 11:52:44.137367 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:52:44.602609 sshd[5682]: Connection closed by 10.0.0.1 port 59600 Jan 24 11:52:44.601768 sshd-session[5678]: pam_unix(sshd:session): session closed for user core Jan 24 11:52:44.628000 audit[5678]: USER_END pid=5678 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:44.629000 audit[5678]: CRED_DISP pid=5678 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:44.638850 systemd[1]: sshd@14-10.0.0.67:22-10.0.0.1:59600.service: Deactivated successfully. Jan 24 11:52:44.638000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.67:22-10.0.0.1:59600 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:44.646551 systemd[1]: session-16.scope: Deactivated successfully. Jan 24 11:52:44.655619 systemd-logind[1621]: Session 16 logged out. Waiting for processes to exit. Jan 24 11:52:44.672484 systemd[1]: Started sshd@15-10.0.0.67:22-10.0.0.1:59606.service - OpenSSH per-connection server daemon (10.0.0.1:59606). Jan 24 11:52:44.670000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.67:22-10.0.0.1:59606 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:44.686100 systemd-logind[1621]: Removed session 16. Jan 24 11:52:44.850000 audit[5695]: USER_ACCT pid=5695 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:44.853000 audit[5695]: CRED_ACQ pid=5695 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:44.853000 audit[5695]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcf056daf0 a2=3 a3=0 items=0 ppid=1 pid=5695 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=17 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:44.853000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:44.858705 sshd[5695]: Accepted publickey for core from 10.0.0.1 port 59606 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:52:44.858090 sshd-session[5695]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:52:44.871908 systemd-logind[1621]: New session 17 of user core. Jan 24 11:52:44.883688 systemd[1]: Started session-17.scope - Session 17 of User core. Jan 24 11:52:44.894000 audit[5695]: USER_START pid=5695 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:44.899000 audit[5699]: CRED_ACQ pid=5699 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:45.157439 containerd[1647]: time="2026-01-24T11:52:45.155994183Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 24 11:52:45.358600 containerd[1647]: time="2026-01-24T11:52:45.357900675Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:52:45.368397 containerd[1647]: time="2026-01-24T11:52:45.365686303Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 24 11:52:45.368397 containerd[1647]: time="2026-01-24T11:52:45.366015276Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 24 11:52:45.368590 kubelet[2895]: E0124 11:52:45.367021 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:52:45.368590 kubelet[2895]: E0124 11:52:45.367357 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:52:45.390630 kubelet[2895]: E0124 11:52:45.388629 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k2g54,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-jfngl_calico-system(293f1d4a-795c-44fc-bda8-3c550f56df72): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 24 11:52:45.417432 kubelet[2895]: E0124 11:52:45.416604 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:52:45.622072 sshd[5699]: Connection closed by 10.0.0.1 port 59606 Jan 24 11:52:45.618615 sshd-session[5695]: pam_unix(sshd:session): session closed for user core Jan 24 11:52:45.627000 audit[5695]: USER_END pid=5695 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:45.629000 audit[5695]: CRED_DISP pid=5695 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:45.638051 systemd-logind[1621]: Session 17 logged out. Waiting for processes to exit. Jan 24 11:52:45.640000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.67:22-10.0.0.1:59606 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:45.640886 systemd[1]: sshd@15-10.0.0.67:22-10.0.0.1:59606.service: Deactivated successfully. Jan 24 11:52:45.647621 systemd[1]: session-17.scope: Deactivated successfully. Jan 24 11:52:45.653184 systemd-logind[1621]: Removed session 17. Jan 24 11:52:46.136094 kubelet[2895]: E0124 11:52:46.136041 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:52:50.674023 kernel: kauditd_printk_skb: 23 callbacks suppressed Jan 24 11:52:50.674214 kernel: audit: type=1130 audit(1769255570.666:854): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.67:22-10.0.0.1:59622 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:50.666000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.67:22-10.0.0.1:59622 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:50.667631 systemd[1]: Started sshd@16-10.0.0.67:22-10.0.0.1:59622.service - OpenSSH per-connection server daemon (10.0.0.1:59622). Jan 24 11:52:50.874221 sshd[5712]: Accepted publickey for core from 10.0.0.1 port 59622 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:52:50.869000 audit[5712]: USER_ACCT pid=5712 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:50.890436 sshd-session[5712]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:52:50.917178 systemd-logind[1621]: New session 18 of user core. Jan 24 11:52:50.875000 audit[5712]: CRED_ACQ pid=5712 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:50.948688 kernel: audit: type=1101 audit(1769255570.869:855): pid=5712 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:50.948816 kernel: audit: type=1103 audit(1769255570.875:856): pid=5712 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:50.956466 kernel: audit: type=1006 audit(1769255570.875:857): pid=5712 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=18 res=1 Jan 24 11:52:50.964129 kernel: audit: type=1300 audit(1769255570.875:857): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcb6d3e020 a2=3 a3=0 items=0 ppid=1 pid=5712 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=18 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:50.875000 audit[5712]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcb6d3e020 a2=3 a3=0 items=0 ppid=1 pid=5712 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=18 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:50.961783 systemd[1]: Started session-18.scope - Session 18 of User core. Jan 24 11:52:51.022392 kernel: audit: type=1327 audit(1769255570.875:857): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:50.875000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:51.031346 kernel: audit: type=1105 audit(1769255571.025:858): pid=5712 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:51.025000 audit[5712]: USER_START pid=5712 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:51.077054 kernel: audit: type=1103 audit(1769255571.045:859): pid=5716 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:51.045000 audit[5716]: CRED_ACQ pid=5716 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:51.148387 kubelet[2895]: E0124 11:52:51.134657 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:51.544986 sshd[5716]: Connection closed by 10.0.0.1 port 59622 Jan 24 11:52:51.547542 sshd-session[5712]: pam_unix(sshd:session): session closed for user core Jan 24 11:52:51.551000 audit[5712]: USER_END pid=5712 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:51.577973 systemd[1]: sshd@16-10.0.0.67:22-10.0.0.1:59622.service: Deactivated successfully. Jan 24 11:52:51.589355 kernel: audit: type=1106 audit(1769255571.551:860): pid=5712 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:51.551000 audit[5712]: CRED_DISP pid=5712 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:51.591627 systemd[1]: session-18.scope: Deactivated successfully. Jan 24 11:52:51.621492 systemd-logind[1621]: Session 18 logged out. Waiting for processes to exit. Jan 24 11:52:51.632895 systemd-logind[1621]: Removed session 18. Jan 24 11:52:51.578000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.67:22-10.0.0.1:59622 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:51.647933 kernel: audit: type=1104 audit(1769255571.551:861): pid=5712 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:53.137381 kubelet[2895]: E0124 11:52:53.137219 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:52:55.132677 kubelet[2895]: E0124 11:52:55.132584 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:52:56.135763 kubelet[2895]: E0124 11:52:56.132625 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:52:56.578238 systemd[1]: Started sshd@17-10.0.0.67:22-10.0.0.1:51172.service - OpenSSH per-connection server daemon (10.0.0.1:51172). Jan 24 11:52:56.587205 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:52:56.587386 kernel: audit: type=1130 audit(1769255576.577:863): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-10.0.0.67:22-10.0.0.1:51172 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:56.577000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-10.0.0.67:22-10.0.0.1:51172 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:56.845000 audit[5733]: USER_ACCT pid=5733 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:56.850908 sshd-session[5733]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:52:56.852398 sshd[5733]: Accepted publickey for core from 10.0.0.1 port 51172 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:52:56.846000 audit[5733]: CRED_ACQ pid=5733 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:56.881006 kernel: audit: type=1101 audit(1769255576.845:864): pid=5733 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:56.881081 kernel: audit: type=1103 audit(1769255576.846:865): pid=5733 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:56.881134 kernel: audit: type=1006 audit(1769255576.846:866): pid=5733 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=19 res=1 Jan 24 11:52:56.888654 kernel: audit: type=1300 audit(1769255576.846:866): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe3ee49b90 a2=3 a3=0 items=0 ppid=1 pid=5733 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=19 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:56.846000 audit[5733]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe3ee49b90 a2=3 a3=0 items=0 ppid=1 pid=5733 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=19 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:52:56.896815 systemd-logind[1621]: New session 19 of user core. Jan 24 11:52:56.915053 kernel: audit: type=1327 audit(1769255576.846:866): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:56.846000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:52:56.927759 systemd[1]: Started session-19.scope - Session 19 of User core. Jan 24 11:52:56.934000 audit[5733]: USER_START pid=5733 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:56.943000 audit[5737]: CRED_ACQ pid=5737 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:56.994453 kernel: audit: type=1105 audit(1769255576.934:867): pid=5733 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:56.994596 kernel: audit: type=1103 audit(1769255576.943:868): pid=5737 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:57.148638 kubelet[2895]: E0124 11:52:57.139506 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:52:57.379427 sshd[5737]: Connection closed by 10.0.0.1 port 51172 Jan 24 11:52:57.383195 sshd-session[5733]: pam_unix(sshd:session): session closed for user core Jan 24 11:52:57.396000 audit[5733]: USER_END pid=5733 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:57.425159 systemd[1]: sshd@17-10.0.0.67:22-10.0.0.1:51172.service: Deactivated successfully. Jan 24 11:52:57.436998 kernel: audit: type=1106 audit(1769255577.396:869): pid=5733 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:57.437372 kernel: audit: type=1104 audit(1769255577.396:870): pid=5733 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:57.396000 audit[5733]: CRED_DISP pid=5733 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:52:57.435004 systemd[1]: session-19.scope: Deactivated successfully. Jan 24 11:52:57.449098 systemd-logind[1621]: Session 19 logged out. Waiting for processes to exit. Jan 24 11:52:57.424000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-10.0.0.67:22-10.0.0.1:51172 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:52:57.474662 systemd-logind[1621]: Removed session 19. Jan 24 11:52:59.145091 containerd[1647]: time="2026-01-24T11:52:59.145032829Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 24 11:52:59.232730 containerd[1647]: time="2026-01-24T11:52:59.232396399Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:52:59.249999 containerd[1647]: time="2026-01-24T11:52:59.248642841Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 24 11:52:59.249999 containerd[1647]: time="2026-01-24T11:52:59.249062315Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 24 11:52:59.252326 kubelet[2895]: E0124 11:52:59.252042 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:52:59.252974 kubelet[2895]: E0124 11:52:59.252832 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:52:59.257395 kubelet[2895]: E0124 11:52:59.257202 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p5kgf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-66955b7fc7-xmscg_calico-system(41fe2d74-dafe-45a2-82d2-5fd56447e59d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 24 11:52:59.258527 kubelet[2895]: E0124 11:52:59.258437 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:53:00.134797 kubelet[2895]: E0124 11:53:00.134557 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:53:00.139511 kubelet[2895]: E0124 11:53:00.139418 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:53:02.423012 systemd[1]: Started sshd@18-10.0.0.67:22-10.0.0.1:41440.service - OpenSSH per-connection server daemon (10.0.0.1:41440). Jan 24 11:53:02.422000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.67:22-10.0.0.1:41440 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:02.436182 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:53:02.436242 kernel: audit: type=1130 audit(1769255582.422:872): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.67:22-10.0.0.1:41440 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:02.744846 sshd[5757]: Accepted publickey for core from 10.0.0.1 port 41440 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:53:02.742000 audit[5757]: USER_ACCT pid=5757 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:02.762000 audit[5757]: CRED_ACQ pid=5757 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:02.792423 sshd-session[5757]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:53:02.817666 kernel: audit: type=1101 audit(1769255582.742:873): pid=5757 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:02.817752 kernel: audit: type=1103 audit(1769255582.762:874): pid=5757 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:02.821643 kernel: audit: type=1006 audit(1769255582.772:875): pid=5757 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=20 res=1 Jan 24 11:53:02.827136 systemd-logind[1621]: New session 20 of user core. Jan 24 11:53:02.772000 audit[5757]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcf7f8f730 a2=3 a3=0 items=0 ppid=1 pid=5757 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=20 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:02.772000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:02.892818 kernel: audit: type=1300 audit(1769255582.772:875): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffcf7f8f730 a2=3 a3=0 items=0 ppid=1 pid=5757 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=20 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:02.892994 kernel: audit: type=1327 audit(1769255582.772:875): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:02.904114 systemd[1]: Started session-20.scope - Session 20 of User core. Jan 24 11:53:02.924000 audit[5757]: USER_START pid=5757 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:02.969402 kernel: audit: type=1105 audit(1769255582.924:876): pid=5757 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:02.969500 kernel: audit: type=1103 audit(1769255582.940:877): pid=5761 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:02.940000 audit[5761]: CRED_ACQ pid=5761 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:03.299371 sshd[5761]: Connection closed by 10.0.0.1 port 41440 Jan 24 11:53:03.302395 sshd-session[5757]: pam_unix(sshd:session): session closed for user core Jan 24 11:53:03.305000 audit[5757]: USER_END pid=5757 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:03.323503 systemd-logind[1621]: Session 20 logged out. Waiting for processes to exit. Jan 24 11:53:03.324609 systemd[1]: sshd@18-10.0.0.67:22-10.0.0.1:41440.service: Deactivated successfully. Jan 24 11:53:03.337820 kernel: audit: type=1106 audit(1769255583.305:878): pid=5757 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:03.337973 kernel: audit: type=1104 audit(1769255583.305:879): pid=5757 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:03.305000 audit[5757]: CRED_DISP pid=5757 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:03.337968 systemd[1]: session-20.scope: Deactivated successfully. Jan 24 11:53:03.322000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.67:22-10.0.0.1:41440 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:03.348960 systemd-logind[1621]: Removed session 20. Jan 24 11:53:04.135545 kubelet[2895]: E0124 11:53:04.131874 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:53:12.898000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.67:22-10.0.0.1:41452 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:12.928510 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:53:12.928644 kernel: audit: type=1130 audit(1769255592.898:881): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.67:22-10.0.0.1:41452 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:12.900361 systemd[1]: Started sshd@19-10.0.0.67:22-10.0.0.1:41452.service - OpenSSH per-connection server daemon (10.0.0.1:41452). Jan 24 11:53:13.069078 kubelet[2895]: E0124 11:53:13.066012 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:53:13.094410 kubelet[2895]: E0124 11:53:13.088790 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:53:13.099951 kubelet[2895]: E0124 11:53:13.096701 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:53:13.100176 containerd[1647]: time="2026-01-24T11:53:13.097706741Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 24 11:53:13.334162 containerd[1647]: time="2026-01-24T11:53:13.334022300Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:53:13.344340 containerd[1647]: time="2026-01-24T11:53:13.343723742Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 24 11:53:13.344340 containerd[1647]: time="2026-01-24T11:53:13.344079176Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 24 11:53:13.345508 kubelet[2895]: E0124 11:53:13.344251 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:53:13.345508 kubelet[2895]: E0124 11:53:13.344412 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:53:13.345508 kubelet[2895]: E0124 11:53:13.344798 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lqn89,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2g45n_calico-system(999c48c5-f3b1-4bfd-91f4-26cafee1eef0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 24 11:53:13.347182 containerd[1647]: time="2026-01-24T11:53:13.346748677Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 24 11:53:13.363335 sshd[5802]: Accepted publickey for core from 10.0.0.1 port 41452 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:53:13.360000 audit[5802]: USER_ACCT pid=5802 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:13.381845 sshd-session[5802]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:53:13.384732 kernel: audit: type=1101 audit(1769255593.360:882): pid=5802 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:13.363000 audit[5802]: CRED_ACQ pid=5802 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:13.408593 kernel: audit: type=1103 audit(1769255593.363:883): pid=5802 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:13.439497 kernel: audit: type=1006 audit(1769255593.363:884): pid=5802 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=21 res=1 Jan 24 11:53:13.597787 kernel: audit: type=1300 audit(1769255593.363:884): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd498a28f0 a2=3 a3=0 items=0 ppid=1 pid=5802 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=21 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:13.763734 kernel: audit: type=1327 audit(1769255593.363:884): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:13.363000 audit[5802]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd498a28f0 a2=3 a3=0 items=0 ppid=1 pid=5802 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=21 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:13.363000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:15.205099 containerd[1647]: time="2026-01-24T11:53:14.093776467Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:53:15.205099 containerd[1647]: time="2026-01-24T11:53:14.149650843Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 24 11:53:15.205099 containerd[1647]: time="2026-01-24T11:53:14.344183548Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 24 11:53:13.868337 systemd-logind[1621]: New session 21 of user core. Jan 24 11:53:15.144847 systemd[1]: Started session-21.scope - Session 21 of User core. Jan 24 11:53:15.235186 kubelet[2895]: E0124 11:53:15.235068 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:53:15.237410 kubelet[2895]: E0124 11:53:15.235921 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:53:15.237410 kubelet[2895]: E0124 11:53:15.236217 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:d91f0bbcac534ce79a38827d5dd29c41,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5n6cl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c99cdc8ff-zh5gl_calico-system(5cdab4ea-dcce-4467-b450-26abe2130c8d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 24 11:53:15.264756 containerd[1647]: time="2026-01-24T11:53:15.260021207Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 24 11:53:15.266000 audit[5802]: USER_START pid=5802 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:15.313756 kubelet[2895]: E0124 11:53:15.271667 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:53:15.334605 kernel: audit: type=1105 audit(1769255595.266:885): pid=5802 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:15.332000 audit[5820]: CRED_ACQ pid=5820 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:15.401167 kernel: audit: type=1103 audit(1769255595.332:886): pid=5820 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:15.509461 containerd[1647]: time="2026-01-24T11:53:15.500771766Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:53:15.551399 containerd[1647]: time="2026-01-24T11:53:15.551330696Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 24 11:53:15.551696 containerd[1647]: time="2026-01-24T11:53:15.551609746Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 24 11:53:15.552374 kubelet[2895]: E0124 11:53:15.552235 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:53:15.553051 kubelet[2895]: E0124 11:53:15.552602 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:53:15.553051 kubelet[2895]: E0124 11:53:15.552877 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lqn89,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2g45n_calico-system(999c48c5-f3b1-4bfd-91f4-26cafee1eef0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 24 11:53:15.553594 containerd[1647]: time="2026-01-24T11:53:15.553551519Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 24 11:53:15.555357 kubelet[2895]: E0124 11:53:15.554804 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:53:15.674797 containerd[1647]: time="2026-01-24T11:53:15.674732133Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:53:15.683694 containerd[1647]: time="2026-01-24T11:53:15.683636989Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 24 11:53:15.686471 containerd[1647]: time="2026-01-24T11:53:15.683878368Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 24 11:53:15.687081 kubelet[2895]: E0124 11:53:15.687029 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:53:15.692730 kubelet[2895]: E0124 11:53:15.692629 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:53:15.930085 containerd[1647]: time="2026-01-24T11:53:15.693885681Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:53:16.255485 kubelet[2895]: E0124 11:53:15.917188 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5n6cl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c99cdc8ff-zh5gl_calico-system(5cdab4ea-dcce-4467-b450-26abe2130c8d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 24 11:53:16.258329 kubelet[2895]: E0124 11:53:16.257855 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:53:16.393583 containerd[1647]: time="2026-01-24T11:53:16.386889385Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:53:16.393583 containerd[1647]: time="2026-01-24T11:53:16.389587740Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:53:16.393583 containerd[1647]: time="2026-01-24T11:53:16.389690451Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:53:16.396224 kubelet[2895]: E0124 11:53:16.395787 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:53:16.396224 kubelet[2895]: E0124 11:53:16.395857 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:53:16.396224 kubelet[2895]: E0124 11:53:16.396111 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l682s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-f485b4455-qhcml_calico-apiserver(c28e9f44-9bc9-4328-bf7a-f4f9d016a87e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:53:16.397674 kubelet[2895]: E0124 11:53:16.397402 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:53:16.594331 sshd[5820]: Connection closed by 10.0.0.1 port 41452 Jan 24 11:53:16.595389 sshd-session[5802]: pam_unix(sshd:session): session closed for user core Jan 24 11:53:16.611000 audit[5802]: USER_END pid=5802 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:16.617796 systemd[1]: sshd@19-10.0.0.67:22-10.0.0.1:41452.service: Deactivated successfully. Jan 24 11:53:16.621604 systemd[1]: session-21.scope: Deactivated successfully. Jan 24 11:53:16.628060 systemd-logind[1621]: Session 21 logged out. Waiting for processes to exit. Jan 24 11:53:16.630603 systemd-logind[1621]: Removed session 21. Jan 24 11:53:16.647229 kernel: audit: type=1106 audit(1769255596.611:887): pid=5802 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:16.647681 kernel: audit: type=1104 audit(1769255596.611:888): pid=5802 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:16.611000 audit[5802]: CRED_DISP pid=5802 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:16.617000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.67:22-10.0.0.1:41452 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:21.636740 systemd[1]: Started sshd@20-10.0.0.67:22-10.0.0.1:50524.service - OpenSSH per-connection server daemon (10.0.0.1:50524). Jan 24 11:53:21.647406 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:53:21.647575 kernel: audit: type=1130 audit(1769255601.636:890): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.67:22-10.0.0.1:50524 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:21.636000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.67:22-10.0.0.1:50524 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:21.896984 kernel: audit: type=1101 audit(1769255601.864:891): pid=5845 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:21.864000 audit[5845]: USER_ACCT pid=5845 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:21.883779 sshd-session[5845]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:53:21.903540 sshd[5845]: Accepted publickey for core from 10.0.0.1 port 50524 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:53:21.878000 audit[5845]: CRED_ACQ pid=5845 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:21.940443 systemd-logind[1621]: New session 22 of user core. Jan 24 11:53:21.942468 kernel: audit: type=1103 audit(1769255601.878:892): pid=5845 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:21.942546 kernel: audit: type=1006 audit(1769255601.880:893): pid=5845 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=22 res=1 Jan 24 11:53:21.880000 audit[5845]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fffe3ec8610 a2=3 a3=0 items=0 ppid=1 pid=5845 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=22 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:21.988527 kernel: audit: type=1300 audit(1769255601.880:893): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fffe3ec8610 a2=3 a3=0 items=0 ppid=1 pid=5845 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=22 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:22.060774 kernel: audit: type=1327 audit(1769255601.880:893): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:21.880000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:22.063051 systemd[1]: Started session-22.scope - Session 22 of User core. Jan 24 11:53:22.080000 audit[5845]: USER_START pid=5845 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:22.100472 kernel: audit: type=1105 audit(1769255602.080:894): pid=5845 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:22.100634 kernel: audit: type=1103 audit(1769255602.087:895): pid=5849 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:22.087000 audit[5849]: CRED_ACQ pid=5849 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:22.632766 sshd[5849]: Connection closed by 10.0.0.1 port 50524 Jan 24 11:53:22.634745 sshd-session[5845]: pam_unix(sshd:session): session closed for user core Jan 24 11:53:22.647000 audit[5845]: USER_END pid=5845 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:22.652009 systemd[1]: sshd@20-10.0.0.67:22-10.0.0.1:50524.service: Deactivated successfully. Jan 24 11:53:22.665812 kernel: audit: type=1106 audit(1769255602.647:896): pid=5845 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:22.678638 kernel: audit: type=1104 audit(1769255602.647:897): pid=5845 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:22.647000 audit[5845]: CRED_DISP pid=5845 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:22.654000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.67:22-10.0.0.1:50524 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:22.694844 systemd[1]: session-22.scope: Deactivated successfully. Jan 24 11:53:22.706518 systemd-logind[1621]: Session 22 logged out. Waiting for processes to exit. Jan 24 11:53:22.757561 systemd-logind[1621]: Removed session 22. Jan 24 11:53:24.198047 kubelet[2895]: E0124 11:53:24.183676 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:53:24.237208 containerd[1647]: time="2026-01-24T11:53:24.233191515Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:53:24.385771 containerd[1647]: time="2026-01-24T11:53:24.385495543Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:53:24.388158 containerd[1647]: time="2026-01-24T11:53:24.387927501Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:53:24.388598 containerd[1647]: time="2026-01-24T11:53:24.388555913Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:53:24.390435 kubelet[2895]: E0124 11:53:24.389828 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:53:24.390515 kubelet[2895]: E0124 11:53:24.390484 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:53:24.406761 kubelet[2895]: E0124 11:53:24.391039 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r4s6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-f485b4455-nc2lh_calico-apiserver(d47a09e2-52ab-4e49-942e-05f2cedb67ae): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:53:24.406761 kubelet[2895]: E0124 11:53:24.392500 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:53:29.699751 containerd[1647]: time="2026-01-24T11:53:29.697638870Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 24 11:53:29.754925 systemd[1]: Started sshd@21-10.0.0.67:22-10.0.0.1:56160.service - OpenSSH per-connection server daemon (10.0.0.1:56160). Jan 24 11:53:29.757000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.67:22-10.0.0.1:56160 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:29.761469 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:53:29.761609 kernel: audit: type=1130 audit(1769255609.757:899): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.67:22-10.0.0.1:56160 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:29.824141 containerd[1647]: time="2026-01-24T11:53:29.824071222Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:53:29.829241 containerd[1647]: time="2026-01-24T11:53:29.828959343Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 24 11:53:29.829597 containerd[1647]: time="2026-01-24T11:53:29.829515081Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 24 11:53:29.848123 kubelet[2895]: E0124 11:53:29.834124 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:53:29.848123 kubelet[2895]: E0124 11:53:29.845461 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 24 11:53:29.848123 kubelet[2895]: E0124 11:53:29.845714 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k2g54,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-jfngl_calico-system(293f1d4a-795c-44fc-bda8-3c550f56df72): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 24 11:53:29.855660 kubelet[2895]: E0124 11:53:29.855239 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:53:29.897732 kubelet[2895]: E0124 11:53:29.897568 2895 kubelet.go:2627] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4.699s" Jan 24 11:53:29.967418 kubelet[2895]: E0124 11:53:29.965746 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:53:29.968465 kubelet[2895]: E0124 11:53:29.967816 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:53:29.977831 kubelet[2895]: E0124 11:53:29.977768 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:53:29.984216 kubelet[2895]: E0124 11:53:29.984146 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:53:30.198590 sshd[5865]: Accepted publickey for core from 10.0.0.1 port 56160 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:53:30.195000 audit[5865]: USER_ACCT pid=5865 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:30.213048 sshd-session[5865]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:53:30.245557 kernel: audit: type=1101 audit(1769255610.195:900): pid=5865 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:30.207000 audit[5865]: CRED_ACQ pid=5865 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:30.342509 kernel: audit: type=1103 audit(1769255610.207:901): pid=5865 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:30.419722 kernel: audit: type=1006 audit(1769255610.207:902): pid=5865 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=23 res=1 Jan 24 11:53:30.420072 kernel: audit: type=1300 audit(1769255610.207:902): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffea89b46d0 a2=3 a3=0 items=0 ppid=1 pid=5865 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=23 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:30.207000 audit[5865]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffea89b46d0 a2=3 a3=0 items=0 ppid=1 pid=5865 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=23 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:30.445208 kernel: audit: type=1327 audit(1769255610.207:902): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:30.207000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:30.431700 systemd-logind[1621]: New session 23 of user core. Jan 24 11:53:30.458635 systemd[1]: Started session-23.scope - Session 23 of User core. Jan 24 11:53:30.528000 audit[5865]: USER_START pid=5865 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:30.594108 kernel: audit: type=1105 audit(1769255610.528:903): pid=5865 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:30.594229 kernel: audit: type=1103 audit(1769255610.543:904): pid=5876 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:30.543000 audit[5876]: CRED_ACQ pid=5876 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:30.952798 sshd[5876]: Connection closed by 10.0.0.1 port 56160 Jan 24 11:53:30.954000 audit[5865]: USER_END pid=5865 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:30.953861 sshd-session[5865]: pam_unix(sshd:session): session closed for user core Jan 24 11:53:30.954000 audit[5865]: CRED_DISP pid=5865 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:30.980344 kernel: audit: type=1106 audit(1769255610.954:905): pid=5865 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:30.980464 kernel: audit: type=1104 audit(1769255610.954:906): pid=5865 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:30.990910 systemd[1]: sshd@21-10.0.0.67:22-10.0.0.1:56160.service: Deactivated successfully. Jan 24 11:53:30.990000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.67:22-10.0.0.1:56160 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:30.999154 systemd[1]: session-23.scope: Deactivated successfully. Jan 24 11:53:31.015673 systemd-logind[1621]: Session 23 logged out. Waiting for processes to exit. Jan 24 11:53:31.021925 systemd-logind[1621]: Removed session 23. Jan 24 11:53:32.130838 kubelet[2895]: E0124 11:53:32.130156 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:53:36.029221 systemd[1]: Started sshd@22-10.0.0.67:22-10.0.0.1:53648.service - OpenSSH per-connection server daemon (10.0.0.1:53648). Jan 24 11:53:36.039245 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:53:36.039437 kernel: audit: type=1130 audit(1769255616.028:908): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.67:22-10.0.0.1:53648 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:36.028000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.67:22-10.0.0.1:53648 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:36.455000 audit[5921]: USER_ACCT pid=5921 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:36.461203 sshd[5921]: Accepted publickey for core from 10.0.0.1 port 53648 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:53:36.468832 sshd-session[5921]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:53:36.490511 kernel: audit: type=1101 audit(1769255616.455:909): pid=5921 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:36.462000 audit[5921]: CRED_ACQ pid=5921 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:36.507869 systemd-logind[1621]: New session 24 of user core. Jan 24 11:53:36.553838 kernel: audit: type=1103 audit(1769255616.462:910): pid=5921 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:36.553994 kernel: audit: type=1006 audit(1769255616.462:911): pid=5921 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=24 res=1 Jan 24 11:53:36.554095 kernel: audit: type=1300 audit(1769255616.462:911): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc04152580 a2=3 a3=0 items=0 ppid=1 pid=5921 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=24 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:36.462000 audit[5921]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc04152580 a2=3 a3=0 items=0 ppid=1 pid=5921 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=24 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:36.588460 kernel: audit: type=1327 audit(1769255616.462:911): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:36.462000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:36.620984 systemd[1]: Started session-24.scope - Session 24 of User core. Jan 24 11:53:36.654000 audit[5921]: USER_START pid=5921 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:36.695361 kernel: audit: type=1105 audit(1769255616.654:912): pid=5921 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:36.695496 kernel: audit: type=1103 audit(1769255616.658:913): pid=5925 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:36.658000 audit[5925]: CRED_ACQ pid=5925 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:37.324543 sshd[5925]: Connection closed by 10.0.0.1 port 53648 Jan 24 11:53:37.323840 sshd-session[5921]: pam_unix(sshd:session): session closed for user core Jan 24 11:53:37.328000 audit[5921]: USER_END pid=5921 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:37.346609 systemd[1]: sshd@22-10.0.0.67:22-10.0.0.1:53648.service: Deactivated successfully. Jan 24 11:53:37.363381 kernel: audit: type=1106 audit(1769255617.328:914): pid=5921 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:37.363473 kernel: audit: type=1104 audit(1769255617.328:915): pid=5921 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:37.328000 audit[5921]: CRED_DISP pid=5921 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:37.356381 systemd[1]: session-24.scope: Deactivated successfully. Jan 24 11:53:37.360477 systemd-logind[1621]: Session 24 logged out. Waiting for processes to exit. Jan 24 11:53:37.363578 systemd-logind[1621]: Removed session 24. Jan 24 11:53:37.345000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.67:22-10.0.0.1:53648 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:38.146689 kubelet[2895]: E0124 11:53:38.145613 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:53:41.136965 kubelet[2895]: E0124 11:53:41.136837 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:53:41.142759 kubelet[2895]: E0124 11:53:41.137463 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:53:41.160893 kubelet[2895]: E0124 11:53:41.159540 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:53:42.136236 kubelet[2895]: E0124 11:53:42.130026 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:53:42.153583 kubelet[2895]: E0124 11:53:42.151562 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:53:42.167157 kubelet[2895]: E0124 11:53:42.167060 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:53:42.414596 systemd[1]: Started sshd@23-10.0.0.67:22-10.0.0.1:43064.service - OpenSSH per-connection server daemon (10.0.0.1:43064). Jan 24 11:53:42.450378 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:53:42.450602 kernel: audit: type=1130 audit(1769255622.429:917): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-10.0.0.67:22-10.0.0.1:43064 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:42.429000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-10.0.0.67:22-10.0.0.1:43064 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:42.781000 audit[5940]: USER_ACCT pid=5940 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:42.788566 sshd[5940]: Accepted publickey for core from 10.0.0.1 port 43064 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:53:42.798448 sshd-session[5940]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:53:42.844486 kernel: audit: type=1101 audit(1769255622.781:918): pid=5940 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:42.844605 kernel: audit: type=1103 audit(1769255622.790:919): pid=5940 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:42.790000 audit[5940]: CRED_ACQ pid=5940 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:42.851434 systemd-logind[1621]: New session 25 of user core. Jan 24 11:53:42.893518 kernel: audit: type=1006 audit(1769255622.790:920): pid=5940 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=25 res=1 Jan 24 11:53:42.790000 audit[5940]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc1115b620 a2=3 a3=0 items=0 ppid=1 pid=5940 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=25 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:42.924954 kernel: audit: type=1300 audit(1769255622.790:920): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc1115b620 a2=3 a3=0 items=0 ppid=1 pid=5940 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=25 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:42.925078 kernel: audit: type=1327 audit(1769255622.790:920): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:42.790000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:42.934775 systemd[1]: Started session-25.scope - Session 25 of User core. Jan 24 11:53:42.952000 audit[5940]: USER_START pid=5940 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:42.973587 kernel: audit: type=1105 audit(1769255622.952:921): pid=5940 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:42.976000 audit[5944]: CRED_ACQ pid=5944 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:42.997332 kernel: audit: type=1103 audit(1769255622.976:922): pid=5944 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:43.145448 kubelet[2895]: E0124 11:53:43.135078 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:53:43.424923 sshd[5944]: Connection closed by 10.0.0.1 port 43064 Jan 24 11:53:43.423567 sshd-session[5940]: pam_unix(sshd:session): session closed for user core Jan 24 11:53:43.429000 audit[5940]: USER_END pid=5940 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:43.455719 systemd[1]: sshd@23-10.0.0.67:22-10.0.0.1:43064.service: Deactivated successfully. Jan 24 11:53:43.484143 systemd[1]: session-25.scope: Deactivated successfully. Jan 24 11:53:43.487577 systemd-logind[1621]: Session 25 logged out. Waiting for processes to exit. Jan 24 11:53:43.500570 kernel: audit: type=1106 audit(1769255623.429:923): pid=5940 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:43.429000 audit[5940]: CRED_DISP pid=5940 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:43.505176 systemd-logind[1621]: Removed session 25. Jan 24 11:53:43.553655 kernel: audit: type=1104 audit(1769255623.429:924): pid=5940 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:43.451000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-10.0.0.67:22-10.0.0.1:43064 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:48.463972 systemd[1]: Started sshd@24-10.0.0.67:22-10.0.0.1:43074.service - OpenSSH per-connection server daemon (10.0.0.1:43074). Jan 24 11:53:48.479406 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:53:48.479538 kernel: audit: type=1130 audit(1769255628.463:926): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@24-10.0.0.67:22-10.0.0.1:43074 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:48.463000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@24-10.0.0.67:22-10.0.0.1:43074 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:48.799000 audit[5957]: USER_ACCT pid=5957 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:48.806794 sshd[5957]: Accepted publickey for core from 10.0.0.1 port 43074 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:53:48.831022 sshd-session[5957]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:53:48.840427 kernel: audit: type=1101 audit(1769255628.799:927): pid=5957 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:48.811000 audit[5957]: CRED_ACQ pid=5957 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:48.878827 systemd-logind[1621]: New session 26 of user core. Jan 24 11:53:48.880366 kernel: audit: type=1103 audit(1769255628.811:928): pid=5957 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:48.953243 kernel: audit: type=1006 audit(1769255628.811:929): pid=5957 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=26 res=1 Jan 24 11:53:48.953369 kernel: audit: type=1300 audit(1769255628.811:929): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff9b171f80 a2=3 a3=0 items=0 ppid=1 pid=5957 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=26 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:48.811000 audit[5957]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff9b171f80 a2=3 a3=0 items=0 ppid=1 pid=5957 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=26 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:48.811000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:48.978200 kernel: audit: type=1327 audit(1769255628.811:929): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:48.978538 systemd[1]: Started session-26.scope - Session 26 of User core. Jan 24 11:53:49.012000 audit[5957]: USER_START pid=5957 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:49.078018 kernel: audit: type=1105 audit(1769255629.012:930): pid=5957 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:49.078315 kernel: audit: type=1103 audit(1769255629.032:931): pid=5961 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:49.032000 audit[5961]: CRED_ACQ pid=5961 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:49.594084 sshd[5961]: Connection closed by 10.0.0.1 port 43074 Jan 24 11:53:49.598549 sshd-session[5957]: pam_unix(sshd:session): session closed for user core Jan 24 11:53:49.608000 audit[5957]: USER_END pid=5957 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:49.665816 kernel: audit: type=1106 audit(1769255629.608:932): pid=5957 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:49.608000 audit[5957]: CRED_DISP pid=5957 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:49.669909 systemd[1]: sshd@24-10.0.0.67:22-10.0.0.1:43074.service: Deactivated successfully. Jan 24 11:53:49.682425 kernel: audit: type=1104 audit(1769255629.608:933): pid=5957 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:49.670000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@24-10.0.0.67:22-10.0.0.1:43074 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:49.683770 systemd[1]: session-26.scope: Deactivated successfully. Jan 24 11:53:49.693007 systemd-logind[1621]: Session 26 logged out. Waiting for processes to exit. Jan 24 11:53:49.695813 systemd-logind[1621]: Removed session 26. Jan 24 11:53:52.147737 kubelet[2895]: E0124 11:53:52.146110 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:53:53.139950 kubelet[2895]: E0124 11:53:53.133857 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:53:53.172456 kubelet[2895]: E0124 11:53:53.172374 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:53:54.656000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-10.0.0.67:22-10.0.0.1:40470 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:54.658898 systemd[1]: Started sshd@25-10.0.0.67:22-10.0.0.1:40470.service - OpenSSH per-connection server daemon (10.0.0.1:40470). Jan 24 11:53:54.683932 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:53:54.684083 kernel: audit: type=1130 audit(1769255634.656:935): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-10.0.0.67:22-10.0.0.1:40470 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:54.849996 sshd[5974]: Accepted publickey for core from 10.0.0.1 port 40470 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:53:54.852745 sshd-session[5974]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:53:54.845000 audit[5974]: USER_ACCT pid=5974 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:54.893363 kernel: audit: type=1101 audit(1769255634.845:936): pid=5974 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:54.848000 audit[5974]: CRED_ACQ pid=5974 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:54.911629 systemd-logind[1621]: New session 27 of user core. Jan 24 11:53:54.937204 kernel: audit: type=1103 audit(1769255634.848:937): pid=5974 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:54.937413 kernel: audit: type=1006 audit(1769255634.848:938): pid=5974 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=27 res=1 Jan 24 11:53:54.848000 audit[5974]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc828dcd90 a2=3 a3=0 items=0 ppid=1 pid=5974 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=27 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:54.977334 kernel: audit: type=1300 audit(1769255634.848:938): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc828dcd90 a2=3 a3=0 items=0 ppid=1 pid=5974 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=27 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:53:54.977498 kernel: audit: type=1327 audit(1769255634.848:938): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:54.848000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:53:55.011067 systemd[1]: Started session-27.scope - Session 27 of User core. Jan 24 11:53:55.046000 audit[5974]: USER_START pid=5974 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:55.096359 kernel: audit: type=1105 audit(1769255635.046:939): pid=5974 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:55.108000 audit[5978]: CRED_ACQ pid=5978 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:55.146366 kernel: audit: type=1103 audit(1769255635.108:940): pid=5978 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:55.866375 sshd[5978]: Connection closed by 10.0.0.1 port 40470 Jan 24 11:53:55.862546 sshd-session[5974]: pam_unix(sshd:session): session closed for user core Jan 24 11:53:55.885000 audit[5974]: USER_END pid=5974 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:55.987608 kernel: audit: type=1106 audit(1769255635.885:941): pid=5974 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:55.987773 kernel: audit: type=1104 audit(1769255635.928:942): pid=5974 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:55.928000 audit[5974]: CRED_DISP pid=5974 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:53:55.959729 systemd[1]: sshd@25-10.0.0.67:22-10.0.0.1:40470.service: Deactivated successfully. Jan 24 11:53:55.987023 systemd[1]: session-27.scope: Deactivated successfully. Jan 24 11:53:55.949000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-10.0.0.67:22-10.0.0.1:40470 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:53:56.047258 systemd-logind[1621]: Session 27 logged out. Waiting for processes to exit. Jan 24 11:53:56.068926 systemd-logind[1621]: Removed session 27. Jan 24 11:53:56.137958 kubelet[2895]: E0124 11:53:56.135377 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:53:56.137958 kubelet[2895]: E0124 11:53:56.135899 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:53:56.152355 kubelet[2895]: E0124 11:53:56.144567 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:53:56.152355 kubelet[2895]: E0124 11:53:56.151748 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:54:00.948196 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:54:00.948499 kernel: audit: type=1130 audit(1769255640.912:944): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-10.0.0.67:22-10.0.0.1:40496 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:00.912000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-10.0.0.67:22-10.0.0.1:40496 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:00.912414 systemd[1]: Started sshd@26-10.0.0.67:22-10.0.0.1:40496.service - OpenSSH per-connection server daemon (10.0.0.1:40496). Jan 24 11:54:01.131888 kubelet[2895]: E0124 11:54:01.131721 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:54:01.144128 sshd[5994]: Accepted publickey for core from 10.0.0.1 port 40496 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:54:01.141000 audit[5994]: USER_ACCT pid=5994 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:01.149981 sshd-session[5994]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:54:01.173593 kernel: audit: type=1101 audit(1769255641.141:945): pid=5994 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:01.142000 audit[5994]: CRED_ACQ pid=5994 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:01.206832 kernel: audit: type=1103 audit(1769255641.142:946): pid=5994 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:01.206949 kernel: audit: type=1006 audit(1769255641.142:947): pid=5994 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=28 res=1 Jan 24 11:54:01.254449 kernel: audit: type=1300 audit(1769255641.142:947): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffdcd9434f0 a2=3 a3=0 items=0 ppid=1 pid=5994 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=28 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:01.254632 kernel: audit: type=1327 audit(1769255641.142:947): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:01.142000 audit[5994]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffdcd9434f0 a2=3 a3=0 items=0 ppid=1 pid=5994 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=28 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:01.142000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:01.242941 systemd-logind[1621]: New session 28 of user core. Jan 24 11:54:01.280768 systemd[1]: Started session-28.scope - Session 28 of User core. Jan 24 11:54:01.329000 audit[5994]: USER_START pid=5994 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:01.373845 kernel: audit: type=1105 audit(1769255641.329:948): pid=5994 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:01.341000 audit[5998]: CRED_ACQ pid=5998 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:01.401393 kernel: audit: type=1103 audit(1769255641.341:949): pid=5998 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:01.795926 sshd[5998]: Connection closed by 10.0.0.1 port 40496 Jan 24 11:54:01.794567 sshd-session[5994]: pam_unix(sshd:session): session closed for user core Jan 24 11:54:01.811000 audit[5994]: USER_END pid=5994 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:01.823769 systemd[1]: sshd@26-10.0.0.67:22-10.0.0.1:40496.service: Deactivated successfully. Jan 24 11:54:01.834706 systemd[1]: session-28.scope: Deactivated successfully. Jan 24 11:54:01.869213 kernel: audit: type=1106 audit(1769255641.811:950): pid=5994 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:01.869409 kernel: audit: type=1104 audit(1769255641.811:951): pid=5994 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:01.811000 audit[5994]: CRED_DISP pid=5994 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:01.850651 systemd-logind[1621]: Session 28 logged out. Waiting for processes to exit. Jan 24 11:54:01.863894 systemd-logind[1621]: Removed session 28. Jan 24 11:54:01.823000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-10.0.0.67:22-10.0.0.1:40496 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:04.147945 kubelet[2895]: E0124 11:54:04.147889 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:54:06.887434 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:54:06.887653 kernel: audit: type=1130 audit(1769255646.856:953): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@27-10.0.0.67:22-10.0.0.1:40082 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:06.856000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@27-10.0.0.67:22-10.0.0.1:40082 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:06.856912 systemd[1]: Started sshd@27-10.0.0.67:22-10.0.0.1:40082.service - OpenSSH per-connection server daemon (10.0.0.1:40082). Jan 24 11:54:07.140523 kubelet[2895]: E0124 11:54:07.137509 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:54:07.140523 kubelet[2895]: E0124 11:54:07.139245 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:54:07.359539 sshd[6037]: Accepted publickey for core from 10.0.0.1 port 40082 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:54:07.355000 audit[6037]: USER_ACCT pid=6037 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:07.366676 sshd-session[6037]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:54:07.387369 kernel: audit: type=1101 audit(1769255647.355:954): pid=6037 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:07.364000 audit[6037]: CRED_ACQ pid=6037 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:07.428336 kernel: audit: type=1103 audit(1769255647.364:955): pid=6037 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:07.428474 kernel: audit: type=1006 audit(1769255647.364:956): pid=6037 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=29 res=1 Jan 24 11:54:07.364000 audit[6037]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe615a1b90 a2=3 a3=0 items=0 ppid=1 pid=6037 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=29 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:07.433722 systemd-logind[1621]: New session 29 of user core. Jan 24 11:54:07.443835 kernel: audit: type=1300 audit(1769255647.364:956): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe615a1b90 a2=3 a3=0 items=0 ppid=1 pid=6037 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=29 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:07.364000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:07.456370 kernel: audit: type=1327 audit(1769255647.364:956): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:07.463674 systemd[1]: Started session-29.scope - Session 29 of User core. Jan 24 11:54:07.479000 audit[6037]: USER_START pid=6037 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:07.504895 kernel: audit: type=1105 audit(1769255647.479:957): pid=6037 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:07.516000 audit[6041]: CRED_ACQ pid=6041 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:07.559340 kernel: audit: type=1103 audit(1769255647.516:958): pid=6041 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:08.037564 sshd[6041]: Connection closed by 10.0.0.1 port 40082 Jan 24 11:54:08.038559 sshd-session[6037]: pam_unix(sshd:session): session closed for user core Jan 24 11:54:08.042000 audit[6037]: USER_END pid=6037 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:08.073721 systemd[1]: sshd@27-10.0.0.67:22-10.0.0.1:40082.service: Deactivated successfully. Jan 24 11:54:08.042000 audit[6037]: CRED_DISP pid=6037 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:08.086721 systemd[1]: session-29.scope: Deactivated successfully. Jan 24 11:54:08.089357 kernel: audit: type=1106 audit(1769255648.042:959): pid=6037 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:08.089445 kernel: audit: type=1104 audit(1769255648.042:960): pid=6037 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:08.089051 systemd-logind[1621]: Session 29 logged out. Waiting for processes to exit. Jan 24 11:54:08.072000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@27-10.0.0.67:22-10.0.0.1:40082 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:08.126795 systemd[1]: Started sshd@28-10.0.0.67:22-10.0.0.1:40096.service - OpenSSH per-connection server daemon (10.0.0.1:40096). Jan 24 11:54:08.125000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@28-10.0.0.67:22-10.0.0.1:40096 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:08.132508 systemd-logind[1621]: Removed session 29. Jan 24 11:54:08.324000 audit[6059]: USER_ACCT pid=6059 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:08.326669 sshd[6059]: Accepted publickey for core from 10.0.0.1 port 40096 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:54:08.332000 audit[6059]: CRED_ACQ pid=6059 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:08.332000 audit[6059]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc890a26d0 a2=3 a3=0 items=0 ppid=1 pid=6059 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=30 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:08.332000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:08.334894 sshd-session[6059]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:54:08.380223 systemd-logind[1621]: New session 30 of user core. Jan 24 11:54:08.391798 systemd[1]: Started session-30.scope - Session 30 of User core. Jan 24 11:54:08.420000 audit[6059]: USER_START pid=6059 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:08.433000 audit[6063]: CRED_ACQ pid=6063 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:10.134513 kubelet[2895]: E0124 11:54:10.133143 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:54:10.678156 sshd[6063]: Connection closed by 10.0.0.1 port 40096 Jan 24 11:54:10.680163 sshd-session[6059]: pam_unix(sshd:session): session closed for user core Jan 24 11:54:10.683000 audit[6059]: USER_END pid=6059 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:10.683000 audit[6059]: CRED_DISP pid=6059 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:10.741878 systemd[1]: sshd@28-10.0.0.67:22-10.0.0.1:40096.service: Deactivated successfully. Jan 24 11:54:10.739000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@28-10.0.0.67:22-10.0.0.1:40096 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:10.753137 systemd[1]: session-30.scope: Deactivated successfully. Jan 24 11:54:10.759391 systemd-logind[1621]: Session 30 logged out. Waiting for processes to exit. Jan 24 11:54:10.768684 systemd[1]: Started sshd@29-10.0.0.67:22-10.0.0.1:40118.service - OpenSSH per-connection server daemon (10.0.0.1:40118). Jan 24 11:54:10.767000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@29-10.0.0.67:22-10.0.0.1:40118 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:10.774632 systemd-logind[1621]: Removed session 30. Jan 24 11:54:11.140001 sshd[6075]: Accepted publickey for core from 10.0.0.1 port 40118 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:54:11.134000 audit[6075]: USER_ACCT pid=6075 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:11.140000 audit[6075]: CRED_ACQ pid=6075 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:11.142000 audit[6075]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd8f18bf90 a2=3 a3=0 items=0 ppid=1 pid=6075 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=31 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:11.142000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:11.146497 kubelet[2895]: E0124 11:54:11.144452 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:54:11.147634 sshd-session[6075]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:54:11.152078 kubelet[2895]: E0124 11:54:11.150163 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:54:11.175472 systemd-logind[1621]: New session 31 of user core. Jan 24 11:54:11.217160 systemd[1]: Started session-31.scope - Session 31 of User core. Jan 24 11:54:11.238000 audit[6075]: USER_START pid=6075 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:11.249000 audit[6079]: CRED_ACQ pid=6079 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:13.988678 kernel: kauditd_printk_skb: 20 callbacks suppressed Jan 24 11:54:13.988833 kernel: audit: type=1325 audit(1769255653.964:977): table=filter:146 family=2 entries=26 op=nft_register_rule pid=6100 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:54:13.964000 audit[6100]: NETFILTER_CFG table=filter:146 family=2 entries=26 op=nft_register_rule pid=6100 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:54:13.996125 kernel: audit: type=1300 audit(1769255653.964:977): arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7ffcb9b59520 a2=0 a3=7ffcb9b5950c items=0 ppid=3040 pid=6100 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:13.964000 audit[6100]: SYSCALL arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7ffcb9b59520 a2=0 a3=7ffcb9b5950c items=0 ppid=3040 pid=6100 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:13.964000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:54:14.080350 kernel: audit: type=1327 audit(1769255653.964:977): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:54:14.014000 audit[6100]: NETFILTER_CFG table=nat:147 family=2 entries=20 op=nft_register_rule pid=6100 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:54:14.127616 kernel: audit: type=1325 audit(1769255654.014:978): table=nat:147 family=2 entries=20 op=nft_register_rule pid=6100 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:54:14.127737 kernel: audit: type=1300 audit(1769255654.014:978): arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffcb9b59520 a2=0 a3=0 items=0 ppid=3040 pid=6100 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:14.014000 audit[6100]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffcb9b59520 a2=0 a3=0 items=0 ppid=3040 pid=6100 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:14.144385 sshd[6079]: Connection closed by 10.0.0.1 port 40118 Jan 24 11:54:14.147587 sshd-session[6075]: pam_unix(sshd:session): session closed for user core Jan 24 11:54:14.014000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:54:14.199862 kernel: audit: type=1327 audit(1769255654.014:978): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:54:14.147000 audit[6075]: USER_END pid=6075 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:14.275966 kernel: audit: type=1106 audit(1769255654.147:979): pid=6075 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:14.147000 audit[6075]: CRED_DISP pid=6075 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:14.304957 systemd[1]: sshd@29-10.0.0.67:22-10.0.0.1:40118.service: Deactivated successfully. Jan 24 11:54:14.297000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@29-10.0.0.67:22-10.0.0.1:40118 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:14.330064 systemd[1]: session-31.scope: Deactivated successfully. Jan 24 11:54:14.340667 systemd-logind[1621]: Session 31 logged out. Waiting for processes to exit. Jan 24 11:54:14.379408 kernel: audit: type=1104 audit(1769255654.147:980): pid=6075 uid=0 auid=500 ses=31 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:14.379525 kernel: audit: type=1131 audit(1769255654.297:981): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@29-10.0.0.67:22-10.0.0.1:40118 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:14.379600 kernel: audit: type=1130 audit(1769255654.364:982): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@30-10.0.0.67:22-10.0.0.1:35368 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:14.364000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@30-10.0.0.67:22-10.0.0.1:35368 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:14.367724 systemd[1]: Started sshd@30-10.0.0.67:22-10.0.0.1:35368.service - OpenSSH per-connection server daemon (10.0.0.1:35368). Jan 24 11:54:14.371941 systemd-logind[1621]: Removed session 31. Jan 24 11:54:14.467000 audit[6107]: NETFILTER_CFG table=filter:148 family=2 entries=38 op=nft_register_rule pid=6107 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:54:14.467000 audit[6107]: SYSCALL arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7fff53aa88b0 a2=0 a3=7fff53aa889c items=0 ppid=3040 pid=6107 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:14.467000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:54:14.483000 audit[6107]: NETFILTER_CFG table=nat:149 family=2 entries=20 op=nft_register_rule pid=6107 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:54:14.483000 audit[6107]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7fff53aa88b0 a2=0 a3=0 items=0 ppid=3040 pid=6107 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:14.483000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:54:14.595000 audit[6106]: USER_ACCT pid=6106 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:14.598382 sshd[6106]: Accepted publickey for core from 10.0.0.1 port 35368 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:54:14.618000 audit[6106]: CRED_ACQ pid=6106 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:14.618000 audit[6106]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc3b4b12c0 a2=3 a3=0 items=0 ppid=1 pid=6106 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=32 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:14.618000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:14.626728 sshd-session[6106]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:54:14.705938 systemd-logind[1621]: New session 32 of user core. Jan 24 11:54:14.737773 systemd[1]: Started session-32.scope - Session 32 of User core. Jan 24 11:54:14.752000 audit[6106]: USER_START pid=6106 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:14.757000 audit[6111]: CRED_ACQ pid=6111 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:16.129618 sshd[6111]: Connection closed by 10.0.0.1 port 35368 Jan 24 11:54:16.130553 sshd-session[6106]: pam_unix(sshd:session): session closed for user core Jan 24 11:54:16.138000 audit[6106]: USER_END pid=6106 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:16.138000 audit[6106]: CRED_DISP pid=6106 uid=0 auid=500 ses=32 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:16.164174 systemd[1]: sshd@30-10.0.0.67:22-10.0.0.1:35368.service: Deactivated successfully. Jan 24 11:54:16.169000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@30-10.0.0.67:22-10.0.0.1:35368 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:16.180778 systemd[1]: session-32.scope: Deactivated successfully. Jan 24 11:54:16.200527 systemd-logind[1621]: Session 32 logged out. Waiting for processes to exit. Jan 24 11:54:16.207000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@31-10.0.0.67:22-10.0.0.1:35374 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:16.208894 systemd[1]: Started sshd@31-10.0.0.67:22-10.0.0.1:35374.service - OpenSSH per-connection server daemon (10.0.0.1:35374). Jan 24 11:54:16.231127 systemd-logind[1621]: Removed session 32. Jan 24 11:54:16.537000 audit[6122]: USER_ACCT pid=6122 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:16.540230 sshd[6122]: Accepted publickey for core from 10.0.0.1 port 35374 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:54:16.540000 audit[6122]: CRED_ACQ pid=6122 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:16.540000 audit[6122]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe258e87c0 a2=3 a3=0 items=0 ppid=1 pid=6122 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=33 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:16.540000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:16.543773 sshd-session[6122]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:54:16.573401 systemd-logind[1621]: New session 33 of user core. Jan 24 11:54:16.584076 systemd[1]: Started session-33.scope - Session 33 of User core. Jan 24 11:54:16.596000 audit[6122]: USER_START pid=6122 uid=0 auid=500 ses=33 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:16.601000 audit[6126]: CRED_ACQ pid=6126 uid=0 auid=500 ses=33 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:16.863612 sshd[6126]: Connection closed by 10.0.0.1 port 35374 Jan 24 11:54:16.866233 sshd-session[6122]: pam_unix(sshd:session): session closed for user core Jan 24 11:54:16.866000 audit[6122]: USER_END pid=6122 uid=0 auid=500 ses=33 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:16.867000 audit[6122]: CRED_DISP pid=6122 uid=0 auid=500 ses=33 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:16.877000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@31-10.0.0.67:22-10.0.0.1:35374 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:16.877157 systemd[1]: sshd@31-10.0.0.67:22-10.0.0.1:35374.service: Deactivated successfully. Jan 24 11:54:16.882083 systemd[1]: session-33.scope: Deactivated successfully. Jan 24 11:54:16.882960 systemd-logind[1621]: Session 33 logged out. Waiting for processes to exit. Jan 24 11:54:16.896983 systemd-logind[1621]: Removed session 33. Jan 24 11:54:18.131422 kubelet[2895]: E0124 11:54:18.130073 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:54:18.132023 kubelet[2895]: E0124 11:54:18.131757 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:54:19.153151 kubelet[2895]: E0124 11:54:19.153094 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:54:21.140598 containerd[1647]: time="2026-01-24T11:54:21.140159980Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 24 11:54:21.262023 containerd[1647]: time="2026-01-24T11:54:21.261193524Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:54:21.273735 containerd[1647]: time="2026-01-24T11:54:21.273618980Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 24 11:54:21.274147 containerd[1647]: time="2026-01-24T11:54:21.273676063Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 24 11:54:21.274504 kubelet[2895]: E0124 11:54:21.274175 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:54:21.280365 kubelet[2895]: E0124 11:54:21.274427 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 24 11:54:21.282467 kubelet[2895]: E0124 11:54:21.282115 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p5kgf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-66955b7fc7-xmscg_calico-system(41fe2d74-dafe-45a2-82d2-5fd56447e59d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 24 11:54:21.283696 kubelet[2895]: E0124 11:54:21.283614 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:54:21.916000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@32-10.0.0.67:22-10.0.0.1:35384 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:21.917932 systemd[1]: Started sshd@32-10.0.0.67:22-10.0.0.1:35384.service - OpenSSH per-connection server daemon (10.0.0.1:35384). Jan 24 11:54:21.928763 kernel: kauditd_printk_skb: 27 callbacks suppressed Jan 24 11:54:21.928922 kernel: audit: type=1130 audit(1769255661.916:1002): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@32-10.0.0.67:22-10.0.0.1:35384 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:22.054000 audit[6147]: USER_ACCT pid=6147 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:22.060432 sshd[6147]: Accepted publickey for core from 10.0.0.1 port 35384 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:54:22.068926 kernel: audit: type=1101 audit(1769255662.054:1003): pid=6147 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:22.068000 audit[6147]: CRED_ACQ pid=6147 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:22.071858 sshd-session[6147]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:54:22.082512 kernel: audit: type=1103 audit(1769255662.068:1004): pid=6147 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:22.082570 kernel: audit: type=1006 audit(1769255662.068:1005): pid=6147 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=34 res=1 Jan 24 11:54:22.068000 audit[6147]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe5886b3c0 a2=3 a3=0 items=0 ppid=1 pid=6147 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=34 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:22.109450 kernel: audit: type=1300 audit(1769255662.068:1005): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe5886b3c0 a2=3 a3=0 items=0 ppid=1 pid=6147 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=34 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:22.109578 kernel: audit: type=1327 audit(1769255662.068:1005): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:22.068000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:22.125412 systemd-logind[1621]: New session 34 of user core. Jan 24 11:54:22.134426 kubelet[2895]: E0124 11:54:22.134249 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:54:22.140658 systemd[1]: Started session-34.scope - Session 34 of User core. Jan 24 11:54:22.155000 audit[6147]: USER_START pid=6147 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:22.200327 kernel: audit: type=1105 audit(1769255662.155:1006): pid=6147 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:22.200484 kernel: audit: type=1103 audit(1769255662.180:1007): pid=6151 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:22.180000 audit[6151]: CRED_ACQ pid=6151 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:22.588445 sshd[6151]: Connection closed by 10.0.0.1 port 35384 Jan 24 11:54:22.589531 sshd-session[6147]: pam_unix(sshd:session): session closed for user core Jan 24 11:54:22.599000 audit[6147]: USER_END pid=6147 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:22.658446 kernel: audit: type=1106 audit(1769255662.599:1008): pid=6147 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:22.634970 systemd[1]: sshd@32-10.0.0.67:22-10.0.0.1:35384.service: Deactivated successfully. Jan 24 11:54:22.636004 systemd-logind[1621]: Session 34 logged out. Waiting for processes to exit. Jan 24 11:54:22.647774 systemd[1]: session-34.scope: Deactivated successfully. Jan 24 11:54:22.600000 audit[6147]: CRED_DISP pid=6147 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:22.674006 systemd-logind[1621]: Removed session 34. Jan 24 11:54:22.679396 kernel: audit: type=1104 audit(1769255662.600:1009): pid=6147 uid=0 auid=500 ses=34 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:22.636000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@32-10.0.0.67:22-10.0.0.1:35384 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:24.135572 kubelet[2895]: E0124 11:54:24.135506 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:54:25.141899 kubelet[2895]: E0124 11:54:25.139991 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:54:27.635380 systemd[1]: Started sshd@33-10.0.0.67:22-10.0.0.1:40124.service - OpenSSH per-connection server daemon (10.0.0.1:40124). Jan 24 11:54:27.634000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@33-10.0.0.67:22-10.0.0.1:40124 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:27.664512 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:54:27.664645 kernel: audit: type=1130 audit(1769255667.634:1011): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@33-10.0.0.67:22-10.0.0.1:40124 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:27.857000 audit[6169]: USER_ACCT pid=6169 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:27.858067 sshd[6169]: Accepted publickey for core from 10.0.0.1 port 40124 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:54:27.865588 sshd-session[6169]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:54:27.890397 kernel: audit: type=1101 audit(1769255667.857:1012): pid=6169 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:27.890563 kernel: audit: type=1103 audit(1769255667.862:1013): pid=6169 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:27.862000 audit[6169]: CRED_ACQ pid=6169 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:27.893075 systemd-logind[1621]: New session 35 of user core. Jan 24 11:54:27.932464 kernel: audit: type=1006 audit(1769255667.862:1014): pid=6169 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=35 res=1 Jan 24 11:54:27.932629 kernel: audit: type=1300 audit(1769255667.862:1014): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd04bfb830 a2=3 a3=0 items=0 ppid=1 pid=6169 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=35 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:27.862000 audit[6169]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd04bfb830 a2=3 a3=0 items=0 ppid=1 pid=6169 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=35 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:27.862000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:27.951340 kernel: audit: type=1327 audit(1769255667.862:1014): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:27.952188 systemd[1]: Started session-35.scope - Session 35 of User core. Jan 24 11:54:27.966000 audit[6169]: USER_START pid=6169 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:27.994488 kernel: audit: type=1105 audit(1769255667.966:1015): pid=6169 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:27.994625 kernel: audit: type=1103 audit(1769255667.975:1016): pid=6173 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:27.975000 audit[6173]: CRED_ACQ pid=6173 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:28.262633 sshd[6173]: Connection closed by 10.0.0.1 port 40124 Jan 24 11:54:28.265503 sshd-session[6169]: pam_unix(sshd:session): session closed for user core Jan 24 11:54:28.319471 kernel: audit: type=1106 audit(1769255668.283:1017): pid=6169 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:28.283000 audit[6169]: USER_END pid=6169 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:28.294184 systemd[1]: sshd@33-10.0.0.67:22-10.0.0.1:40124.service: Deactivated successfully. Jan 24 11:54:28.310128 systemd[1]: session-35.scope: Deactivated successfully. Jan 24 11:54:28.318808 systemd-logind[1621]: Session 35 logged out. Waiting for processes to exit. Jan 24 11:54:28.283000 audit[6169]: CRED_DISP pid=6169 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:28.321091 systemd-logind[1621]: Removed session 35. Jan 24 11:54:28.353749 kernel: audit: type=1104 audit(1769255668.283:1018): pid=6169 uid=0 auid=500 ses=35 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:28.293000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@33-10.0.0.67:22-10.0.0.1:40124 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:30.143828 kubelet[2895]: E0124 11:54:30.143588 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:54:32.129841 kubelet[2895]: E0124 11:54:32.129791 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:54:33.147802 kubelet[2895]: E0124 11:54:33.147110 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:54:33.149467 kubelet[2895]: E0124 11:54:33.149355 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:54:33.150052 kubelet[2895]: E0124 11:54:33.149947 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:54:33.319000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@34-10.0.0.67:22-10.0.0.1:37402 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:33.319934 systemd[1]: Started sshd@34-10.0.0.67:22-10.0.0.1:37402.service - OpenSSH per-connection server daemon (10.0.0.1:37402). Jan 24 11:54:33.333085 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:54:33.333210 kernel: audit: type=1130 audit(1769255673.319:1020): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@34-10.0.0.67:22-10.0.0.1:37402 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:33.493000 audit[6186]: USER_ACCT pid=6186 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:33.507720 sshd[6186]: Accepted publickey for core from 10.0.0.1 port 37402 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:54:33.509072 sshd-session[6186]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:54:33.498000 audit[6186]: CRED_ACQ pid=6186 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:33.551358 kernel: audit: type=1101 audit(1769255673.493:1021): pid=6186 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:33.551478 kernel: audit: type=1103 audit(1769255673.498:1022): pid=6186 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:33.566210 systemd-logind[1621]: New session 36 of user core. Jan 24 11:54:33.579901 kernel: audit: type=1006 audit(1769255673.502:1023): pid=6186 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=36 res=1 Jan 24 11:54:33.580067 kernel: audit: type=1300 audit(1769255673.502:1023): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffdde86a440 a2=3 a3=0 items=0 ppid=1 pid=6186 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=36 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:33.502000 audit[6186]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffdde86a440 a2=3 a3=0 items=0 ppid=1 pid=6186 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=36 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:33.624399 kernel: audit: type=1327 audit(1769255673.502:1023): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:33.502000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:33.623869 systemd[1]: Started session-36.scope - Session 36 of User core. Jan 24 11:54:33.634000 audit[6186]: USER_START pid=6186 uid=0 auid=500 ses=36 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:33.669420 kernel: audit: type=1105 audit(1769255673.634:1024): pid=6186 uid=0 auid=500 ses=36 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:33.643000 audit[6190]: CRED_ACQ pid=6190 uid=0 auid=500 ses=36 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:33.684211 kernel: audit: type=1103 audit(1769255673.643:1025): pid=6190 uid=0 auid=500 ses=36 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:33.959351 sshd[6190]: Connection closed by 10.0.0.1 port 37402 Jan 24 11:54:33.960605 sshd-session[6186]: pam_unix(sshd:session): session closed for user core Jan 24 11:54:33.962000 audit[6186]: USER_END pid=6186 uid=0 auid=500 ses=36 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:33.977646 systemd[1]: sshd@34-10.0.0.67:22-10.0.0.1:37402.service: Deactivated successfully. Jan 24 11:54:33.985850 systemd[1]: session-36.scope: Deactivated successfully. Jan 24 11:54:33.962000 audit[6186]: CRED_DISP pid=6186 uid=0 auid=500 ses=36 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:33.988822 systemd-logind[1621]: Session 36 logged out. Waiting for processes to exit. Jan 24 11:54:34.007188 systemd-logind[1621]: Removed session 36. Jan 24 11:54:34.012806 kernel: audit: type=1106 audit(1769255673.962:1026): pid=6186 uid=0 auid=500 ses=36 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:34.013086 kernel: audit: type=1104 audit(1769255673.962:1027): pid=6186 uid=0 auid=500 ses=36 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:33.972000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@34-10.0.0.67:22-10.0.0.1:37402 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:34.735000 audit[6231]: NETFILTER_CFG table=filter:150 family=2 entries=26 op=nft_register_rule pid=6231 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:54:34.735000 audit[6231]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffc54c0fbd0 a2=0 a3=7ffc54c0fbbc items=0 ppid=3040 pid=6231 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:34.735000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:54:34.743000 audit[6231]: NETFILTER_CFG table=nat:151 family=2 entries=104 op=nft_register_chain pid=6231 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 24 11:54:34.743000 audit[6231]: SYSCALL arch=c000003e syscall=46 success=yes exit=48684 a0=3 a1=7ffc54c0fbd0 a2=0 a3=7ffc54c0fbbc items=0 ppid=3040 pid=6231 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:34.743000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 24 11:54:35.073668 systemd[1743]: Created slice background.slice - User Background Tasks Slice. Jan 24 11:54:35.087307 systemd[1743]: Starting systemd-tmpfiles-clean.service - Cleanup of User's Temporary Files and Directories... Jan 24 11:54:35.187655 systemd[1743]: Finished systemd-tmpfiles-clean.service - Cleanup of User's Temporary Files and Directories. Jan 24 11:54:38.134766 containerd[1647]: time="2026-01-24T11:54:38.134651708Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:54:38.242989 containerd[1647]: time="2026-01-24T11:54:38.242815530Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:54:38.253067 containerd[1647]: time="2026-01-24T11:54:38.252850894Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:54:38.253067 containerd[1647]: time="2026-01-24T11:54:38.253012365Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:54:38.253590 kubelet[2895]: E0124 11:54:38.253417 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:54:38.254394 kubelet[2895]: E0124 11:54:38.253635 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:54:38.255476 kubelet[2895]: E0124 11:54:38.254903 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l682s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-f485b4455-qhcml_calico-apiserver(c28e9f44-9bc9-4328-bf7a-f4f9d016a87e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:54:38.255656 containerd[1647]: time="2026-01-24T11:54:38.255092196Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 24 11:54:38.257532 kubelet[2895]: E0124 11:54:38.257450 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:54:38.335804 containerd[1647]: time="2026-01-24T11:54:38.335735152Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:54:38.350198 containerd[1647]: time="2026-01-24T11:54:38.350049631Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 24 11:54:38.350198 containerd[1647]: time="2026-01-24T11:54:38.350182469Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 24 11:54:38.351155 kubelet[2895]: E0124 11:54:38.351013 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:54:38.351155 kubelet[2895]: E0124 11:54:38.351089 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 24 11:54:38.357998 kubelet[2895]: E0124 11:54:38.351347 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lqn89,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2g45n_calico-system(999c48c5-f3b1-4bfd-91f4-26cafee1eef0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 24 11:54:38.358202 containerd[1647]: time="2026-01-24T11:54:38.356068649Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 24 11:54:38.437510 containerd[1647]: time="2026-01-24T11:54:38.436054270Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:54:38.444129 containerd[1647]: time="2026-01-24T11:54:38.443079304Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 24 11:54:38.444129 containerd[1647]: time="2026-01-24T11:54:38.443357824Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 24 11:54:38.444381 kubelet[2895]: E0124 11:54:38.443629 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:54:38.444381 kubelet[2895]: E0124 11:54:38.443703 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 24 11:54:38.444381 kubelet[2895]: E0124 11:54:38.443869 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lqn89,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-2g45n_calico-system(999c48c5-f3b1-4bfd-91f4-26cafee1eef0): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 24 11:54:38.445756 kubelet[2895]: E0124 11:54:38.445629 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0" Jan 24 11:54:39.046011 kernel: kauditd_printk_skb: 7 callbacks suppressed Jan 24 11:54:39.046815 kernel: audit: type=1130 audit(1769255679.025:1031): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@35-10.0.0.67:22-10.0.0.1:37430 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:39.025000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@35-10.0.0.67:22-10.0.0.1:37430 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:39.025882 systemd[1]: Started sshd@35-10.0.0.67:22-10.0.0.1:37430.service - OpenSSH per-connection server daemon (10.0.0.1:37430). Jan 24 11:54:39.281000 audit[6235]: USER_ACCT pid=6235 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:39.295372 kernel: audit: type=1101 audit(1769255679.281:1032): pid=6235 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:39.295455 sshd[6235]: Accepted publickey for core from 10.0.0.1 port 37430 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:54:39.305056 sshd-session[6235]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:54:39.299000 audit[6235]: CRED_ACQ pid=6235 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:39.336866 kernel: audit: type=1103 audit(1769255679.299:1033): pid=6235 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:39.336976 kernel: audit: type=1006 audit(1769255679.299:1034): pid=6235 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=37 res=1 Jan 24 11:54:39.336592 systemd-logind[1621]: New session 37 of user core. Jan 24 11:54:39.299000 audit[6235]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff62ad6970 a2=3 a3=0 items=0 ppid=1 pid=6235 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=37 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:39.299000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:39.353846 kernel: audit: type=1300 audit(1769255679.299:1034): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff62ad6970 a2=3 a3=0 items=0 ppid=1 pid=6235 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=37 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:39.353907 kernel: audit: type=1327 audit(1769255679.299:1034): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:39.359225 systemd[1]: Started session-37.scope - Session 37 of User core. Jan 24 11:54:39.367000 audit[6235]: USER_START pid=6235 uid=0 auid=500 ses=37 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:39.377000 audit[6239]: CRED_ACQ pid=6239 uid=0 auid=500 ses=37 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:39.395670 kernel: audit: type=1105 audit(1769255679.367:1035): pid=6235 uid=0 auid=500 ses=37 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:39.395764 kernel: audit: type=1103 audit(1769255679.377:1036): pid=6239 uid=0 auid=500 ses=37 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:39.625000 audit[6235]: USER_END pid=6235 uid=0 auid=500 ses=37 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:39.626614 sshd[6239]: Connection closed by 10.0.0.1 port 37430 Jan 24 11:54:39.623972 sshd-session[6235]: pam_unix(sshd:session): session closed for user core Jan 24 11:54:39.634972 systemd[1]: sshd@35-10.0.0.67:22-10.0.0.1:37430.service: Deactivated successfully. Jan 24 11:54:39.643142 systemd[1]: session-37.scope: Deactivated successfully. Jan 24 11:54:39.626000 audit[6235]: CRED_DISP pid=6235 uid=0 auid=500 ses=37 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:39.648511 systemd-logind[1621]: Session 37 logged out. Waiting for processes to exit. Jan 24 11:54:39.656354 systemd-logind[1621]: Removed session 37. Jan 24 11:54:39.662949 kernel: audit: type=1106 audit(1769255679.625:1037): pid=6235 uid=0 auid=500 ses=37 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:39.663089 kernel: audit: type=1104 audit(1769255679.626:1038): pid=6235 uid=0 auid=500 ses=37 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:39.637000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@35-10.0.0.67:22-10.0.0.1:37430 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:44.486683 kubelet[2895]: E0124 11:54:44.486493 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-jfngl" podUID="293f1d4a-795c-44fc-bda8-3c550f56df72" Jan 24 11:54:44.497401 containerd[1647]: time="2026-01-24T11:54:44.494584569Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 24 11:54:44.684000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@36-10.0.0.67:22-10.0.0.1:36150 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:44.688799 containerd[1647]: time="2026-01-24T11:54:44.685882703Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:54:44.686186 systemd[1]: Started sshd@36-10.0.0.67:22-10.0.0.1:36150.service - OpenSSH per-connection server daemon (10.0.0.1:36150). Jan 24 11:54:44.697482 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:54:44.697951 kernel: audit: type=1130 audit(1769255684.684:1040): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@36-10.0.0.67:22-10.0.0.1:36150 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:44.699834 containerd[1647]: time="2026-01-24T11:54:44.699776478Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 24 11:54:44.700199 containerd[1647]: time="2026-01-24T11:54:44.700072611Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 24 11:54:44.700785 kubelet[2895]: E0124 11:54:44.700682 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:54:44.700785 kubelet[2895]: E0124 11:54:44.700754 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 24 11:54:44.701814 kubelet[2895]: E0124 11:54:44.701743 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r4s6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-f485b4455-nc2lh_calico-apiserver(d47a09e2-52ab-4e49-942e-05f2cedb67ae): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 24 11:54:44.707655 kubelet[2895]: E0124 11:54:44.707602 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-nc2lh" podUID="d47a09e2-52ab-4e49-942e-05f2cedb67ae" Jan 24 11:54:45.035000 audit[6252]: USER_ACCT pid=6252 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:45.040345 sshd[6252]: Accepted publickey for core from 10.0.0.1 port 36150 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:54:45.047198 sshd-session[6252]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:54:45.060388 kernel: audit: type=1101 audit(1769255685.035:1041): pid=6252 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:45.045000 audit[6252]: CRED_ACQ pid=6252 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:45.089352 systemd-logind[1621]: New session 38 of user core. Jan 24 11:54:45.122398 kernel: audit: type=1103 audit(1769255685.045:1042): pid=6252 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:45.122540 kernel: audit: type=1006 audit(1769255685.045:1043): pid=6252 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=38 res=1 Jan 24 11:54:45.045000 audit[6252]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffff4aff6d0 a2=3 a3=0 items=0 ppid=1 pid=6252 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=38 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:45.131641 systemd[1]: Started session-38.scope - Session 38 of User core. Jan 24 11:54:45.162998 kernel: audit: type=1300 audit(1769255685.045:1043): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffff4aff6d0 a2=3 a3=0 items=0 ppid=1 pid=6252 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=38 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:45.163130 kernel: audit: type=1327 audit(1769255685.045:1043): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:45.045000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:45.222933 kernel: audit: type=1105 audit(1769255685.159:1044): pid=6252 uid=0 auid=500 ses=38 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:45.159000 audit[6252]: USER_START pid=6252 uid=0 auid=500 ses=38 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:45.187000 audit[6268]: CRED_ACQ pid=6268 uid=0 auid=500 ses=38 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:45.264253 containerd[1647]: time="2026-01-24T11:54:45.232619723Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 24 11:54:45.264477 kernel: audit: type=1103 audit(1769255685.187:1045): pid=6268 uid=0 auid=500 ses=38 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:45.342538 containerd[1647]: time="2026-01-24T11:54:45.340430011Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:54:45.347748 containerd[1647]: time="2026-01-24T11:54:45.347692262Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 24 11:54:45.353059 containerd[1647]: time="2026-01-24T11:54:45.347976673Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 24 11:54:45.353191 kubelet[2895]: E0124 11:54:45.352429 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:54:45.353191 kubelet[2895]: E0124 11:54:45.352502 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 24 11:54:45.353191 kubelet[2895]: E0124 11:54:45.352666 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:d91f0bbcac534ce79a38827d5dd29c41,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5n6cl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c99cdc8ff-zh5gl_calico-system(5cdab4ea-dcce-4467-b450-26abe2130c8d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 24 11:54:45.360483 containerd[1647]: time="2026-01-24T11:54:45.360440860Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 24 11:54:45.481649 containerd[1647]: time="2026-01-24T11:54:45.481594395Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 24 11:54:45.492791 containerd[1647]: time="2026-01-24T11:54:45.492739160Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 24 11:54:45.493132 containerd[1647]: time="2026-01-24T11:54:45.493020335Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 24 11:54:45.539089 kubelet[2895]: E0124 11:54:45.538550 2895 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:54:45.539089 kubelet[2895]: E0124 11:54:45.538615 2895 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 24 11:54:45.539089 kubelet[2895]: E0124 11:54:45.538755 2895 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5n6cl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-5c99cdc8ff-zh5gl_calico-system(5cdab4ea-dcce-4467-b450-26abe2130c8d): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 24 11:54:45.549569 kubelet[2895]: E0124 11:54:45.540480 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-5c99cdc8ff-zh5gl" podUID="5cdab4ea-dcce-4467-b450-26abe2130c8d" Jan 24 11:54:45.901439 sshd[6268]: Connection closed by 10.0.0.1 port 36150 Jan 24 11:54:45.897607 sshd-session[6252]: pam_unix(sshd:session): session closed for user core Jan 24 11:54:45.904000 audit[6252]: USER_END pid=6252 uid=0 auid=500 ses=38 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:45.958107 systemd[1]: sshd@36-10.0.0.67:22-10.0.0.1:36150.service: Deactivated successfully. Jan 24 11:54:45.960086 systemd-logind[1621]: Session 38 logged out. Waiting for processes to exit. Jan 24 11:54:45.993698 systemd[1]: session-38.scope: Deactivated successfully. Jan 24 11:54:46.006048 systemd-logind[1621]: Removed session 38. Jan 24 11:54:46.044539 kernel: audit: type=1106 audit(1769255685.904:1046): pid=6252 uid=0 auid=500 ses=38 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:46.044668 kernel: audit: type=1104 audit(1769255685.904:1047): pid=6252 uid=0 auid=500 ses=38 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:45.904000 audit[6252]: CRED_DISP pid=6252 uid=0 auid=500 ses=38 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:45.959000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@36-10.0.0.67:22-10.0.0.1:36150 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:47.160752 kubelet[2895]: E0124 11:54:47.141205 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-66955b7fc7-xmscg" podUID="41fe2d74-dafe-45a2-82d2-5fd56447e59d" Jan 24 11:54:50.915119 systemd[1]: Started sshd@37-10.0.0.67:22-10.0.0.1:36204.service - OpenSSH per-connection server daemon (10.0.0.1:36204). Jan 24 11:54:50.925656 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 24 11:54:50.926430 kernel: audit: type=1130 audit(1769255690.915:1049): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@37-10.0.0.67:22-10.0.0.1:36204 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:50.915000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@37-10.0.0.67:22-10.0.0.1:36204 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:51.047000 audit[6289]: USER_ACCT pid=6289 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:51.052413 sshd-session[6289]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 24 11:54:51.054188 sshd[6289]: Accepted publickey for core from 10.0.0.1 port 36204 ssh2: RSA SHA256:N4DptLu65muvg2RdNP5t6A9jwGknXmCATYE4jszWH64 Jan 24 11:54:51.063940 kernel: audit: type=1101 audit(1769255691.047:1050): pid=6289 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:51.064030 kernel: audit: type=1103 audit(1769255691.050:1051): pid=6289 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:51.050000 audit[6289]: CRED_ACQ pid=6289 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:51.065738 systemd-logind[1621]: New session 39 of user core. Jan 24 11:54:51.086167 kernel: audit: type=1006 audit(1769255691.050:1052): pid=6289 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=39 res=1 Jan 24 11:54:51.050000 audit[6289]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffffa01a5a0 a2=3 a3=0 items=0 ppid=1 pid=6289 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=39 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:51.094419 systemd[1]: Started session-39.scope - Session 39 of User core. Jan 24 11:54:51.125503 kernel: audit: type=1300 audit(1769255691.050:1052): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffffa01a5a0 a2=3 a3=0 items=0 ppid=1 pid=6289 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=39 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 24 11:54:51.125635 kernel: audit: type=1327 audit(1769255691.050:1052): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:51.050000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 24 11:54:51.122000 audit[6289]: USER_START pid=6289 uid=0 auid=500 ses=39 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:51.151766 kernel: audit: type=1105 audit(1769255691.122:1053): pid=6289 uid=0 auid=500 ses=39 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:51.128000 audit[6293]: CRED_ACQ pid=6293 uid=0 auid=500 ses=39 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:51.167428 kernel: audit: type=1103 audit(1769255691.128:1054): pid=6293 uid=0 auid=500 ses=39 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:51.171034 kubelet[2895]: E0124 11:54:51.164183 2895 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jan 24 11:54:51.365362 sshd[6293]: Connection closed by 10.0.0.1 port 36204 Jan 24 11:54:51.365045 sshd-session[6289]: pam_unix(sshd:session): session closed for user core Jan 24 11:54:51.392349 kernel: audit: type=1106 audit(1769255691.376:1055): pid=6289 uid=0 auid=500 ses=39 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:51.376000 audit[6289]: USER_END pid=6289 uid=0 auid=500 ses=39 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:51.376000 audit[6289]: CRED_DISP pid=6289 uid=0 auid=500 ses=39 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:51.402942 systemd[1]: sshd@37-10.0.0.67:22-10.0.0.1:36204.service: Deactivated successfully. Jan 24 11:54:51.415161 systemd[1]: session-39.scope: Deactivated successfully. Jan 24 11:54:51.429865 systemd-logind[1621]: Session 39 logged out. Waiting for processes to exit. Jan 24 11:54:51.432118 systemd-logind[1621]: Removed session 39. Jan 24 11:54:51.401000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@37-10.0.0.67:22-10.0.0.1:36204 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 24 11:54:51.433935 kernel: audit: type=1104 audit(1769255691.376:1056): pid=6289 uid=0 auid=500 ses=39 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' Jan 24 11:54:52.132350 kubelet[2895]: E0124 11:54:52.132200 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-f485b4455-qhcml" podUID="c28e9f44-9bc9-4328-bf7a-f4f9d016a87e" Jan 24 11:54:53.175256 kubelet[2895]: E0124 11:54:53.174405 2895 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-2g45n" podUID="999c48c5-f3b1-4bfd-91f4-26cafee1eef0"