Dec 16 12:47:04.539899 kernel: Linux version 6.12.61-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.1_p20250801 p4) 14.3.1 20250801, GNU ld (Gentoo 2.45 p3) 2.45.0) #1 SMP PREEMPT_DYNAMIC Tue Dec 16 00:18:19 -00 2025 Dec 16 12:47:04.539921 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=553464fdb0286a5b06b399da29ca659e521c68f08ea70a931c96ddffd00b4357 Dec 16 12:47:04.539931 kernel: BIOS-provided physical RAM map: Dec 16 12:47:04.539937 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Dec 16 12:47:04.539942 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Dec 16 12:47:04.539947 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Dec 16 12:47:04.539953 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007cfdbfff] usable Dec 16 12:47:04.539958 kernel: BIOS-e820: [mem 0x000000007cfdc000-0x000000007cffffff] reserved Dec 16 12:47:04.539964 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Dec 16 12:47:04.539970 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Dec 16 12:47:04.539976 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Dec 16 12:47:04.539981 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Dec 16 12:47:04.539986 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Dec 16 12:47:04.540321 kernel: NX (Execute Disable) protection: active Dec 16 12:47:04.540345 kernel: APIC: Static calls initialized Dec 16 12:47:04.540358 kernel: SMBIOS 3.0.0 present. Dec 16 12:47:04.540370 kernel: DMI: Hetzner vServer/Standard PC (Q35 + ICH9, 2009), BIOS 20171111 11/11/2017 Dec 16 12:47:04.540381 kernel: DMI: Memory slots populated: 1/1 Dec 16 12:47:04.540391 kernel: Hypervisor detected: KVM Dec 16 12:47:04.540402 kernel: last_pfn = 0x7cfdc max_arch_pfn = 0x400000000 Dec 16 12:47:04.540412 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Dec 16 12:47:04.540422 kernel: kvm-clock: using sched offset of 3985632751 cycles Dec 16 12:47:04.540433 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Dec 16 12:47:04.540445 kernel: tsc: Detected 2445.404 MHz processor Dec 16 12:47:04.540459 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Dec 16 12:47:04.540471 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Dec 16 12:47:04.540483 kernel: last_pfn = 0x7cfdc max_arch_pfn = 0x400000000 Dec 16 12:47:04.540495 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Dec 16 12:47:04.540506 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Dec 16 12:47:04.540518 kernel: Using GB pages for direct mapping Dec 16 12:47:04.540529 kernel: ACPI: Early table checksum verification disabled Dec 16 12:47:04.540554 kernel: ACPI: RSDP 0x00000000000F5270 000014 (v00 BOCHS ) Dec 16 12:47:04.540560 kernel: ACPI: RSDT 0x000000007CFE2693 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 16 12:47:04.540567 kernel: ACPI: FACP 0x000000007CFE2483 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Dec 16 12:47:04.540573 kernel: ACPI: DSDT 0x000000007CFE0040 002443 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 16 12:47:04.540579 kernel: ACPI: FACS 0x000000007CFE0000 000040 Dec 16 12:47:04.540586 kernel: ACPI: APIC 0x000000007CFE2577 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Dec 16 12:47:04.540592 kernel: ACPI: HPET 0x000000007CFE25F7 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 16 12:47:04.540599 kernel: ACPI: MCFG 0x000000007CFE262F 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 16 12:47:04.540606 kernel: ACPI: WAET 0x000000007CFE266B 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 16 12:47:04.540614 kernel: ACPI: Reserving FACP table memory at [mem 0x7cfe2483-0x7cfe2576] Dec 16 12:47:04.540621 kernel: ACPI: Reserving DSDT table memory at [mem 0x7cfe0040-0x7cfe2482] Dec 16 12:47:04.540627 kernel: ACPI: Reserving FACS table memory at [mem 0x7cfe0000-0x7cfe003f] Dec 16 12:47:04.540635 kernel: ACPI: Reserving APIC table memory at [mem 0x7cfe2577-0x7cfe25f6] Dec 16 12:47:04.540641 kernel: ACPI: Reserving HPET table memory at [mem 0x7cfe25f7-0x7cfe262e] Dec 16 12:47:04.540647 kernel: ACPI: Reserving MCFG table memory at [mem 0x7cfe262f-0x7cfe266a] Dec 16 12:47:04.540653 kernel: ACPI: Reserving WAET table memory at [mem 0x7cfe266b-0x7cfe2692] Dec 16 12:47:04.540659 kernel: No NUMA configuration found Dec 16 12:47:04.540666 kernel: Faking a node at [mem 0x0000000000000000-0x000000007cfdbfff] Dec 16 12:47:04.540672 kernel: NODE_DATA(0) allocated [mem 0x7cfd4dc0-0x7cfdbfff] Dec 16 12:47:04.540680 kernel: Zone ranges: Dec 16 12:47:04.540686 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Dec 16 12:47:04.540723 kernel: DMA32 [mem 0x0000000001000000-0x000000007cfdbfff] Dec 16 12:47:04.540730 kernel: Normal empty Dec 16 12:47:04.540736 kernel: Device empty Dec 16 12:47:04.540742 kernel: Movable zone start for each node Dec 16 12:47:04.540748 kernel: Early memory node ranges Dec 16 12:47:04.540756 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Dec 16 12:47:04.540762 kernel: node 0: [mem 0x0000000000100000-0x000000007cfdbfff] Dec 16 12:47:04.540768 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007cfdbfff] Dec 16 12:47:04.540775 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Dec 16 12:47:04.540781 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Dec 16 12:47:04.540787 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Dec 16 12:47:04.540794 kernel: ACPI: PM-Timer IO Port: 0x608 Dec 16 12:47:04.540800 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Dec 16 12:47:04.540808 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Dec 16 12:47:04.540814 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Dec 16 12:47:04.540820 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Dec 16 12:47:04.540827 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Dec 16 12:47:04.540833 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Dec 16 12:47:04.540840 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Dec 16 12:47:04.540846 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Dec 16 12:47:04.540853 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Dec 16 12:47:04.540859 kernel: CPU topo: Max. logical packages: 1 Dec 16 12:47:04.540866 kernel: CPU topo: Max. logical dies: 1 Dec 16 12:47:04.540872 kernel: CPU topo: Max. dies per package: 1 Dec 16 12:47:04.540878 kernel: CPU topo: Max. threads per core: 1 Dec 16 12:47:04.540885 kernel: CPU topo: Num. cores per package: 2 Dec 16 12:47:04.540891 kernel: CPU topo: Num. threads per package: 2 Dec 16 12:47:04.540897 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs Dec 16 12:47:04.540904 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Dec 16 12:47:04.540910 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Dec 16 12:47:04.540917 kernel: Booting paravirtualized kernel on KVM Dec 16 12:47:04.540923 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Dec 16 12:47:04.540930 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Dec 16 12:47:04.540936 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u1048576 Dec 16 12:47:04.540942 kernel: pcpu-alloc: s207832 r8192 d29736 u1048576 alloc=1*2097152 Dec 16 12:47:04.540950 kernel: pcpu-alloc: [0] 0 1 Dec 16 12:47:04.540956 kernel: kvm-guest: PV spinlocks disabled, no host support Dec 16 12:47:04.540963 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=553464fdb0286a5b06b399da29ca659e521c68f08ea70a931c96ddffd00b4357 Dec 16 12:47:04.540970 kernel: random: crng init done Dec 16 12:47:04.540977 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Dec 16 12:47:04.540983 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Dec 16 12:47:04.540989 kernel: Fallback order for Node 0: 0 Dec 16 12:47:04.540997 kernel: Built 1 zonelists, mobility grouping on. Total pages: 511866 Dec 16 12:47:04.541003 kernel: Policy zone: DMA32 Dec 16 12:47:04.541010 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Dec 16 12:47:04.541032 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Dec 16 12:47:04.541039 kernel: ftrace: allocating 40103 entries in 157 pages Dec 16 12:47:04.541559 kernel: ftrace: allocated 157 pages with 5 groups Dec 16 12:47:04.541569 kernel: Dynamic Preempt: voluntary Dec 16 12:47:04.541579 kernel: rcu: Preemptible hierarchical RCU implementation. Dec 16 12:47:04.541587 kernel: rcu: RCU event tracing is enabled. Dec 16 12:47:04.541593 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Dec 16 12:47:04.541600 kernel: Trampoline variant of Tasks RCU enabled. Dec 16 12:47:04.541607 kernel: Rude variant of Tasks RCU enabled. Dec 16 12:47:04.541613 kernel: Tracing variant of Tasks RCU enabled. Dec 16 12:47:04.541620 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Dec 16 12:47:04.541627 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Dec 16 12:47:04.541634 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 16 12:47:04.541641 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 16 12:47:04.541648 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Dec 16 12:47:04.541654 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Dec 16 12:47:04.541661 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Dec 16 12:47:04.541667 kernel: Console: colour VGA+ 80x25 Dec 16 12:47:04.541675 kernel: printk: legacy console [tty0] enabled Dec 16 12:47:04.541681 kernel: printk: legacy console [ttyS0] enabled Dec 16 12:47:04.541688 kernel: ACPI: Core revision 20240827 Dec 16 12:47:04.541704 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Dec 16 12:47:04.541719 kernel: APIC: Switch to symmetric I/O mode setup Dec 16 12:47:04.541731 kernel: x2apic enabled Dec 16 12:47:04.541743 kernel: APIC: Switched APIC routing to: physical x2apic Dec 16 12:47:04.541755 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Dec 16 12:47:04.541764 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x233fc319723, max_idle_ns: 440795258057 ns Dec 16 12:47:04.541771 kernel: Calibrating delay loop (skipped) preset value.. 4890.80 BogoMIPS (lpj=2445404) Dec 16 12:47:04.541780 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Dec 16 12:47:04.541787 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Dec 16 12:47:04.541793 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Dec 16 12:47:04.541800 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Dec 16 12:47:04.541808 kernel: Spectre V2 : Mitigation: Retpolines Dec 16 12:47:04.541814 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Dec 16 12:47:04.541821 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Dec 16 12:47:04.541828 kernel: active return thunk: retbleed_return_thunk Dec 16 12:47:04.541834 kernel: RETBleed: Mitigation: untrained return thunk Dec 16 12:47:04.541841 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Dec 16 12:47:04.541847 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Dec 16 12:47:04.541855 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Dec 16 12:47:04.541862 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Dec 16 12:47:04.541869 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Dec 16 12:47:04.541875 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Dec 16 12:47:04.541882 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Dec 16 12:47:04.541889 kernel: Freeing SMP alternatives memory: 32K Dec 16 12:47:04.541895 kernel: pid_max: default: 32768 minimum: 301 Dec 16 12:47:04.541903 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Dec 16 12:47:04.541909 kernel: landlock: Up and running. Dec 16 12:47:04.541916 kernel: SELinux: Initializing. Dec 16 12:47:04.541923 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Dec 16 12:47:04.541929 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Dec 16 12:47:04.541936 kernel: smpboot: CPU0: AMD EPYC-Rome Processor (family: 0x17, model: 0x31, stepping: 0x0) Dec 16 12:47:04.541943 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Dec 16 12:47:04.541975 kernel: ... version: 0 Dec 16 12:47:04.541982 kernel: ... bit width: 48 Dec 16 12:47:04.541989 kernel: ... generic registers: 6 Dec 16 12:47:04.541996 kernel: ... value mask: 0000ffffffffffff Dec 16 12:47:04.542002 kernel: ... max period: 00007fffffffffff Dec 16 12:47:04.542009 kernel: ... fixed-purpose events: 0 Dec 16 12:47:04.548158 kernel: ... event mask: 000000000000003f Dec 16 12:47:04.548205 kernel: signal: max sigframe size: 1776 Dec 16 12:47:04.548231 kernel: rcu: Hierarchical SRCU implementation. Dec 16 12:47:04.548250 kernel: rcu: Max phase no-delay instances is 400. Dec 16 12:47:04.548270 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Dec 16 12:47:04.548293 kernel: smp: Bringing up secondary CPUs ... Dec 16 12:47:04.548318 kernel: smpboot: x86: Booting SMP configuration: Dec 16 12:47:04.548342 kernel: .... node #0, CPUs: #1 Dec 16 12:47:04.548369 kernel: smp: Brought up 1 node, 2 CPUs Dec 16 12:47:04.548393 kernel: smpboot: Total of 2 processors activated (9781.61 BogoMIPS) Dec 16 12:47:04.548417 kernel: Memory: 1934164K/2047464K available (14336K kernel code, 2444K rwdata, 31636K rodata, 15556K init, 2484K bss, 108756K reserved, 0K cma-reserved) Dec 16 12:47:04.548438 kernel: devtmpfs: initialized Dec 16 12:47:04.548464 kernel: x86/mm: Memory block size: 128MB Dec 16 12:47:04.548471 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Dec 16 12:47:04.548478 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Dec 16 12:47:04.548495 kernel: pinctrl core: initialized pinctrl subsystem Dec 16 12:47:04.548518 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Dec 16 12:47:04.548549 kernel: audit: initializing netlink subsys (disabled) Dec 16 12:47:04.548557 kernel: audit: type=2000 audit(1765889220.518:1): state=initialized audit_enabled=0 res=1 Dec 16 12:47:04.548563 kernel: thermal_sys: Registered thermal governor 'step_wise' Dec 16 12:47:04.548572 kernel: thermal_sys: Registered thermal governor 'user_space' Dec 16 12:47:04.548584 kernel: cpuidle: using governor menu Dec 16 12:47:04.548599 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Dec 16 12:47:04.548612 kernel: dca service started, version 1.12.1 Dec 16 12:47:04.548625 kernel: PCI: ECAM [mem 0xb0000000-0xbfffffff] (base 0xb0000000) for domain 0000 [bus 00-ff] Dec 16 12:47:04.548638 kernel: PCI: Using configuration type 1 for base access Dec 16 12:47:04.548650 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Dec 16 12:47:04.548663 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Dec 16 12:47:04.548677 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Dec 16 12:47:04.548695 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Dec 16 12:47:04.548706 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Dec 16 12:47:04.548713 kernel: ACPI: Added _OSI(Module Device) Dec 16 12:47:04.548720 kernel: ACPI: Added _OSI(Processor Device) Dec 16 12:47:04.548727 kernel: ACPI: Added _OSI(Processor Aggregator Device) Dec 16 12:47:04.548734 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Dec 16 12:47:04.548740 kernel: ACPI: Interpreter enabled Dec 16 12:47:04.548747 kernel: ACPI: PM: (supports S0 S5) Dec 16 12:47:04.548755 kernel: ACPI: Using IOAPIC for interrupt routing Dec 16 12:47:04.548762 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Dec 16 12:47:04.548768 kernel: PCI: Using E820 reservations for host bridge windows Dec 16 12:47:04.548775 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Dec 16 12:47:04.548781 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Dec 16 12:47:04.548915 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Dec 16 12:47:04.549009 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Dec 16 12:47:04.553507 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Dec 16 12:47:04.553520 kernel: PCI host bridge to bus 0000:00 Dec 16 12:47:04.553622 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Dec 16 12:47:04.556639 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Dec 16 12:47:04.556727 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Dec 16 12:47:04.556801 kernel: pci_bus 0000:00: root bus resource [mem 0x7d000000-0xafffffff window] Dec 16 12:47:04.556871 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Dec 16 12:47:04.556941 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Dec 16 12:47:04.557011 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Dec 16 12:47:04.557135 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Dec 16 12:47:04.557231 kernel: pci 0000:00:01.0: [1af4:1050] type 00 class 0x030000 conventional PCI endpoint Dec 16 12:47:04.557331 kernel: pci 0000:00:01.0: BAR 0 [mem 0xfb800000-0xfbffffff pref] Dec 16 12:47:04.557439 kernel: pci 0000:00:01.0: BAR 2 [mem 0xfd200000-0xfd203fff 64bit pref] Dec 16 12:47:04.557528 kernel: pci 0000:00:01.0: BAR 4 [mem 0xfea10000-0xfea10fff] Dec 16 12:47:04.557621 kernel: pci 0000:00:01.0: ROM [mem 0xfea00000-0xfea0ffff pref] Dec 16 12:47:04.557720 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Dec 16 12:47:04.557814 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 16 12:47:04.557893 kernel: pci 0000:00:02.0: BAR 0 [mem 0xfea11000-0xfea11fff] Dec 16 12:47:04.557971 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Dec 16 12:47:04.566189 kernel: pci 0000:00:02.0: bridge window [mem 0xfe800000-0xfe9fffff] Dec 16 12:47:04.566289 kernel: pci 0000:00:02.0: bridge window [mem 0xfd000000-0xfd1fffff 64bit pref] Dec 16 12:47:04.566382 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 16 12:47:04.566469 kernel: pci 0000:00:02.1: BAR 0 [mem 0xfea12000-0xfea12fff] Dec 16 12:47:04.566567 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Dec 16 12:47:04.566649 kernel: pci 0000:00:02.1: bridge window [mem 0xfe600000-0xfe7fffff] Dec 16 12:47:04.566727 kernel: pci 0000:00:02.1: bridge window [mem 0xfce00000-0xfcffffff 64bit pref] Dec 16 12:47:04.566811 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 16 12:47:04.566896 kernel: pci 0000:00:02.2: BAR 0 [mem 0xfea13000-0xfea13fff] Dec 16 12:47:04.566974 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Dec 16 12:47:04.567076 kernel: pci 0000:00:02.2: bridge window [mem 0xfe400000-0xfe5fffff] Dec 16 12:47:04.567156 kernel: pci 0000:00:02.2: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref] Dec 16 12:47:04.567240 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 16 12:47:04.567319 kernel: pci 0000:00:02.3: BAR 0 [mem 0xfea14000-0xfea14fff] Dec 16 12:47:04.567401 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Dec 16 12:47:04.567479 kernel: pci 0000:00:02.3: bridge window [mem 0xfe200000-0xfe3fffff] Dec 16 12:47:04.567567 kernel: pci 0000:00:02.3: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref] Dec 16 12:47:04.567652 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 16 12:47:04.567731 kernel: pci 0000:00:02.4: BAR 0 [mem 0xfea15000-0xfea15fff] Dec 16 12:47:04.567809 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Dec 16 12:47:04.567891 kernel: pci 0000:00:02.4: bridge window [mem 0xfe000000-0xfe1fffff] Dec 16 12:47:04.567968 kernel: pci 0000:00:02.4: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref] Dec 16 12:47:04.569186 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 16 12:47:04.569281 kernel: pci 0000:00:02.5: BAR 0 [mem 0xfea16000-0xfea16fff] Dec 16 12:47:04.569364 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Dec 16 12:47:04.569443 kernel: pci 0000:00:02.5: bridge window [mem 0xfde00000-0xfdffffff] Dec 16 12:47:04.575123 kernel: pci 0000:00:02.5: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref] Dec 16 12:47:04.575245 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 16 12:47:04.575331 kernel: pci 0000:00:02.6: BAR 0 [mem 0xfea17000-0xfea17fff] Dec 16 12:47:04.575413 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Dec 16 12:47:04.575520 kernel: pci 0000:00:02.6: bridge window [mem 0xfdc00000-0xfddfffff] Dec 16 12:47:04.575627 kernel: pci 0000:00:02.6: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref] Dec 16 12:47:04.575715 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 16 12:47:04.575793 kernel: pci 0000:00:02.7: BAR 0 [mem 0xfea18000-0xfea18fff] Dec 16 12:47:04.575873 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Dec 16 12:47:04.575955 kernel: pci 0000:00:02.7: bridge window [mem 0xfda00000-0xfdbfffff] Dec 16 12:47:04.576059 kernel: pci 0000:00:02.7: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref] Dec 16 12:47:04.576177 kernel: pci 0000:00:03.0: [1b36:000c] type 01 class 0x060400 PCIe Root Port Dec 16 12:47:04.576263 kernel: pci 0000:00:03.0: BAR 0 [mem 0xfea19000-0xfea19fff] Dec 16 12:47:04.576341 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Dec 16 12:47:04.576419 kernel: pci 0000:00:03.0: bridge window [mem 0xfd800000-0xfd9fffff] Dec 16 12:47:04.576495 kernel: pci 0000:00:03.0: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref] Dec 16 12:47:04.576596 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Dec 16 12:47:04.576680 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Dec 16 12:47:04.576764 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Dec 16 12:47:04.576842 kernel: pci 0000:00:1f.2: BAR 4 [io 0xc040-0xc05f] Dec 16 12:47:04.576919 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xfea1a000-0xfea1afff] Dec 16 12:47:04.577001 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Dec 16 12:47:04.577303 kernel: pci 0000:00:1f.3: BAR 4 [io 0x0700-0x073f] Dec 16 12:47:04.577397 kernel: pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000 PCIe Endpoint Dec 16 12:47:04.577480 kernel: pci 0000:01:00.0: BAR 1 [mem 0xfe880000-0xfe880fff] Dec 16 12:47:04.580119 kernel: pci 0000:01:00.0: BAR 4 [mem 0xfd000000-0xfd003fff 64bit pref] Dec 16 12:47:04.580213 kernel: pci 0000:01:00.0: ROM [mem 0xfe800000-0xfe87ffff pref] Dec 16 12:47:04.580294 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Dec 16 12:47:04.580391 kernel: pci 0000:02:00.0: [1b36:000d] type 00 class 0x0c0330 PCIe Endpoint Dec 16 12:47:04.580509 kernel: pci 0000:02:00.0: BAR 0 [mem 0xfe600000-0xfe603fff 64bit] Dec 16 12:47:04.580638 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Dec 16 12:47:04.580749 kernel: pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000 PCIe Endpoint Dec 16 12:47:04.580859 kernel: pci 0000:03:00.0: BAR 1 [mem 0xfe400000-0xfe400fff] Dec 16 12:47:04.580984 kernel: pci 0000:03:00.0: BAR 4 [mem 0xfcc00000-0xfcc03fff 64bit pref] Dec 16 12:47:04.581110 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Dec 16 12:47:04.581247 kernel: pci 0000:04:00.0: [1af4:1045] type 00 class 0x00ff00 PCIe Endpoint Dec 16 12:47:04.581340 kernel: pci 0000:04:00.0: BAR 4 [mem 0xfca00000-0xfca03fff 64bit pref] Dec 16 12:47:04.581423 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Dec 16 12:47:04.581516 kernel: pci 0000:05:00.0: [1af4:1044] type 00 class 0x00ff00 PCIe Endpoint Dec 16 12:47:04.581620 kernel: pci 0000:05:00.0: BAR 1 [mem 0xfe000000-0xfe000fff] Dec 16 12:47:04.581725 kernel: pci 0000:05:00.0: BAR 4 [mem 0xfc800000-0xfc803fff 64bit pref] Dec 16 12:47:04.581835 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Dec 16 12:47:04.581931 kernel: pci 0000:06:00.0: [1af4:1048] type 00 class 0x010000 PCIe Endpoint Dec 16 12:47:04.582235 kernel: pci 0000:06:00.0: BAR 1 [mem 0xfde00000-0xfde00fff] Dec 16 12:47:04.582352 kernel: pci 0000:06:00.0: BAR 4 [mem 0xfc600000-0xfc603fff 64bit pref] Dec 16 12:47:04.582440 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Dec 16 12:47:04.582450 kernel: acpiphp: Slot [0] registered Dec 16 12:47:04.582557 kernel: pci 0000:07:00.0: [1af4:1041] type 00 class 0x020000 PCIe Endpoint Dec 16 12:47:04.582649 kernel: pci 0000:07:00.0: BAR 1 [mem 0xfdc80000-0xfdc80fff] Dec 16 12:47:04.582734 kernel: pci 0000:07:00.0: BAR 4 [mem 0xfc400000-0xfc403fff 64bit pref] Dec 16 12:47:04.582822 kernel: pci 0000:07:00.0: ROM [mem 0xfdc00000-0xfdc7ffff pref] Dec 16 12:47:04.582909 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Dec 16 12:47:04.582926 kernel: acpiphp: Slot [0-2] registered Dec 16 12:47:04.583048 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Dec 16 12:47:04.583062 kernel: acpiphp: Slot [0-3] registered Dec 16 12:47:04.583146 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Dec 16 12:47:04.583160 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Dec 16 12:47:04.583167 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Dec 16 12:47:04.583174 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Dec 16 12:47:04.583181 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Dec 16 12:47:04.583187 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Dec 16 12:47:04.583194 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Dec 16 12:47:04.583201 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Dec 16 12:47:04.583209 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Dec 16 12:47:04.583216 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Dec 16 12:47:04.583223 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Dec 16 12:47:04.583230 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Dec 16 12:47:04.583236 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Dec 16 12:47:04.583243 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Dec 16 12:47:04.583250 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Dec 16 12:47:04.583257 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Dec 16 12:47:04.583272 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Dec 16 12:47:04.583285 kernel: iommu: Default domain type: Translated Dec 16 12:47:04.583297 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Dec 16 12:47:04.583309 kernel: PCI: Using ACPI for IRQ routing Dec 16 12:47:04.583321 kernel: PCI: pci_cache_line_size set to 64 bytes Dec 16 12:47:04.583333 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Dec 16 12:47:04.583345 kernel: e820: reserve RAM buffer [mem 0x7cfdc000-0x7fffffff] Dec 16 12:47:04.583458 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Dec 16 12:47:04.583630 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Dec 16 12:47:04.583720 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Dec 16 12:47:04.583731 kernel: vgaarb: loaded Dec 16 12:47:04.583738 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Dec 16 12:47:04.583745 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Dec 16 12:47:04.583752 kernel: clocksource: Switched to clocksource kvm-clock Dec 16 12:47:04.583761 kernel: VFS: Disk quotas dquot_6.6.0 Dec 16 12:47:04.583768 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Dec 16 12:47:04.583775 kernel: pnp: PnP ACPI init Dec 16 12:47:04.583866 kernel: system 00:04: [mem 0xb0000000-0xbfffffff window] has been reserved Dec 16 12:47:04.583878 kernel: pnp: PnP ACPI: found 5 devices Dec 16 12:47:04.583885 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Dec 16 12:47:04.583894 kernel: NET: Registered PF_INET protocol family Dec 16 12:47:04.583901 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Dec 16 12:47:04.583908 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Dec 16 12:47:04.583915 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Dec 16 12:47:04.583921 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Dec 16 12:47:04.583928 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Dec 16 12:47:04.583935 kernel: TCP: Hash tables configured (established 16384 bind 16384) Dec 16 12:47:04.583943 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Dec 16 12:47:04.583951 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Dec 16 12:47:04.583958 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Dec 16 12:47:04.583965 kernel: NET: Registered PF_XDP protocol family Dec 16 12:47:04.584079 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Dec 16 12:47:04.584168 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Dec 16 12:47:04.584251 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Dec 16 12:47:04.584362 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x1fff]: assigned Dec 16 12:47:04.584470 kernel: pci 0000:00:02.7: bridge window [io 0x2000-0x2fff]: assigned Dec 16 12:47:04.584580 kernel: pci 0000:00:03.0: bridge window [io 0x3000-0x3fff]: assigned Dec 16 12:47:04.584675 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Dec 16 12:47:04.584778 kernel: pci 0000:00:02.0: bridge window [mem 0xfe800000-0xfe9fffff] Dec 16 12:47:04.584863 kernel: pci 0000:00:02.0: bridge window [mem 0xfd000000-0xfd1fffff 64bit pref] Dec 16 12:47:04.584945 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Dec 16 12:47:04.585050 kernel: pci 0000:00:02.1: bridge window [mem 0xfe600000-0xfe7fffff] Dec 16 12:47:04.585151 kernel: pci 0000:00:02.1: bridge window [mem 0xfce00000-0xfcffffff 64bit pref] Dec 16 12:47:04.585236 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Dec 16 12:47:04.585318 kernel: pci 0000:00:02.2: bridge window [mem 0xfe400000-0xfe5fffff] Dec 16 12:47:04.585407 kernel: pci 0000:00:02.2: bridge window [mem 0xfcc00000-0xfcdfffff 64bit pref] Dec 16 12:47:04.585525 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Dec 16 12:47:04.585648 kernel: pci 0000:00:02.3: bridge window [mem 0xfe200000-0xfe3fffff] Dec 16 12:47:04.585737 kernel: pci 0000:00:02.3: bridge window [mem 0xfca00000-0xfcbfffff 64bit pref] Dec 16 12:47:04.585853 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Dec 16 12:47:04.585994 kernel: pci 0000:00:02.4: bridge window [mem 0xfe000000-0xfe1fffff] Dec 16 12:47:04.586104 kernel: pci 0000:00:02.4: bridge window [mem 0xfc800000-0xfc9fffff 64bit pref] Dec 16 12:47:04.586187 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Dec 16 12:47:04.586269 kernel: pci 0000:00:02.5: bridge window [mem 0xfde00000-0xfdffffff] Dec 16 12:47:04.586357 kernel: pci 0000:00:02.5: bridge window [mem 0xfc600000-0xfc7fffff 64bit pref] Dec 16 12:47:04.586441 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Dec 16 12:47:04.586556 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x1fff] Dec 16 12:47:04.586678 kernel: pci 0000:00:02.6: bridge window [mem 0xfdc00000-0xfddfffff] Dec 16 12:47:04.586768 kernel: pci 0000:00:02.6: bridge window [mem 0xfc400000-0xfc5fffff 64bit pref] Dec 16 12:47:04.586914 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Dec 16 12:47:04.587002 kernel: pci 0000:00:02.7: bridge window [io 0x2000-0x2fff] Dec 16 12:47:04.587108 kernel: pci 0000:00:02.7: bridge window [mem 0xfda00000-0xfdbfffff] Dec 16 12:47:04.589108 kernel: pci 0000:00:02.7: bridge window [mem 0xfc200000-0xfc3fffff 64bit pref] Dec 16 12:47:04.589234 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Dec 16 12:47:04.589321 kernel: pci 0000:00:03.0: bridge window [io 0x3000-0x3fff] Dec 16 12:47:04.589428 kernel: pci 0000:00:03.0: bridge window [mem 0xfd800000-0xfd9fffff] Dec 16 12:47:04.589609 kernel: pci 0000:00:03.0: bridge window [mem 0xfc000000-0xfc1fffff 64bit pref] Dec 16 12:47:04.589762 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Dec 16 12:47:04.589854 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Dec 16 12:47:04.589929 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Dec 16 12:47:04.590000 kernel: pci_bus 0000:00: resource 7 [mem 0x7d000000-0xafffffff window] Dec 16 12:47:04.590094 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Dec 16 12:47:04.590171 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Dec 16 12:47:04.590252 kernel: pci_bus 0000:01: resource 1 [mem 0xfe800000-0xfe9fffff] Dec 16 12:47:04.590325 kernel: pci_bus 0000:01: resource 2 [mem 0xfd000000-0xfd1fffff 64bit pref] Dec 16 12:47:04.590403 kernel: pci_bus 0000:02: resource 1 [mem 0xfe600000-0xfe7fffff] Dec 16 12:47:04.590476 kernel: pci_bus 0000:02: resource 2 [mem 0xfce00000-0xfcffffff 64bit pref] Dec 16 12:47:04.590574 kernel: pci_bus 0000:03: resource 1 [mem 0xfe400000-0xfe5fffff] Dec 16 12:47:04.590666 kernel: pci_bus 0000:03: resource 2 [mem 0xfcc00000-0xfcdfffff 64bit pref] Dec 16 12:47:04.590813 kernel: pci_bus 0000:04: resource 1 [mem 0xfe200000-0xfe3fffff] Dec 16 12:47:04.590893 kernel: pci_bus 0000:04: resource 2 [mem 0xfca00000-0xfcbfffff 64bit pref] Dec 16 12:47:04.590972 kernel: pci_bus 0000:05: resource 1 [mem 0xfe000000-0xfe1fffff] Dec 16 12:47:04.591080 kernel: pci_bus 0000:05: resource 2 [mem 0xfc800000-0xfc9fffff 64bit pref] Dec 16 12:47:04.591203 kernel: pci_bus 0000:06: resource 1 [mem 0xfde00000-0xfdffffff] Dec 16 12:47:04.591277 kernel: pci_bus 0000:06: resource 2 [mem 0xfc600000-0xfc7fffff 64bit pref] Dec 16 12:47:04.591357 kernel: pci_bus 0000:07: resource 0 [io 0x1000-0x1fff] Dec 16 12:47:04.591430 kernel: pci_bus 0000:07: resource 1 [mem 0xfdc00000-0xfddfffff] Dec 16 12:47:04.591501 kernel: pci_bus 0000:07: resource 2 [mem 0xfc400000-0xfc5fffff 64bit pref] Dec 16 12:47:04.591599 kernel: pci_bus 0000:08: resource 0 [io 0x2000-0x2fff] Dec 16 12:47:04.591673 kernel: pci_bus 0000:08: resource 1 [mem 0xfda00000-0xfdbfffff] Dec 16 12:47:04.591745 kernel: pci_bus 0000:08: resource 2 [mem 0xfc200000-0xfc3fffff 64bit pref] Dec 16 12:47:04.591821 kernel: pci_bus 0000:09: resource 0 [io 0x3000-0x3fff] Dec 16 12:47:04.591897 kernel: pci_bus 0000:09: resource 1 [mem 0xfd800000-0xfd9fffff] Dec 16 12:47:04.591970 kernel: pci_bus 0000:09: resource 2 [mem 0xfc000000-0xfc1fffff 64bit pref] Dec 16 12:47:04.591981 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Dec 16 12:47:04.591989 kernel: PCI: CLS 0 bytes, default 64 Dec 16 12:47:04.591996 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x233fc319723, max_idle_ns: 440795258057 ns Dec 16 12:47:04.592003 kernel: Initialise system trusted keyrings Dec 16 12:47:04.592011 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Dec 16 12:47:04.592053 kernel: Key type asymmetric registered Dec 16 12:47:04.592080 kernel: Asymmetric key parser 'x509' registered Dec 16 12:47:04.592089 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Dec 16 12:47:04.592096 kernel: io scheduler mq-deadline registered Dec 16 12:47:04.592103 kernel: io scheduler kyber registered Dec 16 12:47:04.592111 kernel: io scheduler bfq registered Dec 16 12:47:04.592202 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 24 Dec 16 12:47:04.592284 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 24 Dec 16 12:47:04.592367 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 25 Dec 16 12:47:04.592446 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 25 Dec 16 12:47:04.592524 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 26 Dec 16 12:47:04.592622 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 26 Dec 16 12:47:04.592701 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 27 Dec 16 12:47:04.592779 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 27 Dec 16 12:47:04.592856 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 28 Dec 16 12:47:04.592936 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 28 Dec 16 12:47:04.593029 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 29 Dec 16 12:47:04.593114 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 29 Dec 16 12:47:04.593192 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 30 Dec 16 12:47:04.593269 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 30 Dec 16 12:47:04.593345 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 31 Dec 16 12:47:04.593426 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 31 Dec 16 12:47:04.593436 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Dec 16 12:47:04.593511 kernel: pcieport 0000:00:03.0: PME: Signaling with IRQ 32 Dec 16 12:47:04.593601 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 32 Dec 16 12:47:04.593613 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Dec 16 12:47:04.593624 kernel: ACPI: \_SB_.GSIF: Enabled at IRQ 21 Dec 16 12:47:04.593631 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Dec 16 12:47:04.593639 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Dec 16 12:47:04.593648 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Dec 16 12:47:04.593655 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Dec 16 12:47:04.593662 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Dec 16 12:47:04.593746 kernel: rtc_cmos 00:03: RTC can wake from S4 Dec 16 12:47:04.593760 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input1 Dec 16 12:47:04.593832 kernel: rtc_cmos 00:03: registered as rtc0 Dec 16 12:47:04.593905 kernel: rtc_cmos 00:03: setting system clock to 2025-12-16T12:47:02 UTC (1765889222) Dec 16 12:47:04.593979 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Dec 16 12:47:04.593989 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Dec 16 12:47:04.593997 kernel: NET: Registered PF_INET6 protocol family Dec 16 12:47:04.594007 kernel: Segment Routing with IPv6 Dec 16 12:47:04.594028 kernel: In-situ OAM (IOAM) with IPv6 Dec 16 12:47:04.594044 kernel: NET: Registered PF_PACKET protocol family Dec 16 12:47:04.594052 kernel: Key type dns_resolver registered Dec 16 12:47:04.594059 kernel: IPI shorthand broadcast: enabled Dec 16 12:47:04.594066 kernel: sched_clock: Marking stable (2329011705, 249383379)->(2612540266, -34145182) Dec 16 12:47:04.594073 kernel: registered taskstats version 1 Dec 16 12:47:04.594082 kernel: Loading compiled-in X.509 certificates Dec 16 12:47:04.594089 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.61-flatcar: aafd1eb27ea805b8231c3bede9210239fae84df8' Dec 16 12:47:04.594097 kernel: Demotion targets for Node 0: null Dec 16 12:47:04.594104 kernel: Key type .fscrypt registered Dec 16 12:47:04.594111 kernel: Key type fscrypt-provisioning registered Dec 16 12:47:04.594118 kernel: ima: No TPM chip found, activating TPM-bypass! Dec 16 12:47:04.594125 kernel: ima: Allocated hash algorithm: sha1 Dec 16 12:47:04.594133 kernel: ima: No architecture policies found Dec 16 12:47:04.594140 kernel: clk: Disabling unused clocks Dec 16 12:47:04.594147 kernel: Freeing unused kernel image (initmem) memory: 15556K Dec 16 12:47:04.594154 kernel: Write protecting the kernel read-only data: 47104k Dec 16 12:47:04.594161 kernel: Freeing unused kernel image (rodata/data gap) memory: 1132K Dec 16 12:47:04.594169 kernel: Run /init as init process Dec 16 12:47:04.594176 kernel: with arguments: Dec 16 12:47:04.594184 kernel: /init Dec 16 12:47:04.594192 kernel: with environment: Dec 16 12:47:04.594199 kernel: HOME=/ Dec 16 12:47:04.594206 kernel: TERM=linux Dec 16 12:47:04.594213 kernel: ACPI: bus type USB registered Dec 16 12:47:04.594220 kernel: usbcore: registered new interface driver usbfs Dec 16 12:47:04.594227 kernel: usbcore: registered new interface driver hub Dec 16 12:47:04.594234 kernel: usbcore: registered new device driver usb Dec 16 12:47:04.594325 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Dec 16 12:47:04.594406 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 1 Dec 16 12:47:04.594485 kernel: xhci_hcd 0000:02:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Dec 16 12:47:04.594577 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Dec 16 12:47:04.594659 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 2 Dec 16 12:47:04.594738 kernel: xhci_hcd 0000:02:00.0: Host supports USB 3.0 SuperSpeed Dec 16 12:47:04.594843 kernel: hub 1-0:1.0: USB hub found Dec 16 12:47:04.594930 kernel: hub 1-0:1.0: 4 ports detected Dec 16 12:47:04.595049 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Dec 16 12:47:04.595152 kernel: hub 2-0:1.0: USB hub found Dec 16 12:47:04.595239 kernel: hub 2-0:1.0: 4 ports detected Dec 16 12:47:04.595253 kernel: SCSI subsystem initialized Dec 16 12:47:04.595261 kernel: libata version 3.00 loaded. Dec 16 12:47:04.595340 kernel: ahci 0000:00:1f.2: version 3.0 Dec 16 12:47:04.595351 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Dec 16 12:47:04.595427 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Dec 16 12:47:04.595504 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Dec 16 12:47:04.595600 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Dec 16 12:47:04.595691 kernel: scsi host0: ahci Dec 16 12:47:04.595776 kernel: scsi host1: ahci Dec 16 12:47:04.595860 kernel: scsi host2: ahci Dec 16 12:47:04.595943 kernel: scsi host3: ahci Dec 16 12:47:04.596054 kernel: scsi host4: ahci Dec 16 12:47:04.596149 kernel: scsi host5: ahci Dec 16 12:47:04.596160 kernel: ata1: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a100 irq 38 lpm-pol 1 Dec 16 12:47:04.596167 kernel: ata2: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a180 irq 38 lpm-pol 1 Dec 16 12:47:04.596175 kernel: ata3: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a200 irq 38 lpm-pol 1 Dec 16 12:47:04.596182 kernel: ata4: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a280 irq 38 lpm-pol 1 Dec 16 12:47:04.596189 kernel: ata5: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a300 irq 38 lpm-pol 1 Dec 16 12:47:04.596199 kernel: ata6: SATA max UDMA/133 abar m4096@0xfea1a000 port 0xfea1a380 irq 38 lpm-pol 1 Dec 16 12:47:04.596298 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Dec 16 12:47:04.596310 kernel: hid: raw HID events driver (C) Jiri Kosina Dec 16 12:47:04.596317 kernel: ata1: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Dec 16 12:47:04.596324 kernel: ata4: SATA link down (SStatus 0 SControl 300) Dec 16 12:47:04.596332 kernel: ata6: SATA link down (SStatus 0 SControl 300) Dec 16 12:47:04.596341 kernel: ata2: SATA link down (SStatus 0 SControl 300) Dec 16 12:47:04.596348 kernel: ata1.00: LPM support broken, forcing max_power Dec 16 12:47:04.596355 kernel: ata1.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Dec 16 12:47:04.596363 kernel: ata1.00: applying bridge limits Dec 16 12:47:04.596370 kernel: ata3: SATA link down (SStatus 0 SControl 300) Dec 16 12:47:04.596377 kernel: ata5: SATA link down (SStatus 0 SControl 300) Dec 16 12:47:04.596384 kernel: ata1.00: LPM support broken, forcing max_power Dec 16 12:47:04.596392 kernel: ata1.00: configured for UDMA/100 Dec 16 12:47:04.596487 kernel: scsi 0:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Dec 16 12:47:04.596498 kernel: usbcore: registered new interface driver usbhid Dec 16 12:47:04.596505 kernel: usbhid: USB HID core driver Dec 16 12:47:04.596609 kernel: virtio_scsi virtio5: 2/0/0 default/read/poll queues Dec 16 12:47:04.596696 kernel: sr 0:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Dec 16 12:47:04.596788 kernel: scsi host6: Virtio SCSI HBA Dec 16 12:47:04.596800 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Dec 16 12:47:04.596893 kernel: scsi 6:0:0:0: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Dec 16 12:47:04.596980 kernel: sr 0:0:0:0: Attached scsi CD-ROM sr0 Dec 16 12:47:04.597088 kernel: sd 6:0:0:0: Power-on or device reset occurred Dec 16 12:47:04.597175 kernel: sd 6:0:0:0: [sda] 80003072 512-byte logical blocks: (41.0 GB/38.1 GiB) Dec 16 12:47:04.597189 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input3 Dec 16 12:47:04.597273 kernel: sd 6:0:0:0: [sda] Write Protect is off Dec 16 12:47:04.597377 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:02:00.0-1/input0 Dec 16 12:47:04.597465 kernel: sd 6:0:0:0: [sda] Mode Sense: 63 00 00 08 Dec 16 12:47:04.597591 kernel: sd 6:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Dec 16 12:47:04.597607 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Dec 16 12:47:04.597614 kernel: GPT:25804799 != 80003071 Dec 16 12:47:04.597622 kernel: GPT:Alternate GPT header not at the end of the disk. Dec 16 12:47:04.597629 kernel: GPT:25804799 != 80003071 Dec 16 12:47:04.597636 kernel: GPT: Use GNU Parted to correct GPT errors. Dec 16 12:47:04.597643 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Dec 16 12:47:04.597729 kernel: sd 6:0:0:0: [sda] Attached SCSI disk Dec 16 12:47:04.597742 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Dec 16 12:47:04.597749 kernel: device-mapper: uevent: version 1.0.3 Dec 16 12:47:04.597756 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Dec 16 12:47:04.597764 kernel: device-mapper: verity: sha256 using shash "sha256-generic" Dec 16 12:47:04.597771 kernel: raid6: avx2x4 gen() 17901 MB/s Dec 16 12:47:04.597778 kernel: raid6: avx2x2 gen() 18104 MB/s Dec 16 12:47:04.597785 kernel: raid6: avx2x1 gen() 17989 MB/s Dec 16 12:47:04.597792 kernel: raid6: using algorithm avx2x2 gen() 18104 MB/s Dec 16 12:47:04.597800 kernel: raid6: .... xor() 27758 MB/s, rmw enabled Dec 16 12:47:04.597808 kernel: raid6: using avx2x2 recovery algorithm Dec 16 12:47:04.597815 kernel: xor: automatically using best checksumming function avx Dec 16 12:47:04.597822 kernel: Btrfs loaded, zoned=no, fsverity=no Dec 16 12:47:04.597829 kernel: BTRFS: device fsid 57a8262f-2900-48ba-a17e-aafbd70d59c7 devid 1 transid 36 /dev/mapper/usr (254:0) scanned by mount (184) Dec 16 12:47:04.597837 kernel: BTRFS info (device dm-0): first mount of filesystem 57a8262f-2900-48ba-a17e-aafbd70d59c7 Dec 16 12:47:04.597844 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Dec 16 12:47:04.597852 kernel: BTRFS info (device dm-0): enabling ssd optimizations Dec 16 12:47:04.597859 kernel: BTRFS info (device dm-0): disabling log replay at mount time Dec 16 12:47:04.597866 kernel: BTRFS info (device dm-0): enabling free space tree Dec 16 12:47:04.597874 kernel: loop: module loaded Dec 16 12:47:04.597881 kernel: loop0: detected capacity change from 0 to 100528 Dec 16 12:47:04.597888 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Dec 16 12:47:04.597896 systemd[1]: Successfully made /usr/ read-only. Dec 16 12:47:04.597908 systemd[1]: systemd 257.9 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Dec 16 12:47:04.597916 systemd[1]: Detected virtualization kvm. Dec 16 12:47:04.597923 systemd[1]: Detected architecture x86-64. Dec 16 12:47:04.597930 systemd[1]: Running in initrd. Dec 16 12:47:04.597937 systemd[1]: No hostname configured, using default hostname. Dec 16 12:47:04.597946 systemd[1]: Hostname set to . Dec 16 12:47:04.597953 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Dec 16 12:47:04.597961 systemd[1]: Queued start job for default target initrd.target. Dec 16 12:47:04.597968 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Dec 16 12:47:04.597976 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 16 12:47:04.597983 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 16 12:47:04.597991 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Dec 16 12:47:04.598000 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 16 12:47:04.598008 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Dec 16 12:47:04.598029 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Dec 16 12:47:04.598046 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 16 12:47:04.598053 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 16 12:47:04.598063 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Dec 16 12:47:04.598071 systemd[1]: Reached target paths.target - Path Units. Dec 16 12:47:04.598079 systemd[1]: Reached target slices.target - Slice Units. Dec 16 12:47:04.598086 systemd[1]: Reached target swap.target - Swaps. Dec 16 12:47:04.598093 systemd[1]: Reached target timers.target - Timer Units. Dec 16 12:47:04.598101 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Dec 16 12:47:04.598109 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 16 12:47:04.598118 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Dec 16 12:47:04.598125 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Dec 16 12:47:04.598132 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Dec 16 12:47:04.598140 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 16 12:47:04.598147 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 16 12:47:04.598155 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 16 12:47:04.598162 systemd[1]: Reached target sockets.target - Socket Units. Dec 16 12:47:04.598172 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Dec 16 12:47:04.598179 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Dec 16 12:47:04.598187 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 16 12:47:04.598194 systemd[1]: Finished network-cleanup.service - Network Cleanup. Dec 16 12:47:04.598202 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Dec 16 12:47:04.598210 systemd[1]: Starting systemd-fsck-usr.service... Dec 16 12:47:04.598218 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 16 12:47:04.598227 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 16 12:47:04.598234 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 16 12:47:04.598242 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Dec 16 12:47:04.598270 systemd-journald[321]: Collecting audit messages is enabled. Dec 16 12:47:04.598290 kernel: audit: type=1130 audit(1765889224.541:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup-pre comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.598298 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Dec 16 12:47:04.598307 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 16 12:47:04.598314 systemd[1]: Finished systemd-fsck-usr.service. Dec 16 12:47:04.598322 kernel: audit: type=1130 audit(1765889224.556:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.598330 kernel: Bridge firewalling registered Dec 16 12:47:04.598337 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 16 12:47:04.598344 kernel: audit: type=1130 audit(1765889224.566:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.598353 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 16 12:47:04.598361 kernel: audit: type=1130 audit(1765889224.575:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.598369 systemd-journald[321]: Journal started Dec 16 12:47:04.598387 systemd-journald[321]: Runtime Journal (/run/log/journal/afbec2ca433d48b69d29e348d9cb3439) is 4.7M, max 38.1M, 33.4M free. Dec 16 12:47:04.541000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup-pre comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.556000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.566000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.575000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.565179 systemd-modules-load[323]: Inserted module 'br_netfilter' Dec 16 12:47:04.672046 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 16 12:47:04.675045 systemd[1]: Started systemd-journald.service - Journal Service. Dec 16 12:47:04.675000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.682397 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 16 12:47:04.684225 kernel: audit: type=1130 audit(1765889224.675:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.684000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.692206 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 16 12:47:04.699927 kernel: audit: type=1130 audit(1765889224.684:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.699947 kernel: audit: type=1130 audit(1765889224.693:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.693000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.693571 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 16 12:47:04.707411 kernel: audit: type=1130 audit(1765889224.700:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.700000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.703494 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Dec 16 12:47:04.711451 kernel: audit: type=1334 audit(1765889224.709:10): prog-id=6 op=LOAD Dec 16 12:47:04.709000 audit: BPF prog-id=6 op=LOAD Dec 16 12:47:04.710167 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 16 12:47:04.715592 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 16 12:47:04.720799 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 16 12:47:04.736270 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 16 12:47:04.738000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.740115 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 16 12:47:04.741000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.741793 systemd-tmpfiles[345]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Dec 16 12:47:04.744830 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Dec 16 12:47:04.753000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.752459 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 16 12:47:04.765684 dracut-cmdline[358]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=hetzner verity.usrhash=553464fdb0286a5b06b399da29ca659e521c68f08ea70a931c96ddffd00b4357 Dec 16 12:47:04.765945 systemd-resolved[341]: Positive Trust Anchors: Dec 16 12:47:04.765951 systemd-resolved[341]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 16 12:47:04.765954 systemd-resolved[341]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Dec 16 12:47:04.765980 systemd-resolved[341]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 16 12:47:04.782740 systemd-resolved[341]: Defaulting to hostname 'linux'. Dec 16 12:47:04.787000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.785855 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 16 12:47:04.788086 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 16 12:47:04.837050 kernel: Loading iSCSI transport class v2.0-870. Dec 16 12:47:04.849044 kernel: iscsi: registered transport (tcp) Dec 16 12:47:04.868512 kernel: iscsi: registered transport (qla4xxx) Dec 16 12:47:04.868563 kernel: QLogic iSCSI HBA Driver Dec 16 12:47:04.887527 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Dec 16 12:47:04.899833 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Dec 16 12:47:04.900000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.902170 systemd[1]: Reached target network-pre.target - Preparation for Network. Dec 16 12:47:04.930811 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Dec 16 12:47:04.931000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.934114 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Dec 16 12:47:04.935902 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Dec 16 12:47:04.959468 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Dec 16 12:47:04.960000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:04.962000 audit: BPF prog-id=7 op=LOAD Dec 16 12:47:04.962000 audit: BPF prog-id=8 op=LOAD Dec 16 12:47:04.963153 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 16 12:47:04.989584 systemd-udevd[600]: Using default interface naming scheme 'v257'. Dec 16 12:47:04.998139 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 16 12:47:04.999000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:05.000248 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 16 12:47:05.002000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:05.004435 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Dec 16 12:47:05.007000 audit: BPF prog-id=9 op=LOAD Dec 16 12:47:05.010106 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 16 12:47:05.022214 dracut-pre-trigger[699]: rd.md=0: removing MD RAID activation Dec 16 12:47:05.045757 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Dec 16 12:47:05.047000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:05.046275 systemd-networkd[700]: lo: Link UP Dec 16 12:47:05.046278 systemd-networkd[700]: lo: Gained carrier Dec 16 12:47:05.049000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:05.047482 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 16 12:47:05.049473 systemd[1]: Reached target network.target - Network. Dec 16 12:47:05.053145 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 16 12:47:05.116895 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 16 12:47:05.119000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:05.122467 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Dec 16 12:47:05.205732 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Dec 16 12:47:05.226558 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Dec 16 12:47:05.235107 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Dec 16 12:47:05.247047 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Dec 16 12:47:05.251114 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Dec 16 12:47:05.282454 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 16 12:47:05.282582 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 16 12:47:05.286000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:05.290085 disk-uuid[772]: Primary Header is updated. Dec 16 12:47:05.290085 disk-uuid[772]: Secondary Entries is updated. Dec 16 12:47:05.290085 disk-uuid[772]: Secondary Header is updated. Dec 16 12:47:05.286238 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Dec 16 12:47:05.297207 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 16 12:47:05.304885 systemd-networkd[700]: eth1: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Dec 16 12:47:05.306336 kernel: cryptd: max_cpu_qlen set to 1000 Dec 16 12:47:05.306070 systemd-networkd[700]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 16 12:47:05.307721 systemd-networkd[700]: eth1: Link UP Dec 16 12:47:05.309742 systemd-networkd[700]: eth1: Gained carrier Dec 16 12:47:05.309754 systemd-networkd[700]: eth1: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Dec 16 12:47:05.322943 systemd-networkd[700]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Dec 16 12:47:05.322949 systemd-networkd[700]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 16 12:47:05.324259 systemd-networkd[700]: eth0: Link UP Dec 16 12:47:05.340563 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Dec 16 12:47:05.327253 systemd-networkd[700]: eth0: Gained carrier Dec 16 12:47:05.327263 systemd-networkd[700]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Dec 16 12:47:05.454667 kernel: AES CTR mode by8 optimization enabled Dec 16 12:47:05.371245 systemd-networkd[700]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Dec 16 12:47:05.384057 systemd-networkd[700]: eth0: DHCPv4 address 77.42.19.41/32, gateway 172.31.1.1 acquired from 172.31.1.1 Dec 16 12:47:05.464328 kernel: kauditd_printk_skb: 16 callbacks suppressed Dec 16 12:47:05.464346 kernel: audit: type=1130 audit(1765889225.456:27): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:05.456000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:05.398807 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Dec 16 12:47:05.471991 kernel: audit: type=1130 audit(1765889225.464:28): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:05.464000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:05.456539 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 16 12:47:05.465631 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Dec 16 12:47:05.472650 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 16 12:47:05.474340 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 16 12:47:05.477115 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Dec 16 12:47:05.496000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:05.495670 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Dec 16 12:47:05.503237 kernel: audit: type=1130 audit(1765889225.496:29): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.333233 disk-uuid[775]: Warning: The kernel is still using the old partition table. Dec 16 12:47:06.333233 disk-uuid[775]: The new table will be used at the next reboot or after you Dec 16 12:47:06.333233 disk-uuid[775]: run partprobe(8) or kpartx(8) Dec 16 12:47:06.333233 disk-uuid[775]: The operation has completed successfully. Dec 16 12:47:06.342390 systemd[1]: disk-uuid.service: Deactivated successfully. Dec 16 12:47:06.342565 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Dec 16 12:47:06.370619 kernel: audit: type=1130 audit(1765889226.344:30): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.370669 kernel: audit: type=1131 audit(1765889226.344:31): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.344000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.344000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.346196 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Dec 16 12:47:06.422075 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (860) Dec 16 12:47:06.428856 kernel: BTRFS info (device sda6): first mount of filesystem 7e31dbd7-b976-4d4a-a2e9-e2baacf4ad38 Dec 16 12:47:06.428899 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Dec 16 12:47:06.439642 kernel: BTRFS info (device sda6): enabling ssd optimizations Dec 16 12:47:06.439691 kernel: BTRFS info (device sda6): turning on async discard Dec 16 12:47:06.439703 kernel: BTRFS info (device sda6): enabling free space tree Dec 16 12:47:06.451096 kernel: BTRFS info (device sda6): last unmount of filesystem 7e31dbd7-b976-4d4a-a2e9-e2baacf4ad38 Dec 16 12:47:06.451827 systemd[1]: Finished ignition-setup.service - Ignition (setup). Dec 16 12:47:06.460580 kernel: audit: type=1130 audit(1765889226.452:32): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.452000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.455131 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Dec 16 12:47:06.575591 ignition[879]: Ignition 2.24.0 Dec 16 12:47:06.576398 ignition[879]: Stage: fetch-offline Dec 16 12:47:06.577648 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Dec 16 12:47:06.585663 kernel: audit: type=1130 audit(1765889226.578:33): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.578000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.576437 ignition[879]: no configs at "/usr/lib/ignition/base.d" Dec 16 12:47:06.581183 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Dec 16 12:47:06.576447 ignition[879]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 16 12:47:06.576531 ignition[879]: parsed url from cmdline: "" Dec 16 12:47:06.576534 ignition[879]: no config URL provided Dec 16 12:47:06.576538 ignition[879]: reading system config file "/usr/lib/ignition/user.ign" Dec 16 12:47:06.576546 ignition[879]: no config at "/usr/lib/ignition/user.ign" Dec 16 12:47:06.576549 ignition[879]: failed to fetch config: resource requires networking Dec 16 12:47:06.576680 ignition[879]: Ignition finished successfully Dec 16 12:47:06.601632 ignition[886]: Ignition 2.24.0 Dec 16 12:47:06.601644 ignition[886]: Stage: fetch Dec 16 12:47:06.601779 ignition[886]: no configs at "/usr/lib/ignition/base.d" Dec 16 12:47:06.601787 ignition[886]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 16 12:47:06.601857 ignition[886]: parsed url from cmdline: "" Dec 16 12:47:06.601860 ignition[886]: no config URL provided Dec 16 12:47:06.601866 ignition[886]: reading system config file "/usr/lib/ignition/user.ign" Dec 16 12:47:06.601871 ignition[886]: no config at "/usr/lib/ignition/user.ign" Dec 16 12:47:06.601905 ignition[886]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #1 Dec 16 12:47:06.607044 ignition[886]: GET result: OK Dec 16 12:47:06.607146 ignition[886]: parsing config with SHA512: cbf06eeda263a518eea16eb1b184f3d0afdb1ddf30608db629474cd1949578d2d0ae2d3f92a0a0b3f05181147ae237b9631813d1378d0ebb0ba137062e45a662 Dec 16 12:47:06.612206 unknown[886]: fetched base config from "system" Dec 16 12:47:06.612448 ignition[886]: fetch: fetch complete Dec 16 12:47:06.612213 unknown[886]: fetched base config from "system" Dec 16 12:47:06.612452 ignition[886]: fetch: fetch passed Dec 16 12:47:06.612217 unknown[886]: fetched user config from "hetzner" Dec 16 12:47:06.622210 kernel: audit: type=1130 audit(1765889226.615:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.615000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.612481 ignition[886]: Ignition finished successfully Dec 16 12:47:06.614987 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Dec 16 12:47:06.616693 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Dec 16 12:47:06.640722 ignition[892]: Ignition 2.24.0 Dec 16 12:47:06.640734 ignition[892]: Stage: kargs Dec 16 12:47:06.640853 ignition[892]: no configs at "/usr/lib/ignition/base.d" Dec 16 12:47:06.640860 ignition[892]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 16 12:47:06.644928 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Dec 16 12:47:06.646000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.641472 ignition[892]: kargs: kargs passed Dec 16 12:47:06.654188 kernel: audit: type=1130 audit(1765889226.646:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.649132 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Dec 16 12:47:06.641505 ignition[892]: Ignition finished successfully Dec 16 12:47:06.675693 ignition[898]: Ignition 2.24.0 Dec 16 12:47:06.675705 ignition[898]: Stage: disks Dec 16 12:47:06.685844 kernel: audit: type=1130 audit(1765889226.678:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.678000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.677460 systemd[1]: Finished ignition-disks.service - Ignition (disks). Dec 16 12:47:06.675822 ignition[898]: no configs at "/usr/lib/ignition/base.d" Dec 16 12:47:06.678496 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Dec 16 12:47:06.675828 ignition[898]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 16 12:47:06.686728 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Dec 16 12:47:06.676445 ignition[898]: disks: disks passed Dec 16 12:47:06.688540 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 16 12:47:06.676477 ignition[898]: Ignition finished successfully Dec 16 12:47:06.690192 systemd[1]: Reached target sysinit.target - System Initialization. Dec 16 12:47:06.692045 systemd[1]: Reached target basic.target - Basic System. Dec 16 12:47:06.694815 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Dec 16 12:47:06.724663 systemd-fsck[906]: ROOT: clean, 15/1631200 files, 112378/1617920 blocks Dec 16 12:47:06.726925 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Dec 16 12:47:06.728000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:06.729961 systemd[1]: Mounting sysroot.mount - /sysroot... Dec 16 12:47:06.852065 kernel: EXT4-fs (sda9): mounted filesystem 1314c107-11a5-486b-9d52-be9f57b6bf1b r/w with ordered data mode. Quota mode: none. Dec 16 12:47:06.852934 systemd[1]: Mounted sysroot.mount - /sysroot. Dec 16 12:47:06.854247 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Dec 16 12:47:06.858146 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 16 12:47:06.862140 systemd-networkd[700]: eth1: Gained IPv6LL Dec 16 12:47:06.863795 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Dec 16 12:47:06.869221 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Dec 16 12:47:06.876387 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Dec 16 12:47:06.889169 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (914) Dec 16 12:47:06.889193 kernel: BTRFS info (device sda6): first mount of filesystem 7e31dbd7-b976-4d4a-a2e9-e2baacf4ad38 Dec 16 12:47:06.889205 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Dec 16 12:47:06.877135 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Dec 16 12:47:06.892921 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Dec 16 12:47:06.897124 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Dec 16 12:47:06.907658 kernel: BTRFS info (device sda6): enabling ssd optimizations Dec 16 12:47:06.907684 kernel: BTRFS info (device sda6): turning on async discard Dec 16 12:47:06.907695 kernel: BTRFS info (device sda6): enabling free space tree Dec 16 12:47:06.911135 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 16 12:47:06.989456 systemd-networkd[700]: eth0: Gained IPv6LL Dec 16 12:47:06.997651 coreos-metadata[916]: Dec 16 12:47:06.997 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/hostname: Attempt #1 Dec 16 12:47:06.999567 coreos-metadata[916]: Dec 16 12:47:06.999 INFO Fetch successful Dec 16 12:47:06.999567 coreos-metadata[916]: Dec 16 12:47:06.999 INFO wrote hostname ci-4547-0-0-6-b8319fed62 to /sysroot/etc/hostname Dec 16 12:47:07.002030 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Dec 16 12:47:07.004000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-metadata-hostname comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:07.139703 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Dec 16 12:47:07.141000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:07.143532 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Dec 16 12:47:07.149205 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Dec 16 12:47:07.166741 systemd[1]: sysroot-oem.mount: Deactivated successfully. Dec 16 12:47:07.176383 kernel: BTRFS info (device sda6): last unmount of filesystem 7e31dbd7-b976-4d4a-a2e9-e2baacf4ad38 Dec 16 12:47:07.197413 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Dec 16 12:47:07.198000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:07.204856 ignition[1016]: INFO : Ignition 2.24.0 Dec 16 12:47:07.204856 ignition[1016]: INFO : Stage: mount Dec 16 12:47:07.206334 ignition[1016]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 16 12:47:07.206334 ignition[1016]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 16 12:47:07.206334 ignition[1016]: INFO : mount: mount passed Dec 16 12:47:07.206334 ignition[1016]: INFO : Ignition finished successfully Dec 16 12:47:07.208000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:07.206556 systemd[1]: Finished ignition-mount.service - Ignition (mount). Dec 16 12:47:07.211143 systemd[1]: Starting ignition-files.service - Ignition (files)... Dec 16 12:47:07.856275 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Dec 16 12:47:07.899082 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (1027) Dec 16 12:47:07.899143 kernel: BTRFS info (device sda6): first mount of filesystem 7e31dbd7-b976-4d4a-a2e9-e2baacf4ad38 Dec 16 12:47:07.904792 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Dec 16 12:47:07.918154 kernel: BTRFS info (device sda6): enabling ssd optimizations Dec 16 12:47:07.918208 kernel: BTRFS info (device sda6): turning on async discard Dec 16 12:47:07.921880 kernel: BTRFS info (device sda6): enabling free space tree Dec 16 12:47:07.927866 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Dec 16 12:47:07.959272 ignition[1044]: INFO : Ignition 2.24.0 Dec 16 12:47:07.959272 ignition[1044]: INFO : Stage: files Dec 16 12:47:07.961097 ignition[1044]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 16 12:47:07.961097 ignition[1044]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 16 12:47:07.961097 ignition[1044]: DEBUG : files: compiled without relabeling support, skipping Dec 16 12:47:07.964300 ignition[1044]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Dec 16 12:47:07.964300 ignition[1044]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Dec 16 12:47:07.966753 ignition[1044]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Dec 16 12:47:07.966753 ignition[1044]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Dec 16 12:47:07.969268 ignition[1044]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Dec 16 12:47:07.967606 unknown[1044]: wrote ssh authorized keys file for user: core Dec 16 12:47:07.971384 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Dec 16 12:47:07.971384 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Dec 16 12:47:08.332243 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Dec 16 12:47:08.633743 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Dec 16 12:47:08.633743 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Dec 16 12:47:08.636966 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Dec 16 12:47:08.636966 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Dec 16 12:47:08.636966 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Dec 16 12:47:08.636966 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Dec 16 12:47:08.636966 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Dec 16 12:47:08.636966 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Dec 16 12:47:08.636966 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Dec 16 12:47:08.636966 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Dec 16 12:47:08.636966 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Dec 16 12:47:08.636966 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Dec 16 12:47:08.646665 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Dec 16 12:47:08.646665 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Dec 16 12:47:08.646665 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Dec 16 12:47:09.195412 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Dec 16 12:47:09.430281 ignition[1044]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Dec 16 12:47:09.430281 ignition[1044]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Dec 16 12:47:09.433249 ignition[1044]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Dec 16 12:47:09.435112 ignition[1044]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Dec 16 12:47:09.435112 ignition[1044]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Dec 16 12:47:09.435112 ignition[1044]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Dec 16 12:47:09.439107 ignition[1044]: INFO : files: op(d): op(e): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Dec 16 12:47:09.439107 ignition[1044]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Dec 16 12:47:09.439107 ignition[1044]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Dec 16 12:47:09.439107 ignition[1044]: INFO : files: op(f): [started] setting preset to enabled for "prepare-helm.service" Dec 16 12:47:09.439107 ignition[1044]: INFO : files: op(f): [finished] setting preset to enabled for "prepare-helm.service" Dec 16 12:47:09.439107 ignition[1044]: INFO : files: createResultFile: createFiles: op(10): [started] writing file "/sysroot/etc/.ignition-result.json" Dec 16 12:47:09.439107 ignition[1044]: INFO : files: createResultFile: createFiles: op(10): [finished] writing file "/sysroot/etc/.ignition-result.json" Dec 16 12:47:09.439107 ignition[1044]: INFO : files: files passed Dec 16 12:47:09.439107 ignition[1044]: INFO : Ignition finished successfully Dec 16 12:47:09.439000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.438350 systemd[1]: Finished ignition-files.service - Ignition (files). Dec 16 12:47:09.456000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.456000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.442128 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Dec 16 12:47:09.444421 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Dec 16 12:47:09.453727 systemd[1]: ignition-quench.service: Deactivated successfully. Dec 16 12:47:09.453792 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Dec 16 12:47:09.464476 initrd-setup-root-after-ignition[1075]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 16 12:47:09.464476 initrd-setup-root-after-ignition[1075]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Dec 16 12:47:09.466887 initrd-setup-root-after-ignition[1079]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 16 12:47:09.467897 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 16 12:47:09.468000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.469330 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Dec 16 12:47:09.471286 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Dec 16 12:47:09.516592 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Dec 16 12:47:09.516688 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Dec 16 12:47:09.518000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.518000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.518683 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Dec 16 12:47:09.520228 systemd[1]: Reached target initrd.target - Initrd Default Target. Dec 16 12:47:09.522977 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Dec 16 12:47:09.525125 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Dec 16 12:47:09.546450 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 16 12:47:09.547000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.549079 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Dec 16 12:47:09.579543 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Dec 16 12:47:09.579856 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Dec 16 12:47:09.584288 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 16 12:47:09.587773 systemd[1]: Stopped target timers.target - Timer Units. Dec 16 12:47:09.591116 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Dec 16 12:47:09.594000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.591408 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Dec 16 12:47:09.595114 systemd[1]: Stopped target initrd.target - Initrd Default Target. Dec 16 12:47:09.597402 systemd[1]: Stopped target basic.target - Basic System. Dec 16 12:47:09.600697 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Dec 16 12:47:09.603460 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Dec 16 12:47:09.620153 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Dec 16 12:47:09.623568 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Dec 16 12:47:09.627059 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Dec 16 12:47:09.630220 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Dec 16 12:47:09.633312 systemd[1]: Stopped target sysinit.target - System Initialization. Dec 16 12:47:09.636248 systemd[1]: Stopped target local-fs.target - Local File Systems. Dec 16 12:47:09.639193 systemd[1]: Stopped target swap.target - Swaps. Dec 16 12:47:09.644000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.641933 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Dec 16 12:47:09.642264 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Dec 16 12:47:09.645725 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Dec 16 12:47:09.647814 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 16 12:47:09.658000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.651168 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Dec 16 12:47:09.661000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.652009 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 16 12:47:09.664000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.654809 systemd[1]: dracut-initqueue.service: Deactivated successfully. Dec 16 12:47:09.668000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-metadata-hostname comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.655368 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Dec 16 12:47:09.659451 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Dec 16 12:47:09.659680 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Dec 16 12:47:09.661880 systemd[1]: ignition-files.service: Deactivated successfully. Dec 16 12:47:09.662304 systemd[1]: Stopped ignition-files.service - Ignition (files). Dec 16 12:47:09.664835 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Dec 16 12:47:09.665234 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Dec 16 12:47:09.671472 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Dec 16 12:47:09.687000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.678432 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Dec 16 12:47:09.692000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.683287 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Dec 16 12:47:09.696000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.683523 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 16 12:47:09.687923 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Dec 16 12:47:09.688220 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Dec 16 12:47:09.693126 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Dec 16 12:47:09.693300 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Dec 16 12:47:09.720000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.720000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.712127 systemd[1]: initrd-cleanup.service: Deactivated successfully. Dec 16 12:47:09.712264 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Dec 16 12:47:09.730904 systemd[1]: sysroot-boot.mount: Deactivated successfully. Dec 16 12:47:09.736005 systemd[1]: sysroot-boot.service: Deactivated successfully. Dec 16 12:47:09.738437 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Dec 16 12:47:09.744000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.745579 ignition[1099]: INFO : Ignition 2.24.0 Dec 16 12:47:09.745579 ignition[1099]: INFO : Stage: umount Dec 16 12:47:09.745579 ignition[1099]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 16 12:47:09.745579 ignition[1099]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Dec 16 12:47:09.749000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.750276 ignition[1099]: INFO : umount: umount passed Dec 16 12:47:09.750276 ignition[1099]: INFO : Ignition finished successfully Dec 16 12:47:09.751000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.752000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.747426 systemd[1]: ignition-mount.service: Deactivated successfully. Dec 16 12:47:09.754000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.747538 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Dec 16 12:47:09.749409 systemd[1]: ignition-disks.service: Deactivated successfully. Dec 16 12:47:09.758000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.749475 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Dec 16 12:47:09.751327 systemd[1]: ignition-kargs.service: Deactivated successfully. Dec 16 12:47:09.751393 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Dec 16 12:47:09.753061 systemd[1]: ignition-fetch.service: Deactivated successfully. Dec 16 12:47:09.753126 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Dec 16 12:47:09.754874 systemd[1]: Stopped target network.target - Network. Dec 16 12:47:09.756555 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Dec 16 12:47:09.756622 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Dec 16 12:47:09.779000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.758756 systemd[1]: Stopped target paths.target - Path Units. Dec 16 12:47:09.781000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup-pre comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.760908 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Dec 16 12:47:09.784000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.761554 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 16 12:47:09.763752 systemd[1]: Stopped target slices.target - Slice Units. Dec 16 12:47:09.766419 systemd[1]: Stopped target sockets.target - Socket Units. Dec 16 12:47:09.769211 systemd[1]: iscsid.socket: Deactivated successfully. Dec 16 12:47:09.769290 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Dec 16 12:47:09.771748 systemd[1]: iscsiuio.socket: Deactivated successfully. Dec 16 12:47:09.771802 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Dec 16 12:47:09.799000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.774104 systemd[1]: systemd-journald-audit.socket: Deactivated successfully. Dec 16 12:47:09.801000 audit: BPF prog-id=6 op=UNLOAD Dec 16 12:47:09.774137 systemd[1]: Closed systemd-journald-audit.socket - Journal Audit Socket. Dec 16 12:47:09.776406 systemd[1]: ignition-setup.service: Deactivated successfully. Dec 16 12:47:09.776530 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Dec 16 12:47:09.806000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.779138 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Dec 16 12:47:09.779227 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Dec 16 12:47:09.810000 audit: BPF prog-id=9 op=UNLOAD Dec 16 12:47:09.781928 systemd[1]: initrd-setup-root.service: Deactivated successfully. Dec 16 12:47:09.782041 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Dec 16 12:47:09.784601 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Dec 16 12:47:09.787315 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Dec 16 12:47:09.797723 systemd[1]: systemd-resolved.service: Deactivated successfully. Dec 16 12:47:09.824000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.797850 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Dec 16 12:47:09.827000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.804009 systemd[1]: systemd-networkd.service: Deactivated successfully. Dec 16 12:47:09.804290 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Dec 16 12:47:09.833000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.810440 systemd[1]: Stopped target network-pre.target - Preparation for Network. Dec 16 12:47:09.812547 systemd[1]: systemd-networkd.socket: Deactivated successfully. Dec 16 12:47:09.812616 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Dec 16 12:47:09.817251 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Dec 16 12:47:09.820169 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Dec 16 12:47:09.820257 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Dec 16 12:47:09.824737 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 16 12:47:09.824822 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Dec 16 12:47:09.827747 systemd[1]: systemd-modules-load.service: Deactivated successfully. Dec 16 12:47:09.827805 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Dec 16 12:47:09.833908 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 16 12:47:09.855000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.847575 systemd[1]: systemd-udevd.service: Deactivated successfully. Dec 16 12:47:09.859000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.847711 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 16 12:47:09.855637 systemd[1]: network-cleanup.service: Deactivated successfully. Dec 16 12:47:09.855888 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Dec 16 12:47:09.865000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.860287 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Dec 16 12:47:09.867000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.860377 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Dec 16 12:47:09.869000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.861747 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Dec 16 12:47:09.861806 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Dec 16 12:47:09.863757 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Dec 16 12:47:09.877000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.863835 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Dec 16 12:47:09.866058 systemd[1]: dracut-cmdline.service: Deactivated successfully. Dec 16 12:47:09.878000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.866131 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Dec 16 12:47:09.880000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.868182 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 16 12:47:09.882000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.868232 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Dec 16 12:47:09.870859 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Dec 16 12:47:09.885000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.875475 systemd[1]: systemd-network-generator.service: Deactivated successfully. Dec 16 12:47:09.875539 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Dec 16 12:47:09.877171 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Dec 16 12:47:09.877236 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 16 12:47:09.879079 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Dec 16 12:47:09.879119 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 16 12:47:09.880799 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Dec 16 12:47:09.893000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.893000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:09.880839 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Dec 16 12:47:09.884092 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 16 12:47:09.884140 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Dec 16 12:47:09.891891 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Dec 16 12:47:09.892010 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Dec 16 12:47:09.893983 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Dec 16 12:47:09.896353 systemd[1]: Starting initrd-switch-root.service - Switch Root... Dec 16 12:47:09.910713 systemd[1]: Switching root. Dec 16 12:47:09.942091 systemd-journald[321]: Received SIGTERM from PID 1 (systemd). Dec 16 12:47:09.942161 systemd-journald[321]: Journal stopped Dec 16 12:47:10.962748 kernel: SELinux: policy capability network_peer_controls=1 Dec 16 12:47:10.962793 kernel: SELinux: policy capability open_perms=1 Dec 16 12:47:10.962803 kernel: SELinux: policy capability extended_socket_class=1 Dec 16 12:47:10.962818 kernel: SELinux: policy capability always_check_network=0 Dec 16 12:47:10.962826 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 16 12:47:10.962839 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 16 12:47:10.962848 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Dec 16 12:47:10.962861 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 16 12:47:10.962870 kernel: SELinux: policy capability userspace_initial_context=0 Dec 16 12:47:10.962881 systemd[1]: Successfully loaded SELinux policy in 76.623ms. Dec 16 12:47:10.962896 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 5.344ms. Dec 16 12:47:10.962905 systemd[1]: systemd 257.9 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Dec 16 12:47:10.962915 systemd[1]: Detected virtualization kvm. Dec 16 12:47:10.962926 systemd[1]: Detected architecture x86-64. Dec 16 12:47:10.962935 systemd[1]: Detected first boot. Dec 16 12:47:10.962945 systemd[1]: Hostname set to . Dec 16 12:47:10.962954 systemd[1]: Initializing machine ID from SMBIOS/DMI UUID. Dec 16 12:47:10.962964 zram_generator::config[1142]: No configuration found. Dec 16 12:47:10.962975 kernel: Guest personality initialized and is inactive Dec 16 12:47:10.962984 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Dec 16 12:47:10.962993 kernel: Initialized host personality Dec 16 12:47:10.963002 kernel: NET: Registered PF_VSOCK protocol family Dec 16 12:47:10.963011 systemd[1]: Populated /etc with preset unit settings. Dec 16 12:47:10.963037 kernel: kauditd_printk_skb: 57 callbacks suppressed Dec 16 12:47:10.963050 kernel: audit: type=1334 audit(1765889230.623:94): prog-id=12 op=LOAD Dec 16 12:47:10.963059 kernel: audit: type=1334 audit(1765889230.623:95): prog-id=3 op=UNLOAD Dec 16 12:47:10.963068 systemd[1]: initrd-switch-root.service: Deactivated successfully. Dec 16 12:47:10.963077 kernel: audit: type=1334 audit(1765889230.623:96): prog-id=13 op=LOAD Dec 16 12:47:10.963086 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Dec 16 12:47:10.963096 kernel: audit: type=1334 audit(1765889230.623:97): prog-id=14 op=LOAD Dec 16 12:47:10.963104 kernel: audit: type=1334 audit(1765889230.623:98): prog-id=4 op=UNLOAD Dec 16 12:47:10.963114 kernel: audit: type=1334 audit(1765889230.623:99): prog-id=5 op=UNLOAD Dec 16 12:47:10.963123 kernel: audit: type=1131 audit(1765889230.624:100): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.963132 kernel: audit: type=1334 audit(1765889230.633:101): prog-id=12 op=UNLOAD Dec 16 12:47:10.963141 kernel: audit: type=1130 audit(1765889230.648:102): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.963150 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Dec 16 12:47:10.963159 kernel: audit: type=1131 audit(1765889230.648:103): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.963171 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Dec 16 12:47:10.963180 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Dec 16 12:47:10.963190 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Dec 16 12:47:10.963199 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Dec 16 12:47:10.963210 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Dec 16 12:47:10.963220 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Dec 16 12:47:10.963231 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Dec 16 12:47:10.963239 systemd[1]: Created slice user.slice - User and Session Slice. Dec 16 12:47:10.963249 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Dec 16 12:47:10.963258 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Dec 16 12:47:10.963268 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Dec 16 12:47:10.963278 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Dec 16 12:47:10.963287 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Dec 16 12:47:10.963296 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Dec 16 12:47:10.963305 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Dec 16 12:47:10.963314 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Dec 16 12:47:10.963324 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Dec 16 12:47:10.963333 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Dec 16 12:47:10.963342 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Dec 16 12:47:10.963353 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Dec 16 12:47:10.963362 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Dec 16 12:47:10.963372 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Dec 16 12:47:10.963381 systemd[1]: Reached target remote-fs.target - Remote File Systems. Dec 16 12:47:10.963390 systemd[1]: Reached target remote-veritysetup.target - Remote Verity Protected Volumes. Dec 16 12:47:10.963399 systemd[1]: Reached target slices.target - Slice Units. Dec 16 12:47:10.963408 systemd[1]: Reached target swap.target - Swaps. Dec 16 12:47:10.963418 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Dec 16 12:47:10.963429 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Dec 16 12:47:10.963440 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Dec 16 12:47:10.963449 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Dec 16 12:47:10.963458 systemd[1]: Listening on systemd-mountfsd.socket - DDI File System Mounter Socket. Dec 16 12:47:10.963467 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Dec 16 12:47:10.963486 systemd[1]: Listening on systemd-nsresourced.socket - Namespace Resource Manager Socket. Dec 16 12:47:10.963497 systemd[1]: Listening on systemd-oomd.socket - Userspace Out-Of-Memory (OOM) Killer Socket. Dec 16 12:47:10.963507 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Dec 16 12:47:10.963516 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Dec 16 12:47:10.963525 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Dec 16 12:47:10.963534 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Dec 16 12:47:10.963543 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Dec 16 12:47:10.963553 systemd[1]: Mounting media.mount - External Media Directory... Dec 16 12:47:10.963563 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 16 12:47:10.963572 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Dec 16 12:47:10.963581 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Dec 16 12:47:10.963590 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Dec 16 12:47:10.963600 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Dec 16 12:47:10.963613 systemd[1]: Reached target machines.target - Containers. Dec 16 12:47:10.963630 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Dec 16 12:47:10.963648 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 16 12:47:10.963659 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Dec 16 12:47:10.963669 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Dec 16 12:47:10.963678 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 16 12:47:10.963687 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 16 12:47:10.963697 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 16 12:47:10.963707 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Dec 16 12:47:10.963717 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 16 12:47:10.963726 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Dec 16 12:47:10.963736 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Dec 16 12:47:10.963745 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Dec 16 12:47:10.963754 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Dec 16 12:47:10.963765 systemd[1]: Stopped systemd-fsck-usr.service. Dec 16 12:47:10.963774 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Dec 16 12:47:10.963784 systemd[1]: Starting systemd-journald.service - Journal Service... Dec 16 12:47:10.963793 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Dec 16 12:47:10.963802 kernel: fuse: init (API version 7.41) Dec 16 12:47:10.963812 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Dec 16 12:47:10.963821 kernel: ACPI: bus type drm_connector registered Dec 16 12:47:10.963830 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Dec 16 12:47:10.963840 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Dec 16 12:47:10.963849 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Dec 16 12:47:10.963859 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 16 12:47:10.963868 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Dec 16 12:47:10.963878 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Dec 16 12:47:10.963888 systemd[1]: Mounted media.mount - External Media Directory. Dec 16 12:47:10.963898 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Dec 16 12:47:10.963922 systemd-journald[1221]: Collecting audit messages is enabled. Dec 16 12:47:10.963944 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Dec 16 12:47:10.963955 systemd-journald[1221]: Journal started Dec 16 12:47:10.963974 systemd-journald[1221]: Runtime Journal (/run/log/journal/afbec2ca433d48b69d29e348d9cb3439) is 4.7M, max 38.1M, 33.4M free. Dec 16 12:47:10.747000 audit[1]: EVENT_LISTENER pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Dec 16 12:47:10.872000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.878000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.886000 audit: BPF prog-id=14 op=UNLOAD Dec 16 12:47:10.886000 audit: BPF prog-id=13 op=UNLOAD Dec 16 12:47:10.887000 audit: BPF prog-id=15 op=LOAD Dec 16 12:47:10.887000 audit: BPF prog-id=16 op=LOAD Dec 16 12:47:10.887000 audit: BPF prog-id=17 op=LOAD Dec 16 12:47:10.961000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Dec 16 12:47:10.961000 audit[1221]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=6 a1=7ffc17881620 a2=4000 a3=0 items=0 ppid=1 pid=1221 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:10.961000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Dec 16 12:47:10.608222 systemd[1]: Queued start job for default target multi-user.target. Dec 16 12:47:10.623821 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Dec 16 12:47:10.624152 systemd[1]: systemd-journald.service: Deactivated successfully. Dec 16 12:47:10.970149 systemd[1]: Started systemd-journald.service - Journal Service. Dec 16 12:47:10.970000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.971090 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Dec 16 12:47:10.971957 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Dec 16 12:47:10.972000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.972925 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Dec 16 12:47:10.973000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.973892 systemd[1]: modprobe@configfs.service: Deactivated successfully. Dec 16 12:47:10.974087 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Dec 16 12:47:10.974000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.974000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.975169 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 16 12:47:10.975352 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 16 12:47:10.975000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.976000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.976297 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 16 12:47:10.976470 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 16 12:47:10.977000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.977000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.977505 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 16 12:47:10.977665 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 16 12:47:10.978000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.978000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.978632 systemd[1]: modprobe@fuse.service: Deactivated successfully. Dec 16 12:47:10.978791 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Dec 16 12:47:10.979000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.979000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.979847 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 16 12:47:10.980055 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 16 12:47:10.980000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.980000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.981163 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Dec 16 12:47:10.981000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.982164 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Dec 16 12:47:10.982000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.983862 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Dec 16 12:47:10.984000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.985051 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Dec 16 12:47:10.985000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-load-credentials comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:10.991831 systemd[1]: Reached target network-pre.target - Preparation for Network. Dec 16 12:47:10.993148 systemd[1]: Listening on systemd-importd.socket - Disk Image Download Service Socket. Dec 16 12:47:10.995102 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Dec 16 12:47:10.998087 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Dec 16 12:47:10.998733 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Dec 16 12:47:10.998756 systemd[1]: Reached target local-fs.target - Local File Systems. Dec 16 12:47:11.000234 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Dec 16 12:47:11.001448 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 16 12:47:11.001551 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Dec 16 12:47:11.004106 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Dec 16 12:47:11.011105 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Dec 16 12:47:11.012183 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 16 12:47:11.015757 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Dec 16 12:47:11.016467 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 16 12:47:11.018630 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Dec 16 12:47:11.021388 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Dec 16 12:47:11.025463 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Dec 16 12:47:11.028358 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Dec 16 12:47:11.030133 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Dec 16 12:47:11.034373 systemd-journald[1221]: Time spent on flushing to /var/log/journal/afbec2ca433d48b69d29e348d9cb3439 is 20.513ms for 1304 entries. Dec 16 12:47:11.034373 systemd-journald[1221]: System Journal (/var/log/journal/afbec2ca433d48b69d29e348d9cb3439) is 8M, max 588.1M, 580.1M free. Dec 16 12:47:11.074317 systemd-journald[1221]: Received client request to flush runtime journal. Dec 16 12:47:11.074349 kernel: loop1: detected capacity change from 0 to 50784 Dec 16 12:47:11.054000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.057000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.058000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.054226 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Dec 16 12:47:11.055342 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Dec 16 12:47:11.057445 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Dec 16 12:47:11.059927 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Dec 16 12:47:11.062646 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Dec 16 12:47:11.079000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.079004 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Dec 16 12:47:11.086241 systemd-tmpfiles[1269]: ACLs are not supported, ignoring. Dec 16 12:47:11.086255 systemd-tmpfiles[1269]: ACLs are not supported, ignoring. Dec 16 12:47:11.092598 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Dec 16 12:47:11.094000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.099490 kernel: loop2: detected capacity change from 0 to 111560 Dec 16 12:47:11.098508 systemd[1]: Starting systemd-sysusers.service - Create System Users... Dec 16 12:47:11.106000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.105918 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Dec 16 12:47:11.129935 kernel: loop3: detected capacity change from 0 to 224512 Dec 16 12:47:11.129000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.129000 audit: BPF prog-id=18 op=LOAD Dec 16 12:47:11.128341 systemd[1]: Finished systemd-sysusers.service - Create System Users. Dec 16 12:47:11.130000 audit: BPF prog-id=19 op=LOAD Dec 16 12:47:11.130000 audit: BPF prog-id=20 op=LOAD Dec 16 12:47:11.134000 audit: BPF prog-id=21 op=LOAD Dec 16 12:47:11.133126 systemd[1]: Starting systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer... Dec 16 12:47:11.135419 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Dec 16 12:47:11.136977 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Dec 16 12:47:11.146000 audit: BPF prog-id=22 op=LOAD Dec 16 12:47:11.146000 audit: BPF prog-id=23 op=LOAD Dec 16 12:47:11.146000 audit: BPF prog-id=24 op=LOAD Dec 16 12:47:11.149000 audit: BPF prog-id=25 op=LOAD Dec 16 12:47:11.149000 audit: BPF prog-id=26 op=LOAD Dec 16 12:47:11.149000 audit: BPF prog-id=27 op=LOAD Dec 16 12:47:11.148122 systemd[1]: Starting systemd-nsresourced.service - Namespace Resource Manager... Dec 16 12:47:11.151206 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Dec 16 12:47:11.164134 systemd-tmpfiles[1290]: ACLs are not supported, ignoring. Dec 16 12:47:11.164147 systemd-tmpfiles[1290]: ACLs are not supported, ignoring. Dec 16 12:47:11.169000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.172084 kernel: loop4: detected capacity change from 0 to 8 Dec 16 12:47:11.169168 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Dec 16 12:47:11.200055 kernel: loop5: detected capacity change from 0 to 50784 Dec 16 12:47:11.217085 systemd-nsresourced[1292]: Not setting up BPF subsystem, as functionality has been disabled at compile time. Dec 16 12:47:11.219311 kernel: loop6: detected capacity change from 0 to 111560 Dec 16 12:47:11.218385 systemd[1]: Started systemd-nsresourced.service - Namespace Resource Manager. Dec 16 12:47:11.220000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-nsresourced comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.221505 systemd[1]: Started systemd-userdbd.service - User Database Manager. Dec 16 12:47:11.223000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.239049 kernel: loop7: detected capacity change from 0 to 224512 Dec 16 12:47:11.265086 kernel: loop1: detected capacity change from 0 to 8 Dec 16 12:47:11.268662 (sd-merge)[1297]: Using extensions 'containerd-flatcar.raw', 'docker-flatcar.raw', 'kubernetes.raw', 'oem-hetzner.raw'. Dec 16 12:47:11.275835 (sd-merge)[1297]: Merged extensions into '/usr'. Dec 16 12:47:11.287129 systemd[1]: Reload requested from client PID 1268 ('systemd-sysext') (unit systemd-sysext.service)... Dec 16 12:47:11.287140 systemd[1]: Reloading... Dec 16 12:47:11.331486 systemd-oomd[1288]: No swap; memory pressure usage will be degraded Dec 16 12:47:11.362176 systemd-resolved[1289]: Positive Trust Anchors: Dec 16 12:47:11.362429 systemd-resolved[1289]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 16 12:47:11.362488 systemd-resolved[1289]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Dec 16 12:47:11.362550 systemd-resolved[1289]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Dec 16 12:47:11.365059 zram_generator::config[1347]: No configuration found. Dec 16 12:47:11.378263 systemd-resolved[1289]: Using system hostname 'ci-4547-0-0-6-b8319fed62'. Dec 16 12:47:11.539011 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Dec 16 12:47:11.539256 systemd[1]: Reloading finished in 251 ms. Dec 16 12:47:11.565945 systemd[1]: Started systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer. Dec 16 12:47:11.566000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-oomd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.566925 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Dec 16 12:47:11.567000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.567949 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Dec 16 12:47:11.568000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.568994 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Dec 16 12:47:11.569000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.572677 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Dec 16 12:47:11.584111 systemd[1]: Starting ensure-sysext.service... Dec 16 12:47:11.587727 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Dec 16 12:47:11.588000 audit: BPF prog-id=8 op=UNLOAD Dec 16 12:47:11.588000 audit: BPF prog-id=7 op=UNLOAD Dec 16 12:47:11.591000 audit: BPF prog-id=28 op=LOAD Dec 16 12:47:11.591000 audit: BPF prog-id=29 op=LOAD Dec 16 12:47:11.594000 audit: BPF prog-id=30 op=LOAD Dec 16 12:47:11.594000 audit: BPF prog-id=25 op=UNLOAD Dec 16 12:47:11.594000 audit: BPF prog-id=31 op=LOAD Dec 16 12:47:11.594000 audit: BPF prog-id=32 op=LOAD Dec 16 12:47:11.594000 audit: BPF prog-id=26 op=UNLOAD Dec 16 12:47:11.594000 audit: BPF prog-id=27 op=UNLOAD Dec 16 12:47:11.596000 audit: BPF prog-id=33 op=LOAD Dec 16 12:47:11.593509 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Dec 16 12:47:11.597000 audit: BPF prog-id=18 op=UNLOAD Dec 16 12:47:11.597000 audit: BPF prog-id=34 op=LOAD Dec 16 12:47:11.597000 audit: BPF prog-id=35 op=LOAD Dec 16 12:47:11.597000 audit: BPF prog-id=19 op=UNLOAD Dec 16 12:47:11.597000 audit: BPF prog-id=20 op=UNLOAD Dec 16 12:47:11.597000 audit: BPF prog-id=36 op=LOAD Dec 16 12:47:11.597000 audit: BPF prog-id=21 op=UNLOAD Dec 16 12:47:11.599000 audit: BPF prog-id=37 op=LOAD Dec 16 12:47:11.599000 audit: BPF prog-id=22 op=UNLOAD Dec 16 12:47:11.599000 audit: BPF prog-id=38 op=LOAD Dec 16 12:47:11.599000 audit: BPF prog-id=39 op=LOAD Dec 16 12:47:11.599000 audit: BPF prog-id=23 op=UNLOAD Dec 16 12:47:11.599000 audit: BPF prog-id=24 op=UNLOAD Dec 16 12:47:11.599000 audit: BPF prog-id=40 op=LOAD Dec 16 12:47:11.599000 audit: BPF prog-id=15 op=UNLOAD Dec 16 12:47:11.599000 audit: BPF prog-id=41 op=LOAD Dec 16 12:47:11.599000 audit: BPF prog-id=42 op=LOAD Dec 16 12:47:11.599000 audit: BPF prog-id=16 op=UNLOAD Dec 16 12:47:11.599000 audit: BPF prog-id=17 op=UNLOAD Dec 16 12:47:11.605360 systemd-tmpfiles[1385]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Dec 16 12:47:11.605389 systemd[1]: Reload requested from client PID 1384 ('systemctl') (unit ensure-sysext.service)... Dec 16 12:47:11.605389 systemd-tmpfiles[1385]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Dec 16 12:47:11.605401 systemd[1]: Reloading... Dec 16 12:47:11.605650 systemd-tmpfiles[1385]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Dec 16 12:47:11.606622 systemd-tmpfiles[1385]: ACLs are not supported, ignoring. Dec 16 12:47:11.606669 systemd-tmpfiles[1385]: ACLs are not supported, ignoring. Dec 16 12:47:11.613054 systemd-tmpfiles[1385]: Detected autofs mount point /boot during canonicalization of boot. Dec 16 12:47:11.613063 systemd-tmpfiles[1385]: Skipping /boot Dec 16 12:47:11.620223 systemd-tmpfiles[1385]: Detected autofs mount point /boot during canonicalization of boot. Dec 16 12:47:11.620288 systemd-tmpfiles[1385]: Skipping /boot Dec 16 12:47:11.639974 systemd-udevd[1386]: Using default interface naming scheme 'v257'. Dec 16 12:47:11.677052 zram_generator::config[1417]: No configuration found. Dec 16 12:47:11.797062 kernel: mousedev: PS/2 mouse device common for all mice Dec 16 12:47:11.815333 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input5 Dec 16 12:47:11.820046 kernel: ACPI: button: Power Button [PWRF] Dec 16 12:47:11.879317 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Dec 16 12:47:11.879628 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Dec 16 12:47:11.913037 kernel: EDAC MC: Ver: 3.0.0 Dec 16 12:47:11.940035 kernel: [drm] pci: virtio-vga detected at 0000:00:01.0 Dec 16 12:47:11.940081 kernel: virtio-pci 0000:00:01.0: vgaarb: deactivate vga console Dec 16 12:47:11.946041 kernel: Console: switching to colour dummy device 80x25 Dec 16 12:47:11.950030 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Dec 16 12:47:11.950070 kernel: [drm] features: -context_init Dec 16 12:47:11.950093 kernel: [drm] number of scanouts: 1 Dec 16 12:47:11.951514 kernel: [drm] number of cap sets: 0 Dec 16 12:47:11.957029 kernel: [drm] Initialized virtio_gpu 0.1.0 for 0000:00:01.0 on minor 0 Dec 16 12:47:11.960127 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Dec 16 12:47:11.961270 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Dec 16 12:47:11.961306 kernel: Console: switching to colour frame buffer device 160x50 Dec 16 12:47:11.970315 kernel: virtio-pci 0000:00:01.0: [drm] fb0: virtio_gpudrmfb frame buffer device Dec 16 12:47:11.973796 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Dec 16 12:47:11.974539 systemd[1]: Reloading finished in 368 ms. Dec 16 12:47:11.985000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.986000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:11.983710 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Dec 16 12:47:11.985667 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Dec 16 12:47:11.992000 audit: BPF prog-id=43 op=LOAD Dec 16 12:47:11.992000 audit: BPF prog-id=33 op=UNLOAD Dec 16 12:47:11.992000 audit: BPF prog-id=44 op=LOAD Dec 16 12:47:11.992000 audit: BPF prog-id=45 op=LOAD Dec 16 12:47:11.992000 audit: BPF prog-id=34 op=UNLOAD Dec 16 12:47:11.992000 audit: BPF prog-id=35 op=UNLOAD Dec 16 12:47:11.992000 audit: BPF prog-id=46 op=LOAD Dec 16 12:47:11.992000 audit: BPF prog-id=37 op=UNLOAD Dec 16 12:47:11.992000 audit: BPF prog-id=47 op=LOAD Dec 16 12:47:11.992000 audit: BPF prog-id=48 op=LOAD Dec 16 12:47:11.992000 audit: BPF prog-id=38 op=UNLOAD Dec 16 12:47:11.992000 audit: BPF prog-id=39 op=UNLOAD Dec 16 12:47:11.993000 audit: BPF prog-id=49 op=LOAD Dec 16 12:47:11.993000 audit: BPF prog-id=36 op=UNLOAD Dec 16 12:47:11.994000 audit: BPF prog-id=50 op=LOAD Dec 16 12:47:11.994000 audit: BPF prog-id=51 op=LOAD Dec 16 12:47:11.994000 audit: BPF prog-id=28 op=UNLOAD Dec 16 12:47:11.994000 audit: BPF prog-id=29 op=UNLOAD Dec 16 12:47:11.998000 audit: BPF prog-id=52 op=LOAD Dec 16 12:47:11.998000 audit: BPF prog-id=30 op=UNLOAD Dec 16 12:47:11.998000 audit: BPF prog-id=53 op=LOAD Dec 16 12:47:11.998000 audit: BPF prog-id=54 op=LOAD Dec 16 12:47:11.998000 audit: BPF prog-id=31 op=UNLOAD Dec 16 12:47:11.998000 audit: BPF prog-id=32 op=UNLOAD Dec 16 12:47:11.999000 audit: BPF prog-id=55 op=LOAD Dec 16 12:47:11.999000 audit: BPF prog-id=40 op=UNLOAD Dec 16 12:47:11.999000 audit: BPF prog-id=56 op=LOAD Dec 16 12:47:11.999000 audit: BPF prog-id=57 op=LOAD Dec 16 12:47:11.999000 audit: BPF prog-id=41 op=UNLOAD Dec 16 12:47:11.999000 audit: BPF prog-id=42 op=UNLOAD Dec 16 12:47:12.064000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ensure-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:12.062222 systemd[1]: Finished ensure-sysext.service. Dec 16 12:47:12.077851 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 16 12:47:12.078732 systemd[1]: Starting audit-rules.service - Load Audit Rules... Dec 16 12:47:12.092942 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Dec 16 12:47:12.093268 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Dec 16 12:47:12.095074 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Dec 16 12:47:12.098110 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Dec 16 12:47:12.100085 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Dec 16 12:47:12.102434 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Dec 16 12:47:12.104128 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Dec 16 12:47:12.104218 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Dec 16 12:47:12.106909 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Dec 16 12:47:12.108163 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Dec 16 12:47:12.109112 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Dec 16 12:47:12.111127 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Dec 16 12:47:12.113000 audit: BPF prog-id=58 op=LOAD Dec 16 12:47:12.116177 systemd[1]: Starting systemd-networkd.service - Network Configuration... Dec 16 12:47:12.127000 audit: BPF prog-id=59 op=LOAD Dec 16 12:47:12.128649 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Dec 16 12:47:12.133219 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Dec 16 12:47:12.137419 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Dec 16 12:47:12.138175 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 16 12:47:12.140249 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 16 12:47:12.141412 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Dec 16 12:47:12.142000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:12.142000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:12.142799 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 16 12:47:12.143764 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Dec 16 12:47:12.143000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:12.143000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:12.146172 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Dec 16 12:47:12.155582 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 16 12:47:12.160532 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Dec 16 12:47:12.163000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:12.163000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:12.164375 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 16 12:47:12.164000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:12.164755 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Dec 16 12:47:12.165000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:12.168312 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 16 12:47:12.184000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:12.184098 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Dec 16 12:47:12.187000 audit[1532]: SYSTEM_BOOT pid=1532 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Dec 16 12:47:12.191000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:12.190130 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Dec 16 12:47:12.215369 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Dec 16 12:47:12.217000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:12.220000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Dec 16 12:47:12.220000 audit[1558]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7fffa426c970 a2=420 a3=0 items=0 ppid=1512 pid=1558 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:12.220000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Dec 16 12:47:12.220790 augenrules[1558]: No rules Dec 16 12:47:12.221805 systemd[1]: audit-rules.service: Deactivated successfully. Dec 16 12:47:12.222420 systemd[1]: Finished audit-rules.service - Load Audit Rules. Dec 16 12:47:12.238414 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Dec 16 12:47:12.239774 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 16 12:47:12.263672 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Dec 16 12:47:12.269507 systemd-networkd[1527]: lo: Link UP Dec 16 12:47:12.269515 systemd-networkd[1527]: lo: Gained carrier Dec 16 12:47:12.270074 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Dec 16 12:47:12.270523 systemd[1]: Reached target time-set.target - System Time Set. Dec 16 12:47:12.272529 systemd[1]: Started systemd-networkd.service - Network Configuration. Dec 16 12:47:12.273610 systemd[1]: Reached target network.target - Network. Dec 16 12:47:12.274356 systemd-networkd[1527]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Dec 16 12:47:12.274363 systemd-networkd[1527]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 16 12:47:12.275919 systemd-networkd[1527]: eth1: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Dec 16 12:47:12.276083 systemd-networkd[1527]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 16 12:47:12.276105 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Dec 16 12:47:12.278443 systemd-networkd[1527]: eth0: Link UP Dec 16 12:47:12.278618 systemd-networkd[1527]: eth0: Gained carrier Dec 16 12:47:12.278630 systemd-networkd[1527]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Dec 16 12:47:12.279643 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Dec 16 12:47:12.283629 systemd-networkd[1527]: eth1: Link UP Dec 16 12:47:12.284616 systemd-networkd[1527]: eth1: Gained carrier Dec 16 12:47:12.284740 systemd-networkd[1527]: eth1: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Dec 16 12:47:12.303126 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Dec 16 12:47:12.314067 systemd-networkd[1527]: eth1: DHCPv4 address 10.0.0.3/32 acquired from 10.0.0.1 Dec 16 12:47:12.314864 systemd-timesyncd[1531]: Network configuration changed, trying to establish connection. Dec 16 12:47:12.339060 systemd-networkd[1527]: eth0: DHCPv4 address 77.42.19.41/32, gateway 172.31.1.1 acquired from 172.31.1.1 Dec 16 12:47:12.339686 systemd-timesyncd[1531]: Network configuration changed, trying to establish connection. Dec 16 12:47:12.762246 ldconfig[1524]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Dec 16 12:47:12.766238 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Dec 16 12:47:12.771152 systemd[1]: Starting systemd-update-done.service - Update is Completed... Dec 16 12:47:12.782184 systemd[1]: Finished systemd-update-done.service - Update is Completed. Dec 16 12:47:12.783128 systemd[1]: Reached target sysinit.target - System Initialization. Dec 16 12:47:12.785879 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Dec 16 12:47:12.786762 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Dec 16 12:47:12.787583 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Dec 16 12:47:12.788572 systemd[1]: Started logrotate.timer - Daily rotation of log files. Dec 16 12:47:12.789520 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Dec 16 12:47:12.790328 systemd[1]: Started systemd-sysupdate-reboot.timer - Reboot Automatically After System Update. Dec 16 12:47:12.791267 systemd[1]: Started systemd-sysupdate.timer - Automatic System Update. Dec 16 12:47:12.792159 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Dec 16 12:47:12.793284 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Dec 16 12:47:12.793414 systemd[1]: Reached target paths.target - Path Units. Dec 16 12:47:12.794509 systemd[1]: Reached target timers.target - Timer Units. Dec 16 12:47:12.797703 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Dec 16 12:47:12.799558 systemd[1]: Starting docker.socket - Docker Socket for the API... Dec 16 12:47:12.803858 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Dec 16 12:47:12.806793 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Dec 16 12:47:12.808976 systemd[1]: Reached target ssh-access.target - SSH Access Available. Dec 16 12:47:12.815513 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Dec 16 12:47:12.816613 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Dec 16 12:47:12.819886 systemd[1]: Listening on docker.socket - Docker Socket for the API. Dec 16 12:47:12.821372 systemd[1]: Reached target sockets.target - Socket Units. Dec 16 12:47:12.824997 systemd[1]: Reached target basic.target - Basic System. Dec 16 12:47:12.825726 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Dec 16 12:47:12.825755 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Dec 16 12:47:12.826701 systemd[1]: Starting containerd.service - containerd container runtime... Dec 16 12:47:12.831333 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Dec 16 12:47:12.839305 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Dec 16 12:47:12.846327 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Dec 16 12:47:12.854171 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Dec 16 12:47:12.859561 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Dec 16 12:47:12.861372 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Dec 16 12:47:12.864956 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Dec 16 12:47:12.874312 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Dec 16 12:47:12.879384 coreos-metadata[1578]: Dec 16 12:47:12.879 INFO Fetching http://169.254.169.254/hetzner/v1/metadata: Attempt #1 Dec 16 12:47:12.882154 coreos-metadata[1578]: Dec 16 12:47:12.881 INFO Fetch successful Dec 16 12:47:12.883109 coreos-metadata[1578]: Dec 16 12:47:12.883 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/private-networks: Attempt #1 Dec 16 12:47:12.883309 coreos-metadata[1578]: Dec 16 12:47:12.883 INFO Fetch successful Dec 16 12:47:12.886274 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Dec 16 12:47:12.889152 systemd[1]: Started qemu-guest-agent.service - QEMU Guest Agent. Dec 16 12:47:12.898981 google_oslogin_nss_cache[1585]: oslogin_cache_refresh[1585]: Refreshing passwd entry cache Dec 16 12:47:12.898071 oslogin_cache_refresh[1585]: Refreshing passwd entry cache Dec 16 12:47:12.899354 jq[1583]: false Dec 16 12:47:12.900527 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Dec 16 12:47:12.902180 google_oslogin_nss_cache[1585]: oslogin_cache_refresh[1585]: Failure getting users, quitting Dec 16 12:47:12.902235 oslogin_cache_refresh[1585]: Failure getting users, quitting Dec 16 12:47:12.902299 google_oslogin_nss_cache[1585]: oslogin_cache_refresh[1585]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Dec 16 12:47:12.902328 oslogin_cache_refresh[1585]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Dec 16 12:47:12.902408 google_oslogin_nss_cache[1585]: oslogin_cache_refresh[1585]: Refreshing group entry cache Dec 16 12:47:12.902440 oslogin_cache_refresh[1585]: Refreshing group entry cache Dec 16 12:47:12.906571 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Dec 16 12:47:12.912581 systemd[1]: Starting systemd-logind.service - User Login Management... Dec 16 12:47:12.914700 google_oslogin_nss_cache[1585]: oslogin_cache_refresh[1585]: Failure getting groups, quitting Dec 16 12:47:12.914698 oslogin_cache_refresh[1585]: Failure getting groups, quitting Dec 16 12:47:12.914764 google_oslogin_nss_cache[1585]: oslogin_cache_refresh[1585]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Dec 16 12:47:12.914708 oslogin_cache_refresh[1585]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Dec 16 12:47:12.915270 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Dec 16 12:47:12.915769 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Dec 16 12:47:12.916922 systemd[1]: Starting update-engine.service - Update Engine... Dec 16 12:47:12.919079 extend-filesystems[1584]: Found /dev/sda6 Dec 16 12:47:12.926218 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Dec 16 12:47:12.929753 extend-filesystems[1584]: Found /dev/sda9 Dec 16 12:47:12.933675 extend-filesystems[1584]: Checking size of /dev/sda9 Dec 16 12:47:12.946216 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Dec 16 12:47:12.947152 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Dec 16 12:47:12.947321 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Dec 16 12:47:12.947528 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Dec 16 12:47:12.947666 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Dec 16 12:47:12.949854 extend-filesystems[1584]: Resized partition /dev/sda9 Dec 16 12:47:12.959770 jq[1604]: true Dec 16 12:47:12.952618 systemd[1]: motdgen.service: Deactivated successfully. Dec 16 12:47:12.955105 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Dec 16 12:47:12.960897 extend-filesystems[1615]: resize2fs 1.47.3 (8-Jul-2025) Dec 16 12:47:12.971042 update_engine[1602]: I20251216 12:47:12.964229 1602 main.cc:92] Flatcar Update Engine starting Dec 16 12:47:12.965281 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Dec 16 12:47:12.965877 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Dec 16 12:47:12.977267 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 8410107 blocks Dec 16 12:47:12.989320 jq[1621]: true Dec 16 12:47:13.004000 tar[1617]: linux-amd64/LICENSE Dec 16 12:47:13.004729 tar[1617]: linux-amd64/helm Dec 16 12:47:13.064253 dbus-daemon[1579]: [system] SELinux support is enabled Dec 16 12:47:13.064648 systemd[1]: Started dbus.service - D-Bus System Message Bus. Dec 16 12:47:13.068585 systemd-logind[1601]: New seat seat0. Dec 16 12:47:13.071125 systemd-logind[1601]: Watching system buttons on /dev/input/event3 (Power Button) Dec 16 12:47:13.071139 systemd-logind[1601]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Dec 16 12:47:13.071812 systemd[1]: Started systemd-logind.service - User Login Management. Dec 16 12:47:13.079998 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Dec 16 12:47:13.083645 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Dec 16 12:47:13.087183 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Dec 16 12:47:13.087199 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Dec 16 12:47:13.098080 update_engine[1602]: I20251216 12:47:13.096524 1602 update_check_scheduler.cc:74] Next update check in 6m20s Dec 16 12:47:13.102435 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Dec 16 12:47:13.117966 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Dec 16 12:47:13.122485 dbus-daemon[1579]: [system] Successfully activated service 'org.freedesktop.systemd1' Dec 16 12:47:13.119568 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Dec 16 12:47:13.119727 systemd[1]: Started update-engine.service - Update Engine. Dec 16 12:47:13.122816 systemd[1]: Started locksmithd.service - Cluster reboot manager. Dec 16 12:47:13.174665 bash[1660]: Updated "/home/core/.ssh/authorized_keys" Dec 16 12:47:13.176679 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Dec 16 12:47:13.188037 kernel: EXT4-fs (sda9): resized filesystem to 8410107 Dec 16 12:47:13.191220 systemd[1]: Starting sshkeys.service... Dec 16 12:47:13.210494 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Dec 16 12:47:13.216452 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Dec 16 12:47:13.223001 extend-filesystems[1615]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Dec 16 12:47:13.223001 extend-filesystems[1615]: old_desc_blocks = 1, new_desc_blocks = 5 Dec 16 12:47:13.223001 extend-filesystems[1615]: The filesystem on /dev/sda9 is now 8410107 (4k) blocks long. Dec 16 12:47:13.240772 extend-filesystems[1584]: Resized filesystem in /dev/sda9 Dec 16 12:47:13.230675 systemd[1]: extend-filesystems.service: Deactivated successfully. Dec 16 12:47:13.230859 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Dec 16 12:47:13.284116 coreos-metadata[1671]: Dec 16 12:47:13.283 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/public-keys: Attempt #1 Dec 16 12:47:13.285699 coreos-metadata[1671]: Dec 16 12:47:13.285 INFO Fetch successful Dec 16 12:47:13.287465 unknown[1671]: wrote ssh authorized keys file for user: core Dec 16 12:47:13.303288 containerd[1623]: time="2025-12-16T12:47:13Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Dec 16 12:47:13.304583 containerd[1623]: time="2025-12-16T12:47:13.304562005Z" level=info msg="starting containerd" revision=fcd43222d6b07379a4be9786bda52438f0dd16a1 version=v2.1.5 Dec 16 12:47:13.311423 update-ssh-keys[1680]: Updated "/home/core/.ssh/authorized_keys" Dec 16 12:47:13.313069 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Dec 16 12:47:13.319733 systemd[1]: Finished sshkeys.service. Dec 16 12:47:13.334654 locksmithd[1663]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Dec 16 12:47:13.339559 containerd[1623]: time="2025-12-16T12:47:13.339533505Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="8.105µs" Dec 16 12:47:13.339616 containerd[1623]: time="2025-12-16T12:47:13.339603907Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Dec 16 12:47:13.339674 containerd[1623]: time="2025-12-16T12:47:13.339663659Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Dec 16 12:47:13.339728 containerd[1623]: time="2025-12-16T12:47:13.339717329Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Dec 16 12:47:13.340863 containerd[1623]: time="2025-12-16T12:47:13.339863764Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Dec 16 12:47:13.340863 containerd[1623]: time="2025-12-16T12:47:13.339882168Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Dec 16 12:47:13.340863 containerd[1623]: time="2025-12-16T12:47:13.339930699Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Dec 16 12:47:13.340863 containerd[1623]: time="2025-12-16T12:47:13.339939836Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Dec 16 12:47:13.340863 containerd[1623]: time="2025-12-16T12:47:13.340102762Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Dec 16 12:47:13.340863 containerd[1623]: time="2025-12-16T12:47:13.340114885Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Dec 16 12:47:13.340863 containerd[1623]: time="2025-12-16T12:47:13.340123140Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Dec 16 12:47:13.340863 containerd[1623]: time="2025-12-16T12:47:13.340129242Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Dec 16 12:47:13.340863 containerd[1623]: time="2025-12-16T12:47:13.340233247Z" level=info msg="skip loading plugin" error="EROFS unsupported, please `modprobe erofs`: skip plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Dec 16 12:47:13.340863 containerd[1623]: time="2025-12-16T12:47:13.340243747Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Dec 16 12:47:13.340863 containerd[1623]: time="2025-12-16T12:47:13.340306234Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Dec 16 12:47:13.340863 containerd[1623]: time="2025-12-16T12:47:13.340437199Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Dec 16 12:47:13.341226 containerd[1623]: time="2025-12-16T12:47:13.340472796Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Dec 16 12:47:13.341226 containerd[1623]: time="2025-12-16T12:47:13.340480641Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Dec 16 12:47:13.341597 containerd[1623]: time="2025-12-16T12:47:13.341580834Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Dec 16 12:47:13.344221 containerd[1623]: time="2025-12-16T12:47:13.343991736Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Dec 16 12:47:13.344802 containerd[1623]: time="2025-12-16T12:47:13.344786747Z" level=info msg="metadata content store policy set" policy=shared Dec 16 12:47:13.348705 containerd[1623]: time="2025-12-16T12:47:13.348688475Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Dec 16 12:47:13.348784 containerd[1623]: time="2025-12-16T12:47:13.348772553Z" level=info msg="loading plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Dec 16 12:47:13.348983 containerd[1623]: time="2025-12-16T12:47:13.348967247Z" level=info msg="skip loading plugin" error="could not find mkfs.erofs: exec: \"mkfs.erofs\": executable file not found in $PATH: skip plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Dec 16 12:47:13.349089 containerd[1623]: time="2025-12-16T12:47:13.349075801Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Dec 16 12:47:13.349136 containerd[1623]: time="2025-12-16T12:47:13.349126817Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Dec 16 12:47:13.349331 containerd[1623]: time="2025-12-16T12:47:13.349321823Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Dec 16 12:47:13.349371 containerd[1623]: time="2025-12-16T12:47:13.349362579Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Dec 16 12:47:13.349407 containerd[1623]: time="2025-12-16T12:47:13.349399007Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Dec 16 12:47:13.349444 containerd[1623]: time="2025-12-16T12:47:13.349436227Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Dec 16 12:47:13.349497 containerd[1623]: time="2025-12-16T12:47:13.349487604Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Dec 16 12:47:13.349551 containerd[1623]: time="2025-12-16T12:47:13.349540533Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Dec 16 12:47:13.349590 containerd[1623]: time="2025-12-16T12:47:13.349582292Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Dec 16 12:47:13.349624 containerd[1623]: time="2025-12-16T12:47:13.349616926Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Dec 16 12:47:13.349661 containerd[1623]: time="2025-12-16T12:47:13.349652994Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349779141Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349800501Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349811972Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349824165Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349833813Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349840716Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349849182Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349856045Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349863178Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349870602Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349878327Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349895549Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349927339Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349936596Z" level=info msg="Start snapshots syncer" Dec 16 12:47:13.350192 containerd[1623]: time="2025-12-16T12:47:13.349955882Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Dec 16 12:47:13.351116 containerd[1623]: time="2025-12-16T12:47:13.350858394Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"cgroupWritable\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"\",\"binDirs\":[\"/opt/cni/bin\"],\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogLineSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Dec 16 12:47:13.351116 containerd[1623]: time="2025-12-16T12:47:13.350903349Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Dec 16 12:47:13.351685 containerd[1623]: time="2025-12-16T12:47:13.351433273Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Dec 16 12:47:13.351685 containerd[1623]: time="2025-12-16T12:47:13.351560131Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Dec 16 12:47:13.351685 containerd[1623]: time="2025-12-16T12:47:13.351579337Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Dec 16 12:47:13.351685 containerd[1623]: time="2025-12-16T12:47:13.351588103Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Dec 16 12:47:13.351685 containerd[1623]: time="2025-12-16T12:47:13.351595868Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Dec 16 12:47:13.351685 containerd[1623]: time="2025-12-16T12:47:13.351604985Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Dec 16 12:47:13.351685 containerd[1623]: time="2025-12-16T12:47:13.351612028Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Dec 16 12:47:13.351685 containerd[1623]: time="2025-12-16T12:47:13.351620123Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Dec 16 12:47:13.351685 containerd[1623]: time="2025-12-16T12:47:13.351627217Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Dec 16 12:47:13.351685 containerd[1623]: time="2025-12-16T12:47:13.351634981Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Dec 16 12:47:13.352039 containerd[1623]: time="2025-12-16T12:47:13.351979056Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Dec 16 12:47:13.352039 containerd[1623]: time="2025-12-16T12:47:13.351995607Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Dec 16 12:47:13.352039 containerd[1623]: time="2025-12-16T12:47:13.352002391Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Dec 16 12:47:13.352356 containerd[1623]: time="2025-12-16T12:47:13.352011257Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Dec 16 12:47:13.352356 containerd[1623]: time="2025-12-16T12:47:13.352183370Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Dec 16 12:47:13.352356 containerd[1623]: time="2025-12-16T12:47:13.352193639Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Dec 16 12:47:13.352356 containerd[1623]: time="2025-12-16T12:47:13.352202806Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Dec 16 12:47:13.352356 containerd[1623]: time="2025-12-16T12:47:13.352211883Z" level=info msg="runtime interface created" Dec 16 12:47:13.352356 containerd[1623]: time="2025-12-16T12:47:13.352220439Z" level=info msg="created NRI interface" Dec 16 12:47:13.352356 containerd[1623]: time="2025-12-16T12:47:13.352241128Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Dec 16 12:47:13.352356 containerd[1623]: time="2025-12-16T12:47:13.352249954Z" level=info msg="Connect containerd service" Dec 16 12:47:13.352356 containerd[1623]: time="2025-12-16T12:47:13.352265444Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Dec 16 12:47:13.353793 containerd[1623]: time="2025-12-16T12:47:13.353600909Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 16 12:47:13.453237 systemd-networkd[1527]: eth1: Gained IPv6LL Dec 16 12:47:13.453682 systemd-timesyncd[1531]: Network configuration changed, trying to establish connection. Dec 16 12:47:13.457809 sshd_keygen[1626]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Dec 16 12:47:13.456477 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Dec 16 12:47:13.460090 systemd[1]: Reached target network-online.target - Network is Online. Dec 16 12:47:13.465072 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 12:47:13.467913 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Dec 16 12:47:13.488440 containerd[1623]: time="2025-12-16T12:47:13.488411507Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Dec 16 12:47:13.488512 containerd[1623]: time="2025-12-16T12:47:13.488473964Z" level=info msg=serving... address=/run/containerd/containerd.sock Dec 16 12:47:13.488512 containerd[1623]: time="2025-12-16T12:47:13.488497879Z" level=info msg="Start subscribing containerd event" Dec 16 12:47:13.488553 containerd[1623]: time="2025-12-16T12:47:13.488517706Z" level=info msg="Start recovering state" Dec 16 12:47:13.488601 containerd[1623]: time="2025-12-16T12:47:13.488586556Z" level=info msg="Start event monitor" Dec 16 12:47:13.488622 containerd[1623]: time="2025-12-16T12:47:13.488606683Z" level=info msg="Start cni network conf syncer for default" Dec 16 12:47:13.488622 containerd[1623]: time="2025-12-16T12:47:13.488612644Z" level=info msg="Start streaming server" Dec 16 12:47:13.488622 containerd[1623]: time="2025-12-16T12:47:13.488618906Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Dec 16 12:47:13.488667 containerd[1623]: time="2025-12-16T12:47:13.488624126Z" level=info msg="runtime interface starting up..." Dec 16 12:47:13.488667 containerd[1623]: time="2025-12-16T12:47:13.488628303Z" level=info msg="starting plugins..." Dec 16 12:47:13.488667 containerd[1623]: time="2025-12-16T12:47:13.488636820Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Dec 16 12:47:13.488800 systemd[1]: Started containerd.service - containerd container runtime. Dec 16 12:47:13.501865 containerd[1623]: time="2025-12-16T12:47:13.498446247Z" level=info msg="containerd successfully booted in 0.195434s" Dec 16 12:47:13.513509 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Dec 16 12:47:13.517408 systemd[1]: Starting issuegen.service - Generate /run/issue... Dec 16 12:47:13.521785 systemd[1]: Started sshd@0-77.42.19.41:22-147.75.109.163:38370.service - OpenSSH per-connection server daemon (147.75.109.163:38370). Dec 16 12:47:13.539390 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Dec 16 12:47:13.556288 systemd[1]: issuegen.service: Deactivated successfully. Dec 16 12:47:13.556495 systemd[1]: Finished issuegen.service - Generate /run/issue. Dec 16 12:47:13.560836 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Dec 16 12:47:13.591408 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Dec 16 12:47:13.595742 tar[1617]: linux-amd64/README.md Dec 16 12:47:13.596745 systemd[1]: Started getty@tty1.service - Getty on tty1. Dec 16 12:47:13.599618 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Dec 16 12:47:13.601417 systemd[1]: Reached target getty.target - Login Prompts. Dec 16 12:47:13.616153 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Dec 16 12:47:14.094211 systemd-networkd[1527]: eth0: Gained IPv6LL Dec 16 12:47:14.095116 systemd-timesyncd[1531]: Network configuration changed, trying to establish connection. Dec 16 12:47:14.417740 sshd[1718]: Accepted publickey for core from 147.75.109.163 port 38370 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:47:14.420851 sshd-session[1718]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:47:14.441788 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Dec 16 12:47:14.448244 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Dec 16 12:47:14.457805 systemd-logind[1601]: New session 1 of user core. Dec 16 12:47:14.485219 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Dec 16 12:47:14.495205 systemd[1]: Starting user@500.service - User Manager for UID 500... Dec 16 12:47:14.515206 (systemd)[1738]: pam_unix(systemd-user:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:47:14.519699 systemd-logind[1601]: New session 2 of user core. Dec 16 12:47:14.679401 systemd[1738]: Queued start job for default target default.target. Dec 16 12:47:14.685136 systemd[1738]: Created slice app.slice - User Application Slice. Dec 16 12:47:14.685164 systemd[1738]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of User's Temporary Directories. Dec 16 12:47:14.685176 systemd[1738]: Reached target paths.target - Paths. Dec 16 12:47:14.685211 systemd[1738]: Reached target timers.target - Timers. Dec 16 12:47:14.687094 systemd[1738]: Starting dbus.socket - D-Bus User Message Bus Socket... Dec 16 12:47:14.687779 systemd[1738]: Starting systemd-tmpfiles-setup.service - Create User Files and Directories... Dec 16 12:47:14.695933 systemd[1738]: Listening on dbus.socket - D-Bus User Message Bus Socket. Dec 16 12:47:14.696778 systemd[1738]: Reached target sockets.target - Sockets. Dec 16 12:47:14.706885 systemd[1738]: Finished systemd-tmpfiles-setup.service - Create User Files and Directories. Dec 16 12:47:14.707106 systemd[1738]: Reached target basic.target - Basic System. Dec 16 12:47:14.707157 systemd[1738]: Reached target default.target - Main User Target. Dec 16 12:47:14.707177 systemd[1738]: Startup finished in 176ms. Dec 16 12:47:14.707292 systemd[1]: Started user@500.service - User Manager for UID 500. Dec 16 12:47:14.719213 systemd[1]: Started session-1.scope - Session 1 of User core. Dec 16 12:47:14.798610 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 12:47:14.804471 systemd[1]: Reached target multi-user.target - Multi-User System. Dec 16 12:47:14.807137 systemd[1]: Startup finished in 3.797s (kernel) + 6.041s (initrd) + 4.758s (userspace) = 14.598s. Dec 16 12:47:14.810134 (kubelet)[1755]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 12:47:15.218293 systemd[1]: Started sshd@1-77.42.19.41:22-147.75.109.163:45332.service - OpenSSH per-connection server daemon (147.75.109.163:45332). Dec 16 12:47:15.656003 kubelet[1755]: E1216 12:47:15.655839 1755 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 12:47:15.659478 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 12:47:15.659718 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 12:47:15.660298 systemd[1]: kubelet.service: Consumed 1.354s CPU time, 265.4M memory peak. Dec 16 12:47:16.053907 sshd[1766]: Accepted publickey for core from 147.75.109.163 port 45332 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:47:16.056739 sshd-session[1766]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:47:16.067102 systemd-logind[1601]: New session 3 of user core. Dec 16 12:47:16.077270 systemd[1]: Started session-3.scope - Session 3 of User core. Dec 16 12:47:16.534343 sshd[1773]: Connection closed by 147.75.109.163 port 45332 Dec 16 12:47:16.535355 sshd-session[1766]: pam_unix(sshd:session): session closed for user core Dec 16 12:47:16.541214 systemd-logind[1601]: Session 3 logged out. Waiting for processes to exit. Dec 16 12:47:16.541337 systemd[1]: sshd@1-77.42.19.41:22-147.75.109.163:45332.service: Deactivated successfully. Dec 16 12:47:16.544355 systemd[1]: session-3.scope: Deactivated successfully. Dec 16 12:47:16.546666 systemd-logind[1601]: Removed session 3. Dec 16 12:47:16.706720 systemd[1]: Started sshd@2-77.42.19.41:22-147.75.109.163:45346.service - OpenSSH per-connection server daemon (147.75.109.163:45346). Dec 16 12:47:17.562791 sshd[1779]: Accepted publickey for core from 147.75.109.163 port 45346 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:47:17.564777 sshd-session[1779]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:47:17.573333 systemd-logind[1601]: New session 4 of user core. Dec 16 12:47:17.579260 systemd[1]: Started session-4.scope - Session 4 of User core. Dec 16 12:47:18.040861 sshd[1783]: Connection closed by 147.75.109.163 port 45346 Dec 16 12:47:18.041969 sshd-session[1779]: pam_unix(sshd:session): session closed for user core Dec 16 12:47:18.047998 systemd[1]: sshd@2-77.42.19.41:22-147.75.109.163:45346.service: Deactivated successfully. Dec 16 12:47:18.051198 systemd[1]: session-4.scope: Deactivated successfully. Dec 16 12:47:18.053586 systemd-logind[1601]: Session 4 logged out. Waiting for processes to exit. Dec 16 12:47:18.055181 systemd-logind[1601]: Removed session 4. Dec 16 12:47:18.218460 systemd[1]: Started sshd@3-77.42.19.41:22-147.75.109.163:45348.service - OpenSSH per-connection server daemon (147.75.109.163:45348). Dec 16 12:47:19.087047 sshd[1789]: Accepted publickey for core from 147.75.109.163 port 45348 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:47:19.088843 sshd-session[1789]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:47:19.095905 systemd-logind[1601]: New session 5 of user core. Dec 16 12:47:19.100216 systemd[1]: Started session-5.scope - Session 5 of User core. Dec 16 12:47:19.567672 sshd[1793]: Connection closed by 147.75.109.163 port 45348 Dec 16 12:47:19.568304 sshd-session[1789]: pam_unix(sshd:session): session closed for user core Dec 16 12:47:19.572612 systemd-logind[1601]: Session 5 logged out. Waiting for processes to exit. Dec 16 12:47:19.573238 systemd[1]: sshd@3-77.42.19.41:22-147.75.109.163:45348.service: Deactivated successfully. Dec 16 12:47:19.574747 systemd[1]: session-5.scope: Deactivated successfully. Dec 16 12:47:19.575847 systemd-logind[1601]: Removed session 5. Dec 16 12:47:19.739274 systemd[1]: Started sshd@4-77.42.19.41:22-147.75.109.163:45354.service - OpenSSH per-connection server daemon (147.75.109.163:45354). Dec 16 12:47:20.574299 sshd[1799]: Accepted publickey for core from 147.75.109.163 port 45354 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:47:20.575582 sshd-session[1799]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:47:20.580807 systemd-logind[1601]: New session 6 of user core. Dec 16 12:47:20.586154 systemd[1]: Started session-6.scope - Session 6 of User core. Dec 16 12:47:20.913392 sudo[1804]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Dec 16 12:47:20.913766 sudo[1804]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 16 12:47:20.928098 sudo[1804]: pam_unix(sudo:session): session closed for user root Dec 16 12:47:21.087199 sshd[1803]: Connection closed by 147.75.109.163 port 45354 Dec 16 12:47:21.088531 sshd-session[1799]: pam_unix(sshd:session): session closed for user core Dec 16 12:47:21.095966 systemd[1]: sshd@4-77.42.19.41:22-147.75.109.163:45354.service: Deactivated successfully. Dec 16 12:47:21.096305 systemd-logind[1601]: Session 6 logged out. Waiting for processes to exit. Dec 16 12:47:21.098090 systemd[1]: session-6.scope: Deactivated successfully. Dec 16 12:47:21.100445 systemd-logind[1601]: Removed session 6. Dec 16 12:47:21.268755 systemd[1]: Started sshd@5-77.42.19.41:22-147.75.109.163:45362.service - OpenSSH per-connection server daemon (147.75.109.163:45362). Dec 16 12:47:22.119671 sshd[1811]: Accepted publickey for core from 147.75.109.163 port 45362 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:47:22.121356 sshd-session[1811]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:47:22.127460 systemd-logind[1601]: New session 7 of user core. Dec 16 12:47:22.134176 systemd[1]: Started session-7.scope - Session 7 of User core. Dec 16 12:47:22.446121 sudo[1817]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Dec 16 12:47:22.446581 sudo[1817]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 16 12:47:22.449385 sudo[1817]: pam_unix(sudo:session): session closed for user root Dec 16 12:47:22.455401 sudo[1816]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Dec 16 12:47:22.455646 sudo[1816]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 16 12:47:22.463855 systemd[1]: Starting audit-rules.service - Load Audit Rules... Dec 16 12:47:22.491000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Dec 16 12:47:22.492458 kernel: kauditd_printk_skb: 134 callbacks suppressed Dec 16 12:47:22.492506 kernel: audit: type=1305 audit(1765889242.491:234): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Dec 16 12:47:22.502590 kernel: audit: type=1300 audit(1765889242.491:234): arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffc8de06c30 a2=420 a3=0 items=0 ppid=1822 pid=1841 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:22.491000 audit[1841]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffc8de06c30 a2=420 a3=0 items=0 ppid=1822 pid=1841 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:22.502715 augenrules[1841]: No rules Dec 16 12:47:22.506211 kernel: audit: type=1327 audit(1765889242.491:234): proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Dec 16 12:47:22.491000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Dec 16 12:47:22.502989 systemd[1]: audit-rules.service: Deactivated successfully. Dec 16 12:47:22.503258 systemd[1]: Finished audit-rules.service - Load Audit Rules. Dec 16 12:47:22.507163 kernel: audit: type=1130 audit(1765889242.502:235): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:22.502000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:22.506544 sudo[1816]: pam_unix(sudo:session): session closed for user root Dec 16 12:47:22.502000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:22.516641 kernel: audit: type=1131 audit(1765889242.502:236): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:22.516685 kernel: audit: type=1106 audit(1765889242.506:237): pid=1816 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Dec 16 12:47:22.506000 audit[1816]: USER_END pid=1816 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Dec 16 12:47:22.506000 audit[1816]: CRED_DISP pid=1816 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Dec 16 12:47:22.527291 kernel: audit: type=1104 audit(1765889242.506:238): pid=1816 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Dec 16 12:47:22.665857 sshd[1815]: Connection closed by 147.75.109.163 port 45362 Dec 16 12:47:22.666349 sshd-session[1811]: pam_unix(sshd:session): session closed for user core Dec 16 12:47:22.667000 audit[1811]: USER_END pid=1811 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:47:22.669655 systemd-logind[1601]: Session 7 logged out. Waiting for processes to exit. Dec 16 12:47:22.671445 systemd[1]: sshd@5-77.42.19.41:22-147.75.109.163:45362.service: Deactivated successfully. Dec 16 12:47:22.673348 systemd[1]: session-7.scope: Deactivated successfully. Dec 16 12:47:22.667000 audit[1811]: CRED_DISP pid=1811 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:47:22.675439 systemd-logind[1601]: Removed session 7. Dec 16 12:47:22.679832 kernel: audit: type=1106 audit(1765889242.667:239): pid=1811 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:47:22.679881 kernel: audit: type=1104 audit(1765889242.667:240): pid=1811 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:47:22.679902 kernel: audit: type=1131 audit(1765889242.667:241): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@5-77.42.19.41:22-147.75.109.163:45362 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:22.667000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@5-77.42.19.41:22-147.75.109.163:45362 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:22.872000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-77.42.19.41:22-147.75.109.163:33338 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:22.872310 systemd[1]: Started sshd@6-77.42.19.41:22-147.75.109.163:33338.service - OpenSSH per-connection server daemon (147.75.109.163:33338). Dec 16 12:47:23.801000 audit[1850]: USER_ACCT pid=1850 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:47:23.801541 sshd[1850]: Accepted publickey for core from 147.75.109.163 port 33338 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:47:23.802000 audit[1850]: CRED_ACQ pid=1850 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:47:23.802000 audit[1850]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff0fb6a920 a2=3 a3=0 items=0 ppid=1 pid=1850 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=8 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:23.802000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:47:23.802713 sshd-session[1850]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:47:23.807825 systemd-logind[1601]: New session 8 of user core. Dec 16 12:47:23.813176 systemd[1]: Started session-8.scope - Session 8 of User core. Dec 16 12:47:23.815000 audit[1850]: USER_START pid=1850 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:47:23.816000 audit[1854]: CRED_ACQ pid=1854 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:47:24.159000 audit[1855]: USER_ACCT pid=1855 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_unix,pam_faillock acct="core" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Dec 16 12:47:24.159906 sudo[1855]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Dec 16 12:47:24.160000 audit[1855]: CRED_REFR pid=1855 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Dec 16 12:47:24.160000 audit[1855]: USER_START pid=1855 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Dec 16 12:47:24.160238 sudo[1855]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Dec 16 12:47:24.520320 systemd[1]: Starting docker.service - Docker Application Container Engine... Dec 16 12:47:24.533277 (dockerd)[1874]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Dec 16 12:47:24.787219 dockerd[1874]: time="2025-12-16T12:47:24.786513853Z" level=info msg="Starting up" Dec 16 12:47:24.790236 dockerd[1874]: time="2025-12-16T12:47:24.790210676Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Dec 16 12:47:24.799919 dockerd[1874]: time="2025-12-16T12:47:24.799882115Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Dec 16 12:47:24.823234 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2263845643-merged.mount: Deactivated successfully. Dec 16 12:47:24.872971 dockerd[1874]: time="2025-12-16T12:47:24.872909458Z" level=info msg="Loading containers: start." Dec 16 12:47:24.884042 kernel: Initializing XFRM netlink socket Dec 16 12:47:24.935000 audit[1922]: NETFILTER_CFG table=nat:2 family=2 entries=2 op=nft_register_chain pid=1922 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:24.935000 audit[1922]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffe4feb9520 a2=0 a3=0 items=0 ppid=1874 pid=1922 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:24.935000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Dec 16 12:47:24.937000 audit[1924]: NETFILTER_CFG table=filter:3 family=2 entries=2 op=nft_register_chain pid=1924 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:24.937000 audit[1924]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffd77545480 a2=0 a3=0 items=0 ppid=1874 pid=1924 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:24.937000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Dec 16 12:47:24.939000 audit[1926]: NETFILTER_CFG table=filter:4 family=2 entries=1 op=nft_register_chain pid=1926 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:24.939000 audit[1926]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffee23f0020 a2=0 a3=0 items=0 ppid=1874 pid=1926 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:24.939000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Dec 16 12:47:24.941000 audit[1928]: NETFILTER_CFG table=filter:5 family=2 entries=1 op=nft_register_chain pid=1928 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:24.941000 audit[1928]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fffaa5dee90 a2=0 a3=0 items=0 ppid=1874 pid=1928 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:24.941000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D425249444745 Dec 16 12:47:24.942000 audit[1930]: NETFILTER_CFG table=filter:6 family=2 entries=1 op=nft_register_chain pid=1930 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:24.942000 audit[1930]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffcdf7fc380 a2=0 a3=0 items=0 ppid=1874 pid=1930 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:24.942000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D4354 Dec 16 12:47:24.944000 audit[1932]: NETFILTER_CFG table=filter:7 family=2 entries=1 op=nft_register_chain pid=1932 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:24.944000 audit[1932]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffcbc0855c0 a2=0 a3=0 items=0 ppid=1874 pid=1932 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:24.944000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Dec 16 12:47:24.946000 audit[1934]: NETFILTER_CFG table=filter:8 family=2 entries=1 op=nft_register_chain pid=1934 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:24.946000 audit[1934]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffcd4c48090 a2=0 a3=0 items=0 ppid=1874 pid=1934 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:24.946000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Dec 16 12:47:24.948000 audit[1936]: NETFILTER_CFG table=nat:9 family=2 entries=2 op=nft_register_chain pid=1936 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:24.948000 audit[1936]: SYSCALL arch=c000003e syscall=46 success=yes exit=384 a0=3 a1=7ffc8c87fd30 a2=0 a3=0 items=0 ppid=1874 pid=1936 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:24.948000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Dec 16 12:47:24.973000 audit[1939]: NETFILTER_CFG table=nat:10 family=2 entries=2 op=nft_register_chain pid=1939 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:24.973000 audit[1939]: SYSCALL arch=c000003e syscall=46 success=yes exit=472 a0=3 a1=7ffed02982f0 a2=0 a3=0 items=0 ppid=1874 pid=1939 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:24.973000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003132372E302E302E302F38 Dec 16 12:47:24.975000 audit[1941]: NETFILTER_CFG table=filter:11 family=2 entries=2 op=nft_register_chain pid=1941 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:24.975000 audit[1941]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffd790772c0 a2=0 a3=0 items=0 ppid=1874 pid=1941 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:24.975000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D464F5257415244 Dec 16 12:47:24.977000 audit[1943]: NETFILTER_CFG table=filter:12 family=2 entries=1 op=nft_register_rule pid=1943 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:24.977000 audit[1943]: SYSCALL arch=c000003e syscall=46 success=yes exit=236 a0=3 a1=7ffc18e0e020 a2=0 a3=0 items=0 ppid=1874 pid=1943 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:24.977000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D425249444745 Dec 16 12:47:24.979000 audit[1945]: NETFILTER_CFG table=filter:13 family=2 entries=1 op=nft_register_rule pid=1945 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:24.979000 audit[1945]: SYSCALL arch=c000003e syscall=46 success=yes exit=248 a0=3 a1=7ffef8f3e760 a2=0 a3=0 items=0 ppid=1874 pid=1945 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:24.979000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Dec 16 12:47:24.981000 audit[1947]: NETFILTER_CFG table=filter:14 family=2 entries=1 op=nft_register_rule pid=1947 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:24.981000 audit[1947]: SYSCALL arch=c000003e syscall=46 success=yes exit=232 a0=3 a1=7ffc8e16f560 a2=0 a3=0 items=0 ppid=1874 pid=1947 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:24.981000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D4354 Dec 16 12:47:25.009000 audit[1977]: NETFILTER_CFG table=nat:15 family=10 entries=2 op=nft_register_chain pid=1977 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.009000 audit[1977]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffdd6469a50 a2=0 a3=0 items=0 ppid=1874 pid=1977 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.009000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Dec 16 12:47:25.010000 audit[1979]: NETFILTER_CFG table=filter:16 family=10 entries=2 op=nft_register_chain pid=1979 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.010000 audit[1979]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffd85b90570 a2=0 a3=0 items=0 ppid=1874 pid=1979 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.010000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Dec 16 12:47:25.012000 audit[1981]: NETFILTER_CFG table=filter:17 family=10 entries=1 op=nft_register_chain pid=1981 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.012000 audit[1981]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffea1373560 a2=0 a3=0 items=0 ppid=1874 pid=1981 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.012000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Dec 16 12:47:25.013000 audit[1983]: NETFILTER_CFG table=filter:18 family=10 entries=1 op=nft_register_chain pid=1983 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.013000 audit[1983]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff37e58fe0 a2=0 a3=0 items=0 ppid=1874 pid=1983 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.013000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D425249444745 Dec 16 12:47:25.015000 audit[1985]: NETFILTER_CFG table=filter:19 family=10 entries=1 op=nft_register_chain pid=1985 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.015000 audit[1985]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffd888d2aa0 a2=0 a3=0 items=0 ppid=1874 pid=1985 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.015000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D4354 Dec 16 12:47:25.016000 audit[1987]: NETFILTER_CFG table=filter:20 family=10 entries=1 op=nft_register_chain pid=1987 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.016000 audit[1987]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7fffb7d5f610 a2=0 a3=0 items=0 ppid=1874 pid=1987 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.016000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Dec 16 12:47:25.017000 audit[1989]: NETFILTER_CFG table=filter:21 family=10 entries=1 op=nft_register_chain pid=1989 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.017000 audit[1989]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7fff14a3a4f0 a2=0 a3=0 items=0 ppid=1874 pid=1989 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.017000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Dec 16 12:47:25.019000 audit[1991]: NETFILTER_CFG table=nat:22 family=10 entries=2 op=nft_register_chain pid=1991 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.019000 audit[1991]: SYSCALL arch=c000003e syscall=46 success=yes exit=384 a0=3 a1=7fffb09a0100 a2=0 a3=0 items=0 ppid=1874 pid=1991 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.019000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Dec 16 12:47:25.022000 audit[1993]: NETFILTER_CFG table=nat:23 family=10 entries=2 op=nft_register_chain pid=1993 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.022000 audit[1993]: SYSCALL arch=c000003e syscall=46 success=yes exit=484 a0=3 a1=7ffcd8f6c0d0 a2=0 a3=0 items=0 ppid=1874 pid=1993 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.022000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003A3A312F313238 Dec 16 12:47:25.023000 audit[1995]: NETFILTER_CFG table=filter:24 family=10 entries=2 op=nft_register_chain pid=1995 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.023000 audit[1995]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffe25404440 a2=0 a3=0 items=0 ppid=1874 pid=1995 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.023000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D464F5257415244 Dec 16 12:47:25.025000 audit[1997]: NETFILTER_CFG table=filter:25 family=10 entries=1 op=nft_register_rule pid=1997 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.025000 audit[1997]: SYSCALL arch=c000003e syscall=46 success=yes exit=236 a0=3 a1=7ffc5840c6a0 a2=0 a3=0 items=0 ppid=1874 pid=1997 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.025000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D425249444745 Dec 16 12:47:25.026000 audit[1999]: NETFILTER_CFG table=filter:26 family=10 entries=1 op=nft_register_rule pid=1999 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.026000 audit[1999]: SYSCALL arch=c000003e syscall=46 success=yes exit=248 a0=3 a1=7ffe69fcfac0 a2=0 a3=0 items=0 ppid=1874 pid=1999 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.026000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Dec 16 12:47:25.028000 audit[2001]: NETFILTER_CFG table=filter:27 family=10 entries=1 op=nft_register_rule pid=2001 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.028000 audit[2001]: SYSCALL arch=c000003e syscall=46 success=yes exit=232 a0=3 a1=7fff8769c900 a2=0 a3=0 items=0 ppid=1874 pid=2001 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.028000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D4354 Dec 16 12:47:25.032000 audit[2006]: NETFILTER_CFG table=filter:28 family=2 entries=1 op=nft_register_chain pid=2006 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:25.032000 audit[2006]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffdd79388e0 a2=0 a3=0 items=0 ppid=1874 pid=2006 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.032000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Dec 16 12:47:25.034000 audit[2008]: NETFILTER_CFG table=filter:29 family=2 entries=1 op=nft_register_rule pid=2008 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:25.034000 audit[2008]: SYSCALL arch=c000003e syscall=46 success=yes exit=212 a0=3 a1=7ffc0c55bad0 a2=0 a3=0 items=0 ppid=1874 pid=2008 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.034000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E Dec 16 12:47:25.035000 audit[2010]: NETFILTER_CFG table=filter:30 family=2 entries=1 op=nft_register_rule pid=2010 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:25.035000 audit[2010]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7ffe8598c4e0 a2=0 a3=0 items=0 ppid=1874 pid=2010 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.035000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Dec 16 12:47:25.037000 audit[2012]: NETFILTER_CFG table=filter:31 family=10 entries=1 op=nft_register_chain pid=2012 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.037000 audit[2012]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7fffce9e7c00 a2=0 a3=0 items=0 ppid=1874 pid=2012 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.037000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Dec 16 12:47:25.039000 audit[2014]: NETFILTER_CFG table=filter:32 family=10 entries=1 op=nft_register_rule pid=2014 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.039000 audit[2014]: SYSCALL arch=c000003e syscall=46 success=yes exit=212 a0=3 a1=7ffe580035a0 a2=0 a3=0 items=0 ppid=1874 pid=2014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.039000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E Dec 16 12:47:25.041000 audit[2016]: NETFILTER_CFG table=filter:33 family=10 entries=1 op=nft_register_rule pid=2016 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:25.041000 audit[2016]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7ffcec539c10 a2=0 a3=0 items=0 ppid=1874 pid=2016 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.041000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Dec 16 12:47:25.047911 systemd-timesyncd[1531]: Network configuration changed, trying to establish connection. Dec 16 12:47:25.064000 audit[2021]: NETFILTER_CFG table=nat:34 family=2 entries=2 op=nft_register_chain pid=2021 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:25.064000 audit[2021]: SYSCALL arch=c000003e syscall=46 success=yes exit=520 a0=3 a1=7fff4b8b8ec0 a2=0 a3=0 items=0 ppid=1874 pid=2021 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.064000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900504F5354524F5554494E47002D73003137322E31372E302E302F31360000002D6F00646F636B657230002D6A004D415351554552414445 Dec 16 12:47:25.067000 audit[2023]: NETFILTER_CFG table=nat:35 family=2 entries=1 op=nft_register_rule pid=2023 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:25.067000 audit[2023]: SYSCALL arch=c000003e syscall=46 success=yes exit=288 a0=3 a1=7ffd8b04d870 a2=0 a3=0 items=0 ppid=1874 pid=2023 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.067000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900444F434B4552002D6900646F636B657230002D6A0052455455524E Dec 16 12:47:25.076000 audit[2031]: NETFILTER_CFG table=filter:36 family=2 entries=1 op=nft_register_rule pid=2031 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:25.076000 audit[2031]: SYSCALL arch=c000003e syscall=46 success=yes exit=300 a0=3 a1=7fff0bff2320 a2=0 a3=0 items=0 ppid=1874 pid=2031 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.076000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D464F5257415244002D6900646F636B657230002D6A00414343455054 Dec 16 12:47:25.084000 audit[2037]: NETFILTER_CFG table=filter:37 family=2 entries=1 op=nft_register_rule pid=2037 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:25.084000 audit[2037]: SYSCALL arch=c000003e syscall=46 success=yes exit=376 a0=3 a1=7ffdd6b435c0 a2=0 a3=0 items=0 ppid=1874 pid=2037 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.084000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45520000002D6900646F636B657230002D6F00646F636B657230002D6A0044524F50 Dec 16 12:47:25.086000 audit[2039]: NETFILTER_CFG table=filter:38 family=2 entries=1 op=nft_register_rule pid=2039 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:25.086000 audit[2039]: SYSCALL arch=c000003e syscall=46 success=yes exit=512 a0=3 a1=7ffcad148720 a2=0 a3=0 items=0 ppid=1874 pid=2039 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.086000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D4354002D6F00646F636B657230002D6D00636F6E6E747261636B002D2D637473746174650052454C415445442C45535441424C4953484544002D6A00414343455054 Dec 16 12:47:25.087000 audit[2041]: NETFILTER_CFG table=filter:39 family=2 entries=1 op=nft_register_rule pid=2041 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:25.087000 audit[2041]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7fff308acca0 a2=0 a3=0 items=0 ppid=1874 pid=2041 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.087000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D425249444745002D6F00646F636B657230002D6A00444F434B4552 Dec 16 12:47:25.089000 audit[2043]: NETFILTER_CFG table=filter:40 family=2 entries=1 op=nft_register_rule pid=2043 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:25.089000 audit[2043]: SYSCALL arch=c000003e syscall=46 success=yes exit=428 a0=3 a1=7ffc06af6cf0 a2=0 a3=0 items=0 ppid=1874 pid=2043 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.089000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D49534F4C4154494F4E2D53544147452D31002D6900646F636B6572300000002D6F00646F636B657230002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D32 Dec 16 12:47:25.091000 audit[2045]: NETFILTER_CFG table=filter:41 family=2 entries=1 op=nft_register_rule pid=2045 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:25.091000 audit[2045]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffe055152c0 a2=0 a3=0 items=0 ppid=1874 pid=2045 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:25.091000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4900444F434B45522D49534F4C4154494F4E2D53544147452D32002D6F00646F636B657230002D6A0044524F50 Dec 16 12:47:25.091955 systemd-networkd[1527]: docker0: Link UP Dec 16 12:47:25.096945 dockerd[1874]: time="2025-12-16T12:47:25.096910995Z" level=info msg="Loading containers: done." Dec 16 12:47:25.116784 dockerd[1874]: time="2025-12-16T12:47:25.116741578Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Dec 16 12:47:25.116885 dockerd[1874]: time="2025-12-16T12:47:25.116818562Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Dec 16 12:47:25.116905 dockerd[1874]: time="2025-12-16T12:47:25.116885898Z" level=info msg="Initializing buildkit" Dec 16 12:47:25.138063 dockerd[1874]: time="2025-12-16T12:47:25.137990230Z" level=info msg="Completed buildkit initialization" Dec 16 12:47:25.144649 dockerd[1874]: time="2025-12-16T12:47:25.144616969Z" level=info msg="Daemon has completed initialization" Dec 16 12:47:25.144763 dockerd[1874]: time="2025-12-16T12:47:25.144654378Z" level=info msg="API listen on /run/docker.sock" Dec 16 12:47:25.145005 systemd[1]: Started docker.service - Docker Application Container Engine. Dec 16 12:47:25.145000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=docker comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:25.261871 systemd-timesyncd[1531]: Contacted time server 157.230.98.101:123 (2.flatcar.pool.ntp.org). Dec 16 12:47:25.262359 systemd-timesyncd[1531]: Initial clock synchronization to Tue 2025-12-16 12:47:25.623595 UTC. Dec 16 12:47:25.910125 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Dec 16 12:47:25.911812 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 12:47:26.022968 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 12:47:26.022000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:26.031427 (kubelet)[2093]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 12:47:26.074867 kubelet[2093]: E1216 12:47:26.074808 2093 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 12:47:26.077660 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 12:47:26.077795 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 12:47:26.077000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Dec 16 12:47:26.078251 systemd[1]: kubelet.service: Consumed 124ms CPU time, 110.6M memory peak. Dec 16 12:47:26.294591 containerd[1623]: time="2025-12-16T12:47:26.294453284Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.10\"" Dec 16 12:47:26.931508 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3438619017.mount: Deactivated successfully. Dec 16 12:47:27.763677 containerd[1623]: time="2025-12-16T12:47:27.763619296Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:27.764817 containerd[1623]: time="2025-12-16T12:47:27.764600373Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.10: active requests=0, bytes read=27403437" Dec 16 12:47:27.766421 containerd[1623]: time="2025-12-16T12:47:27.766392805Z" level=info msg="ImageCreate event name:\"sha256:77f8b0de97da9ee43e174b170c363c893ab69a20b03878e1bf6b54b10d44ef6f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:27.768728 containerd[1623]: time="2025-12-16T12:47:27.768699997Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:af4ee57c047e31a7f58422b94a9ec4c62221d3deebb16755bdeff720df796189\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:27.769798 containerd[1623]: time="2025-12-16T12:47:27.769765583Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.10\" with image id \"sha256:77f8b0de97da9ee43e174b170c363c893ab69a20b03878e1bf6b54b10d44ef6f\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.10\", repo digest \"registry.k8s.io/kube-apiserver@sha256:af4ee57c047e31a7f58422b94a9ec4c62221d3deebb16755bdeff720df796189\", size \"29068782\" in 1.475269267s" Dec 16 12:47:27.769836 containerd[1623]: time="2025-12-16T12:47:27.769803088Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.10\" returns image reference \"sha256:77f8b0de97da9ee43e174b170c363c893ab69a20b03878e1bf6b54b10d44ef6f\"" Dec 16 12:47:27.770659 containerd[1623]: time="2025-12-16T12:47:27.770633121Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.10\"" Dec 16 12:47:29.058751 containerd[1623]: time="2025-12-16T12:47:29.058691191Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:29.060170 containerd[1623]: time="2025-12-16T12:47:29.060097796Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.10: active requests=0, bytes read=24983855" Dec 16 12:47:29.061210 containerd[1623]: time="2025-12-16T12:47:29.061117127Z" level=info msg="ImageCreate event name:\"sha256:34e0beef266f1ca24c0093506853b1cc0ed91e873aeef655f39721813f10f924\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:29.063438 containerd[1623]: time="2025-12-16T12:47:29.063360383Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:efbd9d1dfcd2940e1c73a1476c880c3c2cdf04cc60722d329b21cd48745c8660\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:29.064365 containerd[1623]: time="2025-12-16T12:47:29.064062123Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.10\" with image id \"sha256:34e0beef266f1ca24c0093506853b1cc0ed91e873aeef655f39721813f10f924\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.10\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:efbd9d1dfcd2940e1c73a1476c880c3c2cdf04cc60722d329b21cd48745c8660\", size \"26649046\" in 1.293324642s" Dec 16 12:47:29.064365 containerd[1623]: time="2025-12-16T12:47:29.064117596Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.10\" returns image reference \"sha256:34e0beef266f1ca24c0093506853b1cc0ed91e873aeef655f39721813f10f924\"" Dec 16 12:47:29.064673 containerd[1623]: time="2025-12-16T12:47:29.064570059Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.10\"" Dec 16 12:47:30.142492 containerd[1623]: time="2025-12-16T12:47:30.142420346Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:30.143812 containerd[1623]: time="2025-12-16T12:47:30.143773247Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.10: active requests=0, bytes read=19396111" Dec 16 12:47:30.144606 containerd[1623]: time="2025-12-16T12:47:30.144557808Z" level=info msg="ImageCreate event name:\"sha256:fd6f6aae834c2ec73b534bc30902f1602089a8f4d1bbd8c521fe2b39968efe4a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:30.147256 containerd[1623]: time="2025-12-16T12:47:30.147221650Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:9c58e1adcad5af66d1d9ca5cf9a4c266e4054b8f19f91a8fff1993549e657b10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:30.148171 containerd[1623]: time="2025-12-16T12:47:30.148150574Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.10\" with image id \"sha256:fd6f6aae834c2ec73b534bc30902f1602089a8f4d1bbd8c521fe2b39968efe4a\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.10\", repo digest \"registry.k8s.io/kube-scheduler@sha256:9c58e1adcad5af66d1d9ca5cf9a4c266e4054b8f19f91a8fff1993549e657b10\", size \"21061302\" in 1.083551327s" Dec 16 12:47:30.148246 containerd[1623]: time="2025-12-16T12:47:30.148228951Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.10\" returns image reference \"sha256:fd6f6aae834c2ec73b534bc30902f1602089a8f4d1bbd8c521fe2b39968efe4a\"" Dec 16 12:47:30.149125 containerd[1623]: time="2025-12-16T12:47:30.149090605Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.10\"" Dec 16 12:47:31.178449 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3010864256.mount: Deactivated successfully. Dec 16 12:47:31.448746 containerd[1623]: time="2025-12-16T12:47:31.448630535Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:31.449862 containerd[1623]: time="2025-12-16T12:47:31.449815754Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.10: active requests=0, bytes read=31157702" Dec 16 12:47:31.450594 containerd[1623]: time="2025-12-16T12:47:31.450554941Z" level=info msg="ImageCreate event name:\"sha256:db4bcdca85a39c02add2db5eed4fc6ab21eb20616fbf8cd2cf824e59e384a956\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:31.452144 containerd[1623]: time="2025-12-16T12:47:31.452091349Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:e3dda1c7b384f9eb5b2fa1c27493b23b80e6204b9fa2ee8791b2de078f468cbf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:31.452757 containerd[1623]: time="2025-12-16T12:47:31.452720253Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.10\" with image id \"sha256:db4bcdca85a39c02add2db5eed4fc6ab21eb20616fbf8cd2cf824e59e384a956\", repo tag \"registry.k8s.io/kube-proxy:v1.32.10\", repo digest \"registry.k8s.io/kube-proxy@sha256:e3dda1c7b384f9eb5b2fa1c27493b23b80e6204b9fa2ee8791b2de078f468cbf\", size \"31160442\" in 1.303598817s" Dec 16 12:47:31.452757 containerd[1623]: time="2025-12-16T12:47:31.452748383Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.10\" returns image reference \"sha256:db4bcdca85a39c02add2db5eed4fc6ab21eb20616fbf8cd2cf824e59e384a956\"" Dec 16 12:47:31.453445 containerd[1623]: time="2025-12-16T12:47:31.453399522Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Dec 16 12:47:31.961651 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2258070703.mount: Deactivated successfully. Dec 16 12:47:32.866466 containerd[1623]: time="2025-12-16T12:47:32.866406502Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:32.867569 containerd[1623]: time="2025-12-16T12:47:32.867408032Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=17570073" Dec 16 12:47:32.868393 containerd[1623]: time="2025-12-16T12:47:32.868354771Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:32.871362 containerd[1623]: time="2025-12-16T12:47:32.871336536Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:32.872279 containerd[1623]: time="2025-12-16T12:47:32.872252404Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.418827542s" Dec 16 12:47:32.872322 containerd[1623]: time="2025-12-16T12:47:32.872280505Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Dec 16 12:47:32.873015 containerd[1623]: time="2025-12-16T12:47:32.872946246Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Dec 16 12:47:33.955940 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1487782073.mount: Deactivated successfully. Dec 16 12:47:33.970220 containerd[1623]: time="2025-12-16T12:47:33.970119634Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 16 12:47:33.971718 containerd[1623]: time="2025-12-16T12:47:33.971434478Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Dec 16 12:47:33.973210 containerd[1623]: time="2025-12-16T12:47:33.973158619Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 16 12:47:33.976824 containerd[1623]: time="2025-12-16T12:47:33.976768960Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Dec 16 12:47:33.978260 containerd[1623]: time="2025-12-16T12:47:33.978018891Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 1.105041489s" Dec 16 12:47:33.978260 containerd[1623]: time="2025-12-16T12:47:33.978123509Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Dec 16 12:47:33.979267 containerd[1623]: time="2025-12-16T12:47:33.979100422Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Dec 16 12:47:34.554961 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1462911616.mount: Deactivated successfully. Dec 16 12:47:36.328699 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Dec 16 12:47:36.331252 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 12:47:36.450000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:36.450696 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 12:47:36.452350 kernel: kauditd_printk_skb: 134 callbacks suppressed Dec 16 12:47:36.452410 kernel: audit: type=1130 audit(1765889256.450:294): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:36.467522 (kubelet)[2287]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Dec 16 12:47:36.553305 kubelet[2287]: E1216 12:47:36.553209 2287 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 16 12:47:36.555940 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 16 12:47:36.556330 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 16 12:47:36.556000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Dec 16 12:47:36.557336 systemd[1]: kubelet.service: Consumed 166ms CPU time, 108.6M memory peak. Dec 16 12:47:36.567314 kernel: audit: type=1131 audit(1765889256.556:295): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Dec 16 12:47:38.190553 containerd[1623]: time="2025-12-16T12:47:38.190491816Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:38.191643 containerd[1623]: time="2025-12-16T12:47:38.191481208Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=55728979" Dec 16 12:47:38.192357 containerd[1623]: time="2025-12-16T12:47:38.192334317Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:38.194525 containerd[1623]: time="2025-12-16T12:47:38.194499563Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:38.195446 containerd[1623]: time="2025-12-16T12:47:38.195426202Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 4.216024893s" Dec 16 12:47:38.195523 containerd[1623]: time="2025-12-16T12:47:38.195510055Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Dec 16 12:47:41.160564 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 12:47:41.160852 systemd[1]: kubelet.service: Consumed 166ms CPU time, 108.6M memory peak. Dec 16 12:47:41.159000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:41.159000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:41.167272 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 12:47:41.170652 kernel: audit: type=1130 audit(1765889261.159:296): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:41.171186 kernel: audit: type=1131 audit(1765889261.159:297): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:41.199554 systemd[1]: Reload requested from client PID 2328 ('systemctl') (unit session-8.scope)... Dec 16 12:47:41.199572 systemd[1]: Reloading... Dec 16 12:47:41.315808 zram_generator::config[2375]: No configuration found. Dec 16 12:47:41.489480 systemd[1]: Reloading finished in 289 ms. Dec 16 12:47:41.507000 audit: BPF prog-id=63 op=LOAD Dec 16 12:47:41.512054 kernel: audit: type=1334 audit(1765889261.507:298): prog-id=63 op=LOAD Dec 16 12:47:41.507000 audit: BPF prog-id=55 op=UNLOAD Dec 16 12:47:41.507000 audit: BPF prog-id=64 op=LOAD Dec 16 12:47:41.516349 kernel: audit: type=1334 audit(1765889261.507:299): prog-id=55 op=UNLOAD Dec 16 12:47:41.516414 kernel: audit: type=1334 audit(1765889261.507:300): prog-id=64 op=LOAD Dec 16 12:47:41.516449 kernel: audit: type=1334 audit(1765889261.507:301): prog-id=65 op=LOAD Dec 16 12:47:41.507000 audit: BPF prog-id=65 op=LOAD Dec 16 12:47:41.507000 audit: BPF prog-id=56 op=UNLOAD Dec 16 12:47:41.522778 kernel: audit: type=1334 audit(1765889261.507:302): prog-id=56 op=UNLOAD Dec 16 12:47:41.522828 kernel: audit: type=1334 audit(1765889261.507:303): prog-id=57 op=UNLOAD Dec 16 12:47:41.507000 audit: BPF prog-id=57 op=UNLOAD Dec 16 12:47:41.507000 audit: BPF prog-id=66 op=LOAD Dec 16 12:47:41.524375 kernel: audit: type=1334 audit(1765889261.507:304): prog-id=66 op=LOAD Dec 16 12:47:41.508000 audit: BPF prog-id=43 op=UNLOAD Dec 16 12:47:41.528227 kernel: audit: type=1334 audit(1765889261.508:305): prog-id=43 op=UNLOAD Dec 16 12:47:41.528289 kernel: audit: type=1334 audit(1765889261.508:306): prog-id=67 op=LOAD Dec 16 12:47:41.508000 audit: BPF prog-id=67 op=LOAD Dec 16 12:47:41.532530 kernel: audit: type=1334 audit(1765889261.508:307): prog-id=68 op=LOAD Dec 16 12:47:41.508000 audit: BPF prog-id=68 op=LOAD Dec 16 12:47:41.508000 audit: BPF prog-id=44 op=UNLOAD Dec 16 12:47:41.508000 audit: BPF prog-id=45 op=UNLOAD Dec 16 12:47:41.508000 audit: BPF prog-id=69 op=LOAD Dec 16 12:47:41.508000 audit: BPF prog-id=52 op=UNLOAD Dec 16 12:47:41.508000 audit: BPF prog-id=70 op=LOAD Dec 16 12:47:41.508000 audit: BPF prog-id=71 op=LOAD Dec 16 12:47:41.508000 audit: BPF prog-id=53 op=UNLOAD Dec 16 12:47:41.508000 audit: BPF prog-id=54 op=UNLOAD Dec 16 12:47:41.510000 audit: BPF prog-id=72 op=LOAD Dec 16 12:47:41.510000 audit: BPF prog-id=60 op=UNLOAD Dec 16 12:47:41.510000 audit: BPF prog-id=73 op=LOAD Dec 16 12:47:41.510000 audit: BPF prog-id=74 op=LOAD Dec 16 12:47:41.510000 audit: BPF prog-id=61 op=UNLOAD Dec 16 12:47:41.510000 audit: BPF prog-id=62 op=UNLOAD Dec 16 12:47:41.519000 audit: BPF prog-id=75 op=LOAD Dec 16 12:47:41.519000 audit: BPF prog-id=49 op=UNLOAD Dec 16 12:47:41.519000 audit: BPF prog-id=76 op=LOAD Dec 16 12:47:41.519000 audit: BPF prog-id=46 op=UNLOAD Dec 16 12:47:41.519000 audit: BPF prog-id=77 op=LOAD Dec 16 12:47:41.519000 audit: BPF prog-id=78 op=LOAD Dec 16 12:47:41.519000 audit: BPF prog-id=47 op=UNLOAD Dec 16 12:47:41.519000 audit: BPF prog-id=48 op=UNLOAD Dec 16 12:47:41.520000 audit: BPF prog-id=79 op=LOAD Dec 16 12:47:41.520000 audit: BPF prog-id=58 op=UNLOAD Dec 16 12:47:41.522000 audit: BPF prog-id=80 op=LOAD Dec 16 12:47:41.522000 audit: BPF prog-id=59 op=UNLOAD Dec 16 12:47:41.522000 audit: BPF prog-id=81 op=LOAD Dec 16 12:47:41.522000 audit: BPF prog-id=82 op=LOAD Dec 16 12:47:41.522000 audit: BPF prog-id=50 op=UNLOAD Dec 16 12:47:41.522000 audit: BPF prog-id=51 op=UNLOAD Dec 16 12:47:41.539000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:41.542905 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 12:47:41.544710 systemd[1]: kubelet.service: Deactivated successfully. Dec 16 12:47:41.545038 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 12:47:41.544000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:41.545149 systemd[1]: kubelet.service: Consumed 92ms CPU time, 98.4M memory peak. Dec 16 12:47:41.546953 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 12:47:41.648103 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 12:47:41.647000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:41.655222 (kubelet)[2431]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Dec 16 12:47:41.714815 kubelet[2431]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 16 12:47:41.714815 kubelet[2431]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Dec 16 12:47:41.714815 kubelet[2431]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 16 12:47:41.715196 kubelet[2431]: I1216 12:47:41.714876 2431 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 16 12:47:41.909102 kubelet[2431]: I1216 12:47:41.908737 2431 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Dec 16 12:47:41.909102 kubelet[2431]: I1216 12:47:41.908763 2431 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 16 12:47:41.909354 kubelet[2431]: I1216 12:47:41.909340 2431 server.go:954] "Client rotation is on, will bootstrap in background" Dec 16 12:47:41.951173 kubelet[2431]: I1216 12:47:41.951136 2431 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 16 12:47:41.952442 kubelet[2431]: E1216 12:47:41.952395 2431 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://77.42.19.41:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 77.42.19.41:6443: connect: connection refused" logger="UnhandledError" Dec 16 12:47:41.965752 kubelet[2431]: I1216 12:47:41.965702 2431 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 16 12:47:41.971185 kubelet[2431]: I1216 12:47:41.971134 2431 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Dec 16 12:47:41.973216 kubelet[2431]: I1216 12:47:41.973168 2431 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 16 12:47:41.973451 kubelet[2431]: I1216 12:47:41.973207 2431 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4547-0-0-6-b8319fed62","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 16 12:47:41.975186 kubelet[2431]: I1216 12:47:41.975150 2431 topology_manager.go:138] "Creating topology manager with none policy" Dec 16 12:47:41.975186 kubelet[2431]: I1216 12:47:41.975172 2431 container_manager_linux.go:304] "Creating device plugin manager" Dec 16 12:47:41.976654 kubelet[2431]: I1216 12:47:41.976622 2431 state_mem.go:36] "Initialized new in-memory state store" Dec 16 12:47:41.980876 kubelet[2431]: I1216 12:47:41.980861 2431 kubelet.go:446] "Attempting to sync node with API server" Dec 16 12:47:41.980916 kubelet[2431]: I1216 12:47:41.980887 2431 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 16 12:47:41.982323 kubelet[2431]: I1216 12:47:41.982290 2431 kubelet.go:352] "Adding apiserver pod source" Dec 16 12:47:41.982323 kubelet[2431]: I1216 12:47:41.982317 2431 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 16 12:47:41.989149 kubelet[2431]: W1216 12:47:41.988948 2431 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://77.42.19.41:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4547-0-0-6-b8319fed62&limit=500&resourceVersion=0": dial tcp 77.42.19.41:6443: connect: connection refused Dec 16 12:47:41.989338 kubelet[2431]: E1216 12:47:41.989289 2431 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://77.42.19.41:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4547-0-0-6-b8319fed62&limit=500&resourceVersion=0\": dial tcp 77.42.19.41:6443: connect: connection refused" logger="UnhandledError" Dec 16 12:47:41.991092 kubelet[2431]: I1216 12:47:41.991058 2431 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.1.5" apiVersion="v1" Dec 16 12:47:41.994487 kubelet[2431]: I1216 12:47:41.994463 2431 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 16 12:47:41.994541 kubelet[2431]: W1216 12:47:41.994515 2431 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Dec 16 12:47:41.996570 kubelet[2431]: W1216 12:47:41.996464 2431 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://77.42.19.41:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 77.42.19.41:6443: connect: connection refused Dec 16 12:47:41.996570 kubelet[2431]: E1216 12:47:41.996505 2431 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://77.42.19.41:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 77.42.19.41:6443: connect: connection refused" logger="UnhandledError" Dec 16 12:47:41.996685 kubelet[2431]: I1216 12:47:41.996609 2431 watchdog_linux.go:99] "Systemd watchdog is not enabled" Dec 16 12:47:41.996685 kubelet[2431]: I1216 12:47:41.996638 2431 server.go:1287] "Started kubelet" Dec 16 12:47:41.997059 kubelet[2431]: I1216 12:47:41.996778 2431 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Dec 16 12:47:42.002072 kubelet[2431]: I1216 12:47:42.002005 2431 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 16 12:47:42.002510 kubelet[2431]: I1216 12:47:42.002473 2431 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 16 12:47:42.009140 kubelet[2431]: I1216 12:47:42.009109 2431 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 16 12:47:42.009737 kubelet[2431]: I1216 12:47:42.009705 2431 server.go:479] "Adding debug handlers to kubelet server" Dec 16 12:47:42.013359 kubelet[2431]: E1216 12:47:42.010428 2431 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://77.42.19.41:6443/api/v1/namespaces/default/events\": dial tcp 77.42.19.41:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4547-0-0-6-b8319fed62.1881b2ed7dfc2fb0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4547-0-0-6-b8319fed62,UID:ci-4547-0-0-6-b8319fed62,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4547-0-0-6-b8319fed62,},FirstTimestamp:2025-12-16 12:47:41.996617648 +0000 UTC m=+0.329358992,LastTimestamp:2025-12-16 12:47:41.996617648 +0000 UTC m=+0.329358992,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4547-0-0-6-b8319fed62,}" Dec 16 12:47:42.015159 kubelet[2431]: I1216 12:47:42.014052 2431 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Dec 16 12:47:42.015000 audit[2442]: NETFILTER_CFG table=mangle:42 family=2 entries=2 op=nft_register_chain pid=2442 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:42.015000 audit[2442]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7fff17b17650 a2=0 a3=0 items=0 ppid=2431 pid=2442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.015000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Dec 16 12:47:42.016000 audit[2443]: NETFILTER_CFG table=filter:43 family=2 entries=1 op=nft_register_chain pid=2443 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:42.016000 audit[2443]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc4cf9f860 a2=0 a3=0 items=0 ppid=2431 pid=2443 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.016000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 Dec 16 12:47:42.019683 kubelet[2431]: I1216 12:47:42.018119 2431 volume_manager.go:297] "Starting Kubelet Volume Manager" Dec 16 12:47:42.019683 kubelet[2431]: E1216 12:47:42.018296 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:42.018000 audit[2445]: NETFILTER_CFG table=filter:44 family=2 entries=2 op=nft_register_chain pid=2445 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:42.018000 audit[2445]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffd09ce0600 a2=0 a3=0 items=0 ppid=2431 pid=2445 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.018000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Dec 16 12:47:42.024413 kubelet[2431]: I1216 12:47:42.023340 2431 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Dec 16 12:47:42.024413 kubelet[2431]: I1216 12:47:42.023396 2431 reconciler.go:26] "Reconciler: start to sync state" Dec 16 12:47:42.026109 kubelet[2431]: E1216 12:47:42.025619 2431 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://77.42.19.41:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4547-0-0-6-b8319fed62?timeout=10s\": dial tcp 77.42.19.41:6443: connect: connection refused" interval="200ms" Dec 16 12:47:42.026109 kubelet[2431]: W1216 12:47:42.025962 2431 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://77.42.19.41:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 77.42.19.41:6443: connect: connection refused Dec 16 12:47:42.026109 kubelet[2431]: E1216 12:47:42.025987 2431 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://77.42.19.41:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 77.42.19.41:6443: connect: connection refused" logger="UnhandledError" Dec 16 12:47:42.027000 audit[2447]: NETFILTER_CFG table=filter:45 family=2 entries=2 op=nft_register_chain pid=2447 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:42.027000 audit[2447]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7fffe3bb9a10 a2=0 a3=0 items=0 ppid=2431 pid=2447 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.027000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Dec 16 12:47:42.030523 kubelet[2431]: I1216 12:47:42.030484 2431 factory.go:221] Registration of the systemd container factory successfully Dec 16 12:47:42.030580 kubelet[2431]: I1216 12:47:42.030562 2431 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 16 12:47:42.036535 kubelet[2431]: E1216 12:47:42.036507 2431 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Dec 16 12:47:42.037441 kubelet[2431]: I1216 12:47:42.037407 2431 factory.go:221] Registration of the containerd container factory successfully Dec 16 12:47:42.038000 audit[2450]: NETFILTER_CFG table=filter:46 family=2 entries=1 op=nft_register_rule pid=2450 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:42.038000 audit[2450]: SYSCALL arch=c000003e syscall=46 success=yes exit=924 a0=3 a1=7fff2d6a5670 a2=0 a3=0 items=0 ppid=2431 pid=2450 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.038000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D41004B5542452D4649524557414C4C002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E7400626C6F636B20696E636F6D696E67206C6F63616C6E657420636F6E6E656374696F6E73002D2D647374003132372E302E302E302F38 Dec 16 12:47:42.040869 kubelet[2431]: I1216 12:47:42.040826 2431 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 16 12:47:42.040000 audit[2451]: NETFILTER_CFG table=mangle:47 family=10 entries=2 op=nft_register_chain pid=2451 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:42.040000 audit[2451]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffd9a346ca0 a2=0 a3=0 items=0 ppid=2431 pid=2451 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.040000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Dec 16 12:47:42.042172 kubelet[2431]: I1216 12:47:42.042153 2431 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 16 12:47:42.042172 kubelet[2431]: I1216 12:47:42.042171 2431 status_manager.go:227] "Starting to sync pod status with apiserver" Dec 16 12:47:42.042221 kubelet[2431]: I1216 12:47:42.042192 2431 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Dec 16 12:47:42.042221 kubelet[2431]: I1216 12:47:42.042201 2431 kubelet.go:2382] "Starting kubelet main sync loop" Dec 16 12:47:42.042297 kubelet[2431]: E1216 12:47:42.042246 2431 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 16 12:47:42.042000 audit[2452]: NETFILTER_CFG table=mangle:48 family=2 entries=1 op=nft_register_chain pid=2452 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:42.042000 audit[2452]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffcd487cce0 a2=0 a3=0 items=0 ppid=2431 pid=2452 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.042000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Dec 16 12:47:42.045000 audit[2454]: NETFILTER_CFG table=nat:49 family=2 entries=1 op=nft_register_chain pid=2454 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:42.045000 audit[2454]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd8f278200 a2=0 a3=0 items=0 ppid=2431 pid=2454 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.045000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Dec 16 12:47:42.046000 audit[2455]: NETFILTER_CFG table=filter:50 family=2 entries=1 op=nft_register_chain pid=2455 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:42.046000 audit[2455]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe600bc520 a2=0 a3=0 items=0 ppid=2431 pid=2455 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.046000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Dec 16 12:47:42.047000 audit[2456]: NETFILTER_CFG table=mangle:51 family=10 entries=1 op=nft_register_chain pid=2456 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:42.047000 audit[2456]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffd648af0e0 a2=0 a3=0 items=0 ppid=2431 pid=2456 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.047000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Dec 16 12:47:42.048000 audit[2457]: NETFILTER_CFG table=nat:52 family=10 entries=1 op=nft_register_chain pid=2457 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:42.048000 audit[2457]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff3005a0f0 a2=0 a3=0 items=0 ppid=2431 pid=2457 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.048000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Dec 16 12:47:42.050000 audit[2458]: NETFILTER_CFG table=filter:53 family=10 entries=1 op=nft_register_chain pid=2458 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:42.050000 audit[2458]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fff179d3a70 a2=0 a3=0 items=0 ppid=2431 pid=2458 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.050000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Dec 16 12:47:42.054240 kubelet[2431]: W1216 12:47:42.054181 2431 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://77.42.19.41:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 77.42.19.41:6443: connect: connection refused Dec 16 12:47:42.054240 kubelet[2431]: E1216 12:47:42.054231 2431 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://77.42.19.41:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 77.42.19.41:6443: connect: connection refused" logger="UnhandledError" Dec 16 12:47:42.068543 kubelet[2431]: I1216 12:47:42.068517 2431 cpu_manager.go:221] "Starting CPU manager" policy="none" Dec 16 12:47:42.068543 kubelet[2431]: I1216 12:47:42.068530 2431 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Dec 16 12:47:42.068543 kubelet[2431]: I1216 12:47:42.068544 2431 state_mem.go:36] "Initialized new in-memory state store" Dec 16 12:47:42.073077 kubelet[2431]: I1216 12:47:42.073049 2431 policy_none.go:49] "None policy: Start" Dec 16 12:47:42.073077 kubelet[2431]: I1216 12:47:42.073066 2431 memory_manager.go:186] "Starting memorymanager" policy="None" Dec 16 12:47:42.073077 kubelet[2431]: I1216 12:47:42.073075 2431 state_mem.go:35] "Initializing new in-memory state store" Dec 16 12:47:42.079377 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Dec 16 12:47:42.091062 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Dec 16 12:47:42.094497 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Dec 16 12:47:42.103657 kubelet[2431]: I1216 12:47:42.103620 2431 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 16 12:47:42.103831 kubelet[2431]: I1216 12:47:42.103805 2431 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 16 12:47:42.103875 kubelet[2431]: I1216 12:47:42.103822 2431 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 16 12:47:42.104402 kubelet[2431]: I1216 12:47:42.104373 2431 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 16 12:47:42.105551 kubelet[2431]: E1216 12:47:42.105475 2431 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Dec 16 12:47:42.105551 kubelet[2431]: E1216 12:47:42.105513 2431 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:42.155299 systemd[1]: Created slice kubepods-burstable-pod453aa1d0fb03557a95eec686c6aa1b9e.slice - libcontainer container kubepods-burstable-pod453aa1d0fb03557a95eec686c6aa1b9e.slice. Dec 16 12:47:42.165253 kubelet[2431]: E1216 12:47:42.164980 2431 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4547-0-0-6-b8319fed62\" not found" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.170165 systemd[1]: Created slice kubepods-burstable-pod44dba4f6404faf0f772a9c1172df6052.slice - libcontainer container kubepods-burstable-pod44dba4f6404faf0f772a9c1172df6052.slice. Dec 16 12:47:42.173073 kubelet[2431]: E1216 12:47:42.172696 2431 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4547-0-0-6-b8319fed62\" not found" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.175681 systemd[1]: Created slice kubepods-burstable-podf48a6eef5f648efedb49b62f05055f93.slice - libcontainer container kubepods-burstable-podf48a6eef5f648efedb49b62f05055f93.slice. Dec 16 12:47:42.177711 kubelet[2431]: E1216 12:47:42.177659 2431 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4547-0-0-6-b8319fed62\" not found" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.207422 kubelet[2431]: I1216 12:47:42.207377 2431 kubelet_node_status.go:75] "Attempting to register node" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.209159 kubelet[2431]: E1216 12:47:42.209095 2431 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://77.42.19.41:6443/api/v1/nodes\": dial tcp 77.42.19.41:6443: connect: connection refused" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.226653 kubelet[2431]: E1216 12:47:42.226605 2431 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://77.42.19.41:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4547-0-0-6-b8319fed62?timeout=10s\": dial tcp 77.42.19.41:6443: connect: connection refused" interval="400ms" Dec 16 12:47:42.324460 kubelet[2431]: I1216 12:47:42.324406 2431 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/44dba4f6404faf0f772a9c1172df6052-ca-certs\") pod \"kube-controller-manager-ci-4547-0-0-6-b8319fed62\" (UID: \"44dba4f6404faf0f772a9c1172df6052\") " pod="kube-system/kube-controller-manager-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.324460 kubelet[2431]: I1216 12:47:42.324454 2431 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/44dba4f6404faf0f772a9c1172df6052-k8s-certs\") pod \"kube-controller-manager-ci-4547-0-0-6-b8319fed62\" (UID: \"44dba4f6404faf0f772a9c1172df6052\") " pod="kube-system/kube-controller-manager-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.324460 kubelet[2431]: I1216 12:47:42.324472 2431 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/44dba4f6404faf0f772a9c1172df6052-kubeconfig\") pod \"kube-controller-manager-ci-4547-0-0-6-b8319fed62\" (UID: \"44dba4f6404faf0f772a9c1172df6052\") " pod="kube-system/kube-controller-manager-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.324695 kubelet[2431]: I1216 12:47:42.324489 2431 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/44dba4f6404faf0f772a9c1172df6052-flexvolume-dir\") pod \"kube-controller-manager-ci-4547-0-0-6-b8319fed62\" (UID: \"44dba4f6404faf0f772a9c1172df6052\") " pod="kube-system/kube-controller-manager-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.324695 kubelet[2431]: I1216 12:47:42.324506 2431 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/44dba4f6404faf0f772a9c1172df6052-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4547-0-0-6-b8319fed62\" (UID: \"44dba4f6404faf0f772a9c1172df6052\") " pod="kube-system/kube-controller-manager-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.324695 kubelet[2431]: I1216 12:47:42.324526 2431 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f48a6eef5f648efedb49b62f05055f93-kubeconfig\") pod \"kube-scheduler-ci-4547-0-0-6-b8319fed62\" (UID: \"f48a6eef5f648efedb49b62f05055f93\") " pod="kube-system/kube-scheduler-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.324695 kubelet[2431]: I1216 12:47:42.324540 2431 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/453aa1d0fb03557a95eec686c6aa1b9e-ca-certs\") pod \"kube-apiserver-ci-4547-0-0-6-b8319fed62\" (UID: \"453aa1d0fb03557a95eec686c6aa1b9e\") " pod="kube-system/kube-apiserver-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.324695 kubelet[2431]: I1216 12:47:42.324554 2431 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/453aa1d0fb03557a95eec686c6aa1b9e-k8s-certs\") pod \"kube-apiserver-ci-4547-0-0-6-b8319fed62\" (UID: \"453aa1d0fb03557a95eec686c6aa1b9e\") " pod="kube-system/kube-apiserver-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.324806 kubelet[2431]: I1216 12:47:42.324568 2431 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/453aa1d0fb03557a95eec686c6aa1b9e-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4547-0-0-6-b8319fed62\" (UID: \"453aa1d0fb03557a95eec686c6aa1b9e\") " pod="kube-system/kube-apiserver-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.411432 kubelet[2431]: I1216 12:47:42.411383 2431 kubelet_node_status.go:75] "Attempting to register node" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.411724 kubelet[2431]: E1216 12:47:42.411671 2431 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://77.42.19.41:6443/api/v1/nodes\": dial tcp 77.42.19.41:6443: connect: connection refused" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.467255 containerd[1623]: time="2025-12-16T12:47:42.467156501Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4547-0-0-6-b8319fed62,Uid:453aa1d0fb03557a95eec686c6aa1b9e,Namespace:kube-system,Attempt:0,}" Dec 16 12:47:42.474008 containerd[1623]: time="2025-12-16T12:47:42.473884134Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4547-0-0-6-b8319fed62,Uid:44dba4f6404faf0f772a9c1172df6052,Namespace:kube-system,Attempt:0,}" Dec 16 12:47:42.478712 containerd[1623]: time="2025-12-16T12:47:42.478648817Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4547-0-0-6-b8319fed62,Uid:f48a6eef5f648efedb49b62f05055f93,Namespace:kube-system,Attempt:0,}" Dec 16 12:47:42.573890 containerd[1623]: time="2025-12-16T12:47:42.573686136Z" level=info msg="connecting to shim 572b02fe65e614ab942b1985a6f3b351518d0a7bc06d4764c65afa080e172397" address="unix:///run/containerd/s/3f5b2b7fb4c529f7a8b552685c156d93aca8f3164ee4e0635d9ffb58f5d23524" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:47:42.578567 containerd[1623]: time="2025-12-16T12:47:42.578211168Z" level=info msg="connecting to shim a4b11037aaa136258d84d308b5125184ee3d601f2045debbce68a868e343d15a" address="unix:///run/containerd/s/69557a881e3a77eba4a2735934dcb6011ee3355c49c3fae5455d922b75574085" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:47:42.582079 containerd[1623]: time="2025-12-16T12:47:42.582054054Z" level=info msg="connecting to shim 6efced880bf1f4a41b37c0727cf9666702e9894485c3ea01b86ebc95c98572a0" address="unix:///run/containerd/s/b5347f464cdd6618ef5e2005179b2352f9bf42ba5343acdd210f13b665b9f099" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:47:42.627189 kubelet[2431]: E1216 12:47:42.627148 2431 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://77.42.19.41:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4547-0-0-6-b8319fed62?timeout=10s\": dial tcp 77.42.19.41:6443: connect: connection refused" interval="800ms" Dec 16 12:47:42.664230 systemd[1]: Started cri-containerd-572b02fe65e614ab942b1985a6f3b351518d0a7bc06d4764c65afa080e172397.scope - libcontainer container 572b02fe65e614ab942b1985a6f3b351518d0a7bc06d4764c65afa080e172397. Dec 16 12:47:42.666142 systemd[1]: Started cri-containerd-6efced880bf1f4a41b37c0727cf9666702e9894485c3ea01b86ebc95c98572a0.scope - libcontainer container 6efced880bf1f4a41b37c0727cf9666702e9894485c3ea01b86ebc95c98572a0. Dec 16 12:47:42.668159 systemd[1]: Started cri-containerd-a4b11037aaa136258d84d308b5125184ee3d601f2045debbce68a868e343d15a.scope - libcontainer container a4b11037aaa136258d84d308b5125184ee3d601f2045debbce68a868e343d15a. Dec 16 12:47:42.687000 audit: BPF prog-id=83 op=LOAD Dec 16 12:47:42.687000 audit: BPF prog-id=84 op=LOAD Dec 16 12:47:42.687000 audit[2525]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=2499 pid=2525 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.687000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665666365643838306266316634613431623337633037323763663936 Dec 16 12:47:42.687000 audit: BPF prog-id=84 op=UNLOAD Dec 16 12:47:42.687000 audit[2525]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2499 pid=2525 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.687000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665666365643838306266316634613431623337633037323763663936 Dec 16 12:47:42.691000 audit: BPF prog-id=85 op=LOAD Dec 16 12:47:42.691000 audit[2525]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=2499 pid=2525 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.691000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665666365643838306266316634613431623337633037323763663936 Dec 16 12:47:42.691000 audit: BPF prog-id=86 op=LOAD Dec 16 12:47:42.691000 audit: BPF prog-id=87 op=LOAD Dec 16 12:47:42.691000 audit[2525]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=2499 pid=2525 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.691000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665666365643838306266316634613431623337633037323763663936 Dec 16 12:47:42.691000 audit: BPF prog-id=87 op=UNLOAD Dec 16 12:47:42.691000 audit[2525]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2499 pid=2525 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.691000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665666365643838306266316634613431623337633037323763663936 Dec 16 12:47:42.691000 audit: BPF prog-id=85 op=UNLOAD Dec 16 12:47:42.691000 audit[2525]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2499 pid=2525 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.691000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665666365643838306266316634613431623337633037323763663936 Dec 16 12:47:42.691000 audit: BPF prog-id=88 op=LOAD Dec 16 12:47:42.691000 audit[2525]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=2499 pid=2525 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.691000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665666365643838306266316634613431623337633037323763663936 Dec 16 12:47:42.691000 audit: BPF prog-id=89 op=LOAD Dec 16 12:47:42.691000 audit[2510]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=2477 pid=2510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.691000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3537326230326665363565363134616239343262313938356136663362 Dec 16 12:47:42.692000 audit: BPF prog-id=89 op=UNLOAD Dec 16 12:47:42.692000 audit[2510]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2477 pid=2510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.692000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3537326230326665363565363134616239343262313938356136663362 Dec 16 12:47:42.692000 audit: BPF prog-id=90 op=LOAD Dec 16 12:47:42.692000 audit[2510]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=2477 pid=2510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.692000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3537326230326665363565363134616239343262313938356136663362 Dec 16 12:47:42.692000 audit: BPF prog-id=91 op=LOAD Dec 16 12:47:42.692000 audit[2510]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=2477 pid=2510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.692000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3537326230326665363565363134616239343262313938356136663362 Dec 16 12:47:42.692000 audit: BPF prog-id=91 op=UNLOAD Dec 16 12:47:42.692000 audit[2510]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2477 pid=2510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.692000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3537326230326665363565363134616239343262313938356136663362 Dec 16 12:47:42.692000 audit: BPF prog-id=90 op=UNLOAD Dec 16 12:47:42.692000 audit[2510]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2477 pid=2510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.692000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3537326230326665363565363134616239343262313938356136663362 Dec 16 12:47:42.693000 audit: BPF prog-id=92 op=LOAD Dec 16 12:47:42.693000 audit[2510]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=2477 pid=2510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.693000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3537326230326665363565363134616239343262313938356136663362 Dec 16 12:47:42.698000 audit: BPF prog-id=93 op=LOAD Dec 16 12:47:42.698000 audit: BPF prog-id=94 op=LOAD Dec 16 12:47:42.698000 audit[2521]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=2492 pid=2521 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.698000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6134623131303337616161313336323538643834643330386235313235 Dec 16 12:47:42.698000 audit: BPF prog-id=94 op=UNLOAD Dec 16 12:47:42.698000 audit[2521]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2492 pid=2521 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.698000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6134623131303337616161313336323538643834643330386235313235 Dec 16 12:47:42.699000 audit: BPF prog-id=95 op=LOAD Dec 16 12:47:42.699000 audit[2521]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=2492 pid=2521 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.699000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6134623131303337616161313336323538643834643330386235313235 Dec 16 12:47:42.699000 audit: BPF prog-id=96 op=LOAD Dec 16 12:47:42.699000 audit[2521]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=2492 pid=2521 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.699000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6134623131303337616161313336323538643834643330386235313235 Dec 16 12:47:42.699000 audit: BPF prog-id=96 op=UNLOAD Dec 16 12:47:42.699000 audit[2521]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2492 pid=2521 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.699000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6134623131303337616161313336323538643834643330386235313235 Dec 16 12:47:42.699000 audit: BPF prog-id=95 op=UNLOAD Dec 16 12:47:42.699000 audit[2521]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2492 pid=2521 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.699000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6134623131303337616161313336323538643834643330386235313235 Dec 16 12:47:42.699000 audit: BPF prog-id=97 op=LOAD Dec 16 12:47:42.699000 audit[2521]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=2492 pid=2521 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.699000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6134623131303337616161313336323538643834643330386235313235 Dec 16 12:47:42.739151 containerd[1623]: time="2025-12-16T12:47:42.738709107Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4547-0-0-6-b8319fed62,Uid:453aa1d0fb03557a95eec686c6aa1b9e,Namespace:kube-system,Attempt:0,} returns sandbox id \"6efced880bf1f4a41b37c0727cf9666702e9894485c3ea01b86ebc95c98572a0\"" Dec 16 12:47:42.743277 containerd[1623]: time="2025-12-16T12:47:42.743246437Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4547-0-0-6-b8319fed62,Uid:44dba4f6404faf0f772a9c1172df6052,Namespace:kube-system,Attempt:0,} returns sandbox id \"572b02fe65e614ab942b1985a6f3b351518d0a7bc06d4764c65afa080e172397\"" Dec 16 12:47:42.745532 containerd[1623]: time="2025-12-16T12:47:42.745494475Z" level=info msg="CreateContainer within sandbox \"6efced880bf1f4a41b37c0727cf9666702e9894485c3ea01b86ebc95c98572a0\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Dec 16 12:47:42.747251 containerd[1623]: time="2025-12-16T12:47:42.747224987Z" level=info msg="CreateContainer within sandbox \"572b02fe65e614ab942b1985a6f3b351518d0a7bc06d4764c65afa080e172397\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Dec 16 12:47:42.758912 containerd[1623]: time="2025-12-16T12:47:42.758891161Z" level=info msg="Container a6add43b584d0e8ff6fb00f86c9e01071c4a920f2c24cc94c9d8965295505031: CDI devices from CRI Config.CDIDevices: []" Dec 16 12:47:42.760026 containerd[1623]: time="2025-12-16T12:47:42.759808454Z" level=info msg="Container 3ae4e2b63eb15af5aef407ce4eeca8aedf4b7487d0d825498439bad680b44862: CDI devices from CRI Config.CDIDevices: []" Dec 16 12:47:42.768724 containerd[1623]: time="2025-12-16T12:47:42.768644080Z" level=info msg="CreateContainer within sandbox \"6efced880bf1f4a41b37c0727cf9666702e9894485c3ea01b86ebc95c98572a0\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"3ae4e2b63eb15af5aef407ce4eeca8aedf4b7487d0d825498439bad680b44862\"" Dec 16 12:47:42.769464 containerd[1623]: time="2025-12-16T12:47:42.769446370Z" level=info msg="StartContainer for \"3ae4e2b63eb15af5aef407ce4eeca8aedf4b7487d0d825498439bad680b44862\"" Dec 16 12:47:42.770460 containerd[1623]: time="2025-12-16T12:47:42.770429507Z" level=info msg="CreateContainer within sandbox \"572b02fe65e614ab942b1985a6f3b351518d0a7bc06d4764c65afa080e172397\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"a6add43b584d0e8ff6fb00f86c9e01071c4a920f2c24cc94c9d8965295505031\"" Dec 16 12:47:42.770790 containerd[1623]: time="2025-12-16T12:47:42.770752414Z" level=info msg="StartContainer for \"a6add43b584d0e8ff6fb00f86c9e01071c4a920f2c24cc94c9d8965295505031\"" Dec 16 12:47:42.771557 containerd[1623]: time="2025-12-16T12:47:42.771508956Z" level=info msg="connecting to shim a6add43b584d0e8ff6fb00f86c9e01071c4a920f2c24cc94c9d8965295505031" address="unix:///run/containerd/s/3f5b2b7fb4c529f7a8b552685c156d93aca8f3164ee4e0635d9ffb58f5d23524" protocol=ttrpc version=3 Dec 16 12:47:42.772098 containerd[1623]: time="2025-12-16T12:47:42.772058420Z" level=info msg="connecting to shim 3ae4e2b63eb15af5aef407ce4eeca8aedf4b7487d0d825498439bad680b44862" address="unix:///run/containerd/s/b5347f464cdd6618ef5e2005179b2352f9bf42ba5343acdd210f13b665b9f099" protocol=ttrpc version=3 Dec 16 12:47:42.781183 containerd[1623]: time="2025-12-16T12:47:42.781077531Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4547-0-0-6-b8319fed62,Uid:f48a6eef5f648efedb49b62f05055f93,Namespace:kube-system,Attempt:0,} returns sandbox id \"a4b11037aaa136258d84d308b5125184ee3d601f2045debbce68a868e343d15a\"" Dec 16 12:47:42.787420 containerd[1623]: time="2025-12-16T12:47:42.787364411Z" level=info msg="CreateContainer within sandbox \"a4b11037aaa136258d84d308b5125184ee3d601f2045debbce68a868e343d15a\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Dec 16 12:47:42.793423 systemd[1]: Started cri-containerd-3ae4e2b63eb15af5aef407ce4eeca8aedf4b7487d0d825498439bad680b44862.scope - libcontainer container 3ae4e2b63eb15af5aef407ce4eeca8aedf4b7487d0d825498439bad680b44862. Dec 16 12:47:42.801054 containerd[1623]: time="2025-12-16T12:47:42.801002611Z" level=info msg="Container 3b63931a1b4c85b68b640f0164e2951f0c6e73d5afd99f62522a8b46d33182ed: CDI devices from CRI Config.CDIDevices: []" Dec 16 12:47:42.807174 containerd[1623]: time="2025-12-16T12:47:42.806740409Z" level=info msg="CreateContainer within sandbox \"a4b11037aaa136258d84d308b5125184ee3d601f2045debbce68a868e343d15a\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"3b63931a1b4c85b68b640f0164e2951f0c6e73d5afd99f62522a8b46d33182ed\"" Dec 16 12:47:42.807923 containerd[1623]: time="2025-12-16T12:47:42.807896759Z" level=info msg="StartContainer for \"3b63931a1b4c85b68b640f0164e2951f0c6e73d5afd99f62522a8b46d33182ed\"" Dec 16 12:47:42.808708 containerd[1623]: time="2025-12-16T12:47:42.808683345Z" level=info msg="connecting to shim 3b63931a1b4c85b68b640f0164e2951f0c6e73d5afd99f62522a8b46d33182ed" address="unix:///run/containerd/s/69557a881e3a77eba4a2735934dcb6011ee3355c49c3fae5455d922b75574085" protocol=ttrpc version=3 Dec 16 12:47:42.810177 systemd[1]: Started cri-containerd-a6add43b584d0e8ff6fb00f86c9e01071c4a920f2c24cc94c9d8965295505031.scope - libcontainer container a6add43b584d0e8ff6fb00f86c9e01071c4a920f2c24cc94c9d8965295505031. Dec 16 12:47:42.814755 kubelet[2431]: I1216 12:47:42.814732 2431 kubelet_node_status.go:75] "Attempting to register node" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.815162 kubelet[2431]: E1216 12:47:42.815110 2431 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://77.42.19.41:6443/api/v1/nodes\": dial tcp 77.42.19.41:6443: connect: connection refused" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:42.818000 audit: BPF prog-id=98 op=LOAD Dec 16 12:47:42.818000 audit: BPF prog-id=99 op=LOAD Dec 16 12:47:42.818000 audit[2600]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=2499 pid=2600 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.818000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3361653465326236336562313561663561656634303763653465656361 Dec 16 12:47:42.819000 audit: BPF prog-id=99 op=UNLOAD Dec 16 12:47:42.819000 audit[2600]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2499 pid=2600 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.819000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3361653465326236336562313561663561656634303763653465656361 Dec 16 12:47:42.819000 audit: BPF prog-id=100 op=LOAD Dec 16 12:47:42.819000 audit[2600]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=2499 pid=2600 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.819000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3361653465326236336562313561663561656634303763653465656361 Dec 16 12:47:42.819000 audit: BPF prog-id=101 op=LOAD Dec 16 12:47:42.819000 audit[2600]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=2499 pid=2600 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.819000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3361653465326236336562313561663561656634303763653465656361 Dec 16 12:47:42.819000 audit: BPF prog-id=101 op=UNLOAD Dec 16 12:47:42.819000 audit[2600]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2499 pid=2600 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.819000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3361653465326236336562313561663561656634303763653465656361 Dec 16 12:47:42.819000 audit: BPF prog-id=100 op=UNLOAD Dec 16 12:47:42.819000 audit[2600]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2499 pid=2600 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.819000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3361653465326236336562313561663561656634303763653465656361 Dec 16 12:47:42.819000 audit: BPF prog-id=102 op=LOAD Dec 16 12:47:42.819000 audit[2600]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=2499 pid=2600 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.819000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3361653465326236336562313561663561656634303763653465656361 Dec 16 12:47:42.833160 systemd[1]: Started cri-containerd-3b63931a1b4c85b68b640f0164e2951f0c6e73d5afd99f62522a8b46d33182ed.scope - libcontainer container 3b63931a1b4c85b68b640f0164e2951f0c6e73d5afd99f62522a8b46d33182ed. Dec 16 12:47:42.844000 audit: BPF prog-id=103 op=LOAD Dec 16 12:47:42.844000 audit: BPF prog-id=104 op=LOAD Dec 16 12:47:42.844000 audit[2601]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=2477 pid=2601 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.844000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6136616464343362353834643065386666366662303066383663396530 Dec 16 12:47:42.844000 audit: BPF prog-id=104 op=UNLOAD Dec 16 12:47:42.844000 audit[2601]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2477 pid=2601 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.844000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6136616464343362353834643065386666366662303066383663396530 Dec 16 12:47:42.844000 audit: BPF prog-id=105 op=LOAD Dec 16 12:47:42.844000 audit[2601]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=2477 pid=2601 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.844000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6136616464343362353834643065386666366662303066383663396530 Dec 16 12:47:42.844000 audit: BPF prog-id=106 op=LOAD Dec 16 12:47:42.844000 audit[2601]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=2477 pid=2601 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.844000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6136616464343362353834643065386666366662303066383663396530 Dec 16 12:47:42.844000 audit: BPF prog-id=106 op=UNLOAD Dec 16 12:47:42.844000 audit[2601]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2477 pid=2601 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.844000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6136616464343362353834643065386666366662303066383663396530 Dec 16 12:47:42.844000 audit: BPF prog-id=105 op=UNLOAD Dec 16 12:47:42.844000 audit[2601]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2477 pid=2601 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.844000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6136616464343362353834643065386666366662303066383663396530 Dec 16 12:47:42.844000 audit: BPF prog-id=107 op=LOAD Dec 16 12:47:42.844000 audit[2601]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=2477 pid=2601 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.844000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6136616464343362353834643065386666366662303066383663396530 Dec 16 12:47:42.850000 audit: BPF prog-id=108 op=LOAD Dec 16 12:47:42.852000 audit: BPF prog-id=109 op=LOAD Dec 16 12:47:42.852000 audit[2630]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=2492 pid=2630 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.852000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3362363339333161316234633835623638623634306630313634653239 Dec 16 12:47:42.852000 audit: BPF prog-id=109 op=UNLOAD Dec 16 12:47:42.852000 audit[2630]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2492 pid=2630 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.852000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3362363339333161316234633835623638623634306630313634653239 Dec 16 12:47:42.852000 audit: BPF prog-id=110 op=LOAD Dec 16 12:47:42.852000 audit[2630]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=2492 pid=2630 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.852000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3362363339333161316234633835623638623634306630313634653239 Dec 16 12:47:42.852000 audit: BPF prog-id=111 op=LOAD Dec 16 12:47:42.852000 audit[2630]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=2492 pid=2630 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.852000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3362363339333161316234633835623638623634306630313634653239 Dec 16 12:47:42.852000 audit: BPF prog-id=111 op=UNLOAD Dec 16 12:47:42.852000 audit[2630]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2492 pid=2630 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.852000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3362363339333161316234633835623638623634306630313634653239 Dec 16 12:47:42.852000 audit: BPF prog-id=110 op=UNLOAD Dec 16 12:47:42.852000 audit[2630]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2492 pid=2630 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.852000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3362363339333161316234633835623638623634306630313634653239 Dec 16 12:47:42.852000 audit: BPF prog-id=112 op=LOAD Dec 16 12:47:42.852000 audit[2630]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=2492 pid=2630 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:42.852000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3362363339333161316234633835623638623634306630313634653239 Dec 16 12:47:42.882228 containerd[1623]: time="2025-12-16T12:47:42.882191431Z" level=info msg="StartContainer for \"3ae4e2b63eb15af5aef407ce4eeca8aedf4b7487d0d825498439bad680b44862\" returns successfully" Dec 16 12:47:42.894524 containerd[1623]: time="2025-12-16T12:47:42.894494104Z" level=info msg="StartContainer for \"a6add43b584d0e8ff6fb00f86c9e01071c4a920f2c24cc94c9d8965295505031\" returns successfully" Dec 16 12:47:42.915572 containerd[1623]: time="2025-12-16T12:47:42.915434430Z" level=info msg="StartContainer for \"3b63931a1b4c85b68b640f0164e2951f0c6e73d5afd99f62522a8b46d33182ed\" returns successfully" Dec 16 12:47:43.038291 kubelet[2431]: W1216 12:47:43.038212 2431 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://77.42.19.41:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 77.42.19.41:6443: connect: connection refused Dec 16 12:47:43.038291 kubelet[2431]: E1216 12:47:43.038272 2431 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://77.42.19.41:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 77.42.19.41:6443: connect: connection refused" logger="UnhandledError" Dec 16 12:47:43.060814 kubelet[2431]: E1216 12:47:43.060570 2431 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4547-0-0-6-b8319fed62\" not found" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:43.067037 kubelet[2431]: E1216 12:47:43.066376 2431 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4547-0-0-6-b8319fed62\" not found" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:43.068854 kubelet[2431]: E1216 12:47:43.068834 2431 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4547-0-0-6-b8319fed62\" not found" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:43.177409 kubelet[2431]: W1216 12:47:43.177356 2431 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://77.42.19.41:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 77.42.19.41:6443: connect: connection refused Dec 16 12:47:43.177409 kubelet[2431]: E1216 12:47:43.177419 2431 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://77.42.19.41:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 77.42.19.41:6443: connect: connection refused" logger="UnhandledError" Dec 16 12:47:43.210979 kubelet[2431]: W1216 12:47:43.210934 2431 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://77.42.19.41:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4547-0-0-6-b8319fed62&limit=500&resourceVersion=0": dial tcp 77.42.19.41:6443: connect: connection refused Dec 16 12:47:43.211080 kubelet[2431]: E1216 12:47:43.210993 2431 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://77.42.19.41:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4547-0-0-6-b8319fed62&limit=500&resourceVersion=0\": dial tcp 77.42.19.41:6443: connect: connection refused" logger="UnhandledError" Dec 16 12:47:43.618220 kubelet[2431]: I1216 12:47:43.618192 2431 kubelet_node_status.go:75] "Attempting to register node" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:44.071623 kubelet[2431]: E1216 12:47:44.071587 2431 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4547-0-0-6-b8319fed62\" not found" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:44.072913 kubelet[2431]: E1216 12:47:44.072882 2431 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4547-0-0-6-b8319fed62\" not found" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:44.582717 kubelet[2431]: E1216 12:47:44.582674 2431 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4547-0-0-6-b8319fed62\" not found" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:44.731568 kubelet[2431]: I1216 12:47:44.731533 2431 kubelet_node_status.go:78] "Successfully registered node" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:44.731568 kubelet[2431]: E1216 12:47:44.731569 2431 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ci-4547-0-0-6-b8319fed62\": node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:44.754071 kubelet[2431]: E1216 12:47:44.754048 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:44.855116 kubelet[2431]: E1216 12:47:44.854955 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:44.955214 kubelet[2431]: E1216 12:47:44.955166 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:45.055610 kubelet[2431]: E1216 12:47:45.055529 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:45.156265 kubelet[2431]: E1216 12:47:45.156144 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:45.208059 kubelet[2431]: E1216 12:47:45.207236 2431 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4547-0-0-6-b8319fed62\" not found" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:45.256904 kubelet[2431]: E1216 12:47:45.256867 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:45.279055 kubelet[2431]: E1216 12:47:45.279005 2431 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4547-0-0-6-b8319fed62\" not found" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:45.357655 kubelet[2431]: E1216 12:47:45.357617 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:45.457916 kubelet[2431]: E1216 12:47:45.457792 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:45.558234 kubelet[2431]: E1216 12:47:45.558174 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:45.659509 kubelet[2431]: E1216 12:47:45.659438 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:45.760216 kubelet[2431]: E1216 12:47:45.760169 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:45.860550 kubelet[2431]: E1216 12:47:45.860504 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:45.961524 kubelet[2431]: E1216 12:47:45.961489 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:46.062504 kubelet[2431]: E1216 12:47:46.062375 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:46.162916 kubelet[2431]: E1216 12:47:46.162875 2431 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4547-0-0-6-b8319fed62\" not found" Dec 16 12:47:46.219011 kubelet[2431]: I1216 12:47:46.218977 2431 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:46.236042 kubelet[2431]: I1216 12:47:46.235677 2431 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:46.243970 kubelet[2431]: I1216 12:47:46.243944 2431 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:46.617708 systemd[1]: Reload requested from client PID 2698 ('systemctl') (unit session-8.scope)... Dec 16 12:47:46.617726 systemd[1]: Reloading... Dec 16 12:47:46.707072 zram_generator::config[2746]: No configuration found. Dec 16 12:47:46.809044 kubelet[2431]: I1216 12:47:46.808680 2431 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:46.814258 kubelet[2431]: E1216 12:47:46.814225 2431 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4547-0-0-6-b8319fed62\" already exists" pod="kube-system/kube-apiserver-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:46.898962 systemd[1]: Reloading finished in 280 ms. Dec 16 12:47:46.924349 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 12:47:46.938243 systemd[1]: kubelet.service: Deactivated successfully. Dec 16 12:47:46.938435 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 12:47:46.937000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:46.939005 systemd[1]: kubelet.service: Consumed 605ms CPU time, 129.1M memory peak. Dec 16 12:47:46.942671 kernel: kauditd_printk_skb: 201 callbacks suppressed Dec 16 12:47:46.942733 kernel: audit: type=1131 audit(1765889266.937:401): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:46.945141 kernel: audit: type=1334 audit(1765889266.943:402): prog-id=113 op=LOAD Dec 16 12:47:46.943000 audit: BPF prog-id=113 op=LOAD Dec 16 12:47:46.944264 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Dec 16 12:47:46.943000 audit: BPF prog-id=66 op=UNLOAD Dec 16 12:47:46.950077 kernel: audit: type=1334 audit(1765889266.943:403): prog-id=66 op=UNLOAD Dec 16 12:47:46.943000 audit: BPF prog-id=114 op=LOAD Dec 16 12:47:46.955377 kernel: audit: type=1334 audit(1765889266.943:404): prog-id=114 op=LOAD Dec 16 12:47:46.955425 kernel: audit: type=1334 audit(1765889266.943:405): prog-id=115 op=LOAD Dec 16 12:47:46.943000 audit: BPF prog-id=115 op=LOAD Dec 16 12:47:46.957961 kernel: audit: type=1334 audit(1765889266.943:406): prog-id=67 op=UNLOAD Dec 16 12:47:46.943000 audit: BPF prog-id=67 op=UNLOAD Dec 16 12:47:46.943000 audit: BPF prog-id=68 op=UNLOAD Dec 16 12:47:46.946000 audit: BPF prog-id=116 op=LOAD Dec 16 12:47:46.962654 kernel: audit: type=1334 audit(1765889266.943:407): prog-id=68 op=UNLOAD Dec 16 12:47:46.962687 kernel: audit: type=1334 audit(1765889266.946:408): prog-id=116 op=LOAD Dec 16 12:47:46.962727 kernel: audit: type=1334 audit(1765889266.946:409): prog-id=79 op=UNLOAD Dec 16 12:47:46.946000 audit: BPF prog-id=79 op=UNLOAD Dec 16 12:47:46.949000 audit: BPF prog-id=117 op=LOAD Dec 16 12:47:46.965091 kernel: audit: type=1334 audit(1765889266.949:410): prog-id=117 op=LOAD Dec 16 12:47:46.949000 audit: BPF prog-id=75 op=UNLOAD Dec 16 12:47:46.950000 audit: BPF prog-id=118 op=LOAD Dec 16 12:47:46.950000 audit: BPF prog-id=63 op=UNLOAD Dec 16 12:47:46.950000 audit: BPF prog-id=119 op=LOAD Dec 16 12:47:46.950000 audit: BPF prog-id=120 op=LOAD Dec 16 12:47:46.950000 audit: BPF prog-id=64 op=UNLOAD Dec 16 12:47:46.950000 audit: BPF prog-id=65 op=UNLOAD Dec 16 12:47:46.951000 audit: BPF prog-id=121 op=LOAD Dec 16 12:47:46.951000 audit: BPF prog-id=76 op=UNLOAD Dec 16 12:47:46.951000 audit: BPF prog-id=122 op=LOAD Dec 16 12:47:46.951000 audit: BPF prog-id=123 op=LOAD Dec 16 12:47:46.951000 audit: BPF prog-id=77 op=UNLOAD Dec 16 12:47:46.951000 audit: BPF prog-id=78 op=UNLOAD Dec 16 12:47:46.951000 audit: BPF prog-id=124 op=LOAD Dec 16 12:47:46.951000 audit: BPF prog-id=69 op=UNLOAD Dec 16 12:47:46.951000 audit: BPF prog-id=125 op=LOAD Dec 16 12:47:46.952000 audit: BPF prog-id=126 op=LOAD Dec 16 12:47:46.952000 audit: BPF prog-id=70 op=UNLOAD Dec 16 12:47:46.952000 audit: BPF prog-id=71 op=UNLOAD Dec 16 12:47:46.953000 audit: BPF prog-id=127 op=LOAD Dec 16 12:47:46.953000 audit: BPF prog-id=72 op=UNLOAD Dec 16 12:47:46.953000 audit: BPF prog-id=128 op=LOAD Dec 16 12:47:46.953000 audit: BPF prog-id=129 op=LOAD Dec 16 12:47:46.953000 audit: BPF prog-id=73 op=UNLOAD Dec 16 12:47:46.953000 audit: BPF prog-id=74 op=UNLOAD Dec 16 12:47:46.953000 audit: BPF prog-id=130 op=LOAD Dec 16 12:47:46.953000 audit: BPF prog-id=131 op=LOAD Dec 16 12:47:46.953000 audit: BPF prog-id=81 op=UNLOAD Dec 16 12:47:46.953000 audit: BPF prog-id=82 op=UNLOAD Dec 16 12:47:46.954000 audit: BPF prog-id=132 op=LOAD Dec 16 12:47:46.954000 audit: BPF prog-id=80 op=UNLOAD Dec 16 12:47:47.045892 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Dec 16 12:47:47.046000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:47:47.053249 (kubelet)[2796]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Dec 16 12:47:47.110150 kubelet[2796]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 16 12:47:47.110150 kubelet[2796]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Dec 16 12:47:47.110150 kubelet[2796]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 16 12:47:47.110150 kubelet[2796]: I1216 12:47:47.109835 2796 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 16 12:47:47.115605 kubelet[2796]: I1216 12:47:47.115576 2796 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Dec 16 12:47:47.115605 kubelet[2796]: I1216 12:47:47.115596 2796 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 16 12:47:47.115807 kubelet[2796]: I1216 12:47:47.115787 2796 server.go:954] "Client rotation is on, will bootstrap in background" Dec 16 12:47:47.116987 kubelet[2796]: I1216 12:47:47.116962 2796 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 16 12:47:47.122332 kubelet[2796]: I1216 12:47:47.121926 2796 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 16 12:47:47.125613 kubelet[2796]: I1216 12:47:47.125602 2796 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 16 12:47:47.128585 kubelet[2796]: I1216 12:47:47.128571 2796 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Dec 16 12:47:47.128788 kubelet[2796]: I1216 12:47:47.128760 2796 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 16 12:47:47.129081 kubelet[2796]: I1216 12:47:47.128839 2796 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4547-0-0-6-b8319fed62","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 16 12:47:47.129233 kubelet[2796]: I1216 12:47:47.129223 2796 topology_manager.go:138] "Creating topology manager with none policy" Dec 16 12:47:47.129395 kubelet[2796]: I1216 12:47:47.129385 2796 container_manager_linux.go:304] "Creating device plugin manager" Dec 16 12:47:47.129493 kubelet[2796]: I1216 12:47:47.129484 2796 state_mem.go:36] "Initialized new in-memory state store" Dec 16 12:47:47.129657 kubelet[2796]: I1216 12:47:47.129647 2796 kubelet.go:446] "Attempting to sync node with API server" Dec 16 12:47:47.129714 kubelet[2796]: I1216 12:47:47.129707 2796 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 16 12:47:47.129775 kubelet[2796]: I1216 12:47:47.129768 2796 kubelet.go:352] "Adding apiserver pod source" Dec 16 12:47:47.129820 kubelet[2796]: I1216 12:47:47.129814 2796 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 16 12:47:47.131778 kubelet[2796]: I1216 12:47:47.131760 2796 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.1.5" apiVersion="v1" Dec 16 12:47:47.133919 kubelet[2796]: I1216 12:47:47.133897 2796 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 16 12:47:47.134310 kubelet[2796]: I1216 12:47:47.134294 2796 watchdog_linux.go:99] "Systemd watchdog is not enabled" Dec 16 12:47:47.134348 kubelet[2796]: I1216 12:47:47.134324 2796 server.go:1287] "Started kubelet" Dec 16 12:47:47.137729 kubelet[2796]: E1216 12:47:47.137685 2796 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Dec 16 12:47:47.137845 kubelet[2796]: I1216 12:47:47.137818 2796 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 16 12:47:47.138249 kubelet[2796]: I1216 12:47:47.138237 2796 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 16 12:47:47.138332 kubelet[2796]: I1216 12:47:47.138318 2796 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Dec 16 12:47:47.138627 kubelet[2796]: I1216 12:47:47.138607 2796 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 16 12:47:47.141038 kubelet[2796]: I1216 12:47:47.139769 2796 server.go:479] "Adding debug handlers to kubelet server" Dec 16 12:47:47.144430 kubelet[2796]: I1216 12:47:47.144417 2796 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Dec 16 12:47:47.146652 kubelet[2796]: I1216 12:47:47.146633 2796 volume_manager.go:297] "Starting Kubelet Volume Manager" Dec 16 12:47:47.149851 kubelet[2796]: I1216 12:47:47.149794 2796 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Dec 16 12:47:47.150108 kubelet[2796]: I1216 12:47:47.150088 2796 reconciler.go:26] "Reconciler: start to sync state" Dec 16 12:47:47.151579 kubelet[2796]: I1216 12:47:47.151562 2796 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 16 12:47:47.152118 kubelet[2796]: I1216 12:47:47.152091 2796 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 16 12:47:47.152920 kubelet[2796]: I1216 12:47:47.152901 2796 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 16 12:47:47.152955 kubelet[2796]: I1216 12:47:47.152925 2796 status_manager.go:227] "Starting to sync pod status with apiserver" Dec 16 12:47:47.152955 kubelet[2796]: I1216 12:47:47.152940 2796 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Dec 16 12:47:47.152955 kubelet[2796]: I1216 12:47:47.152945 2796 kubelet.go:2382] "Starting kubelet main sync loop" Dec 16 12:47:47.153013 kubelet[2796]: E1216 12:47:47.152974 2796 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 16 12:47:47.154423 kubelet[2796]: I1216 12:47:47.154411 2796 factory.go:221] Registration of the containerd container factory successfully Dec 16 12:47:47.154472 kubelet[2796]: I1216 12:47:47.154466 2796 factory.go:221] Registration of the systemd container factory successfully Dec 16 12:47:47.207223 kubelet[2796]: I1216 12:47:47.207187 2796 cpu_manager.go:221] "Starting CPU manager" policy="none" Dec 16 12:47:47.207223 kubelet[2796]: I1216 12:47:47.207230 2796 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Dec 16 12:47:47.207355 kubelet[2796]: I1216 12:47:47.207244 2796 state_mem.go:36] "Initialized new in-memory state store" Dec 16 12:47:47.207389 kubelet[2796]: I1216 12:47:47.207354 2796 state_mem.go:88] "Updated default CPUSet" cpuSet="" Dec 16 12:47:47.207389 kubelet[2796]: I1216 12:47:47.207362 2796 state_mem.go:96] "Updated CPUSet assignments" assignments={} Dec 16 12:47:47.207389 kubelet[2796]: I1216 12:47:47.207390 2796 policy_none.go:49] "None policy: Start" Dec 16 12:47:47.207459 kubelet[2796]: I1216 12:47:47.207397 2796 memory_manager.go:186] "Starting memorymanager" policy="None" Dec 16 12:47:47.207459 kubelet[2796]: I1216 12:47:47.207405 2796 state_mem.go:35] "Initializing new in-memory state store" Dec 16 12:47:47.207504 kubelet[2796]: I1216 12:47:47.207479 2796 state_mem.go:75] "Updated machine memory state" Dec 16 12:47:47.211266 kubelet[2796]: I1216 12:47:47.211235 2796 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 16 12:47:47.211790 kubelet[2796]: I1216 12:47:47.211673 2796 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 16 12:47:47.211790 kubelet[2796]: I1216 12:47:47.211733 2796 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 16 12:47:47.212097 kubelet[2796]: I1216 12:47:47.212078 2796 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 16 12:47:47.215888 kubelet[2796]: E1216 12:47:47.214068 2796 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Dec 16 12:47:47.253756 kubelet[2796]: I1216 12:47:47.253720 2796 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.254879 kubelet[2796]: I1216 12:47:47.254846 2796 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.255007 kubelet[2796]: I1216 12:47:47.254994 2796 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.261246 kubelet[2796]: E1216 12:47:47.261220 2796 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4547-0-0-6-b8319fed62\" already exists" pod="kube-system/kube-scheduler-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.262006 kubelet[2796]: E1216 12:47:47.261986 2796 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4547-0-0-6-b8319fed62\" already exists" pod="kube-system/kube-apiserver-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.262170 kubelet[2796]: E1216 12:47:47.262113 2796 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4547-0-0-6-b8319fed62\" already exists" pod="kube-system/kube-controller-manager-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.320032 kubelet[2796]: I1216 12:47:47.319977 2796 kubelet_node_status.go:75] "Attempting to register node" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.327618 kubelet[2796]: I1216 12:47:47.327565 2796 kubelet_node_status.go:124] "Node was previously registered" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.327910 kubelet[2796]: I1216 12:47:47.327898 2796 kubelet_node_status.go:78] "Successfully registered node" node="ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.352338 kubelet[2796]: I1216 12:47:47.352300 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/44dba4f6404faf0f772a9c1172df6052-flexvolume-dir\") pod \"kube-controller-manager-ci-4547-0-0-6-b8319fed62\" (UID: \"44dba4f6404faf0f772a9c1172df6052\") " pod="kube-system/kube-controller-manager-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.352338 kubelet[2796]: I1216 12:47:47.352334 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f48a6eef5f648efedb49b62f05055f93-kubeconfig\") pod \"kube-scheduler-ci-4547-0-0-6-b8319fed62\" (UID: \"f48a6eef5f648efedb49b62f05055f93\") " pod="kube-system/kube-scheduler-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.352458 kubelet[2796]: I1216 12:47:47.352349 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/453aa1d0fb03557a95eec686c6aa1b9e-k8s-certs\") pod \"kube-apiserver-ci-4547-0-0-6-b8319fed62\" (UID: \"453aa1d0fb03557a95eec686c6aa1b9e\") " pod="kube-system/kube-apiserver-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.352458 kubelet[2796]: I1216 12:47:47.352365 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/453aa1d0fb03557a95eec686c6aa1b9e-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4547-0-0-6-b8319fed62\" (UID: \"453aa1d0fb03557a95eec686c6aa1b9e\") " pod="kube-system/kube-apiserver-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.352458 kubelet[2796]: I1216 12:47:47.352381 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/44dba4f6404faf0f772a9c1172df6052-ca-certs\") pod \"kube-controller-manager-ci-4547-0-0-6-b8319fed62\" (UID: \"44dba4f6404faf0f772a9c1172df6052\") " pod="kube-system/kube-controller-manager-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.352458 kubelet[2796]: I1216 12:47:47.352395 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/44dba4f6404faf0f772a9c1172df6052-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4547-0-0-6-b8319fed62\" (UID: \"44dba4f6404faf0f772a9c1172df6052\") " pod="kube-system/kube-controller-manager-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.352458 kubelet[2796]: I1216 12:47:47.352410 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/453aa1d0fb03557a95eec686c6aa1b9e-ca-certs\") pod \"kube-apiserver-ci-4547-0-0-6-b8319fed62\" (UID: \"453aa1d0fb03557a95eec686c6aa1b9e\") " pod="kube-system/kube-apiserver-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.352570 kubelet[2796]: I1216 12:47:47.352423 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/44dba4f6404faf0f772a9c1172df6052-k8s-certs\") pod \"kube-controller-manager-ci-4547-0-0-6-b8319fed62\" (UID: \"44dba4f6404faf0f772a9c1172df6052\") " pod="kube-system/kube-controller-manager-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:47.352570 kubelet[2796]: I1216 12:47:47.352437 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/44dba4f6404faf0f772a9c1172df6052-kubeconfig\") pod \"kube-controller-manager-ci-4547-0-0-6-b8319fed62\" (UID: \"44dba4f6404faf0f772a9c1172df6052\") " pod="kube-system/kube-controller-manager-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:48.131041 kubelet[2796]: I1216 12:47:48.130932 2796 apiserver.go:52] "Watching apiserver" Dec 16 12:47:48.150866 kubelet[2796]: I1216 12:47:48.150805 2796 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Dec 16 12:47:48.192272 kubelet[2796]: I1216 12:47:48.192232 2796 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:48.192515 kubelet[2796]: I1216 12:47:48.192489 2796 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:48.200906 kubelet[2796]: E1216 12:47:48.200869 2796 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4547-0-0-6-b8319fed62\" already exists" pod="kube-system/kube-apiserver-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:48.206507 kubelet[2796]: E1216 12:47:48.206485 2796 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4547-0-0-6-b8319fed62\" already exists" pod="kube-system/kube-scheduler-ci-4547-0-0-6-b8319fed62" Dec 16 12:47:48.226967 kubelet[2796]: I1216 12:47:48.226326 2796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4547-0-0-6-b8319fed62" podStartSLOduration=2.22631552 podStartE2EDuration="2.22631552s" podCreationTimestamp="2025-12-16 12:47:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-16 12:47:48.225631335 +0000 UTC m=+1.159359362" watchObservedRunningTime="2025-12-16 12:47:48.22631552 +0000 UTC m=+1.160043528" Dec 16 12:47:48.226967 kubelet[2796]: I1216 12:47:48.226409 2796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4547-0-0-6-b8319fed62" podStartSLOduration=2.226404109 podStartE2EDuration="2.226404109s" podCreationTimestamp="2025-12-16 12:47:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-16 12:47:48.219369634 +0000 UTC m=+1.153097651" watchObservedRunningTime="2025-12-16 12:47:48.226404109 +0000 UTC m=+1.160132126" Dec 16 12:47:48.236311 kubelet[2796]: I1216 12:47:48.234658 2796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4547-0-0-6-b8319fed62" podStartSLOduration=2.234638496 podStartE2EDuration="2.234638496s" podCreationTimestamp="2025-12-16 12:47:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-16 12:47:48.234613389 +0000 UTC m=+1.168341396" watchObservedRunningTime="2025-12-16 12:47:48.234638496 +0000 UTC m=+1.168366503" Dec 16 12:47:53.544833 kubelet[2796]: I1216 12:47:53.544756 2796 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Dec 16 12:47:53.547078 containerd[1623]: time="2025-12-16T12:47:53.546984056Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Dec 16 12:47:53.548468 kubelet[2796]: I1216 12:47:53.548125 2796 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Dec 16 12:47:54.186749 systemd[1]: Created slice kubepods-besteffort-pod397a3e9e_f1b9_4b91_b7c0_8067a7d61aa1.slice - libcontainer container kubepods-besteffort-pod397a3e9e_f1b9_4b91_b7c0_8067a7d61aa1.slice. Dec 16 12:47:54.197112 kubelet[2796]: I1216 12:47:54.196894 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/397a3e9e-f1b9-4b91-b7c0-8067a7d61aa1-xtables-lock\") pod \"kube-proxy-sfv8g\" (UID: \"397a3e9e-f1b9-4b91-b7c0-8067a7d61aa1\") " pod="kube-system/kube-proxy-sfv8g" Dec 16 12:47:54.197112 kubelet[2796]: I1216 12:47:54.196943 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/397a3e9e-f1b9-4b91-b7c0-8067a7d61aa1-lib-modules\") pod \"kube-proxy-sfv8g\" (UID: \"397a3e9e-f1b9-4b91-b7c0-8067a7d61aa1\") " pod="kube-system/kube-proxy-sfv8g" Dec 16 12:47:54.197112 kubelet[2796]: I1216 12:47:54.196974 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/397a3e9e-f1b9-4b91-b7c0-8067a7d61aa1-kube-proxy\") pod \"kube-proxy-sfv8g\" (UID: \"397a3e9e-f1b9-4b91-b7c0-8067a7d61aa1\") " pod="kube-system/kube-proxy-sfv8g" Dec 16 12:47:54.197112 kubelet[2796]: I1216 12:47:54.196998 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jnv4\" (UniqueName: \"kubernetes.io/projected/397a3e9e-f1b9-4b91-b7c0-8067a7d61aa1-kube-api-access-8jnv4\") pod \"kube-proxy-sfv8g\" (UID: \"397a3e9e-f1b9-4b91-b7c0-8067a7d61aa1\") " pod="kube-system/kube-proxy-sfv8g" Dec 16 12:47:54.502946 containerd[1623]: time="2025-12-16T12:47:54.501428601Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-sfv8g,Uid:397a3e9e-f1b9-4b91-b7c0-8067a7d61aa1,Namespace:kube-system,Attempt:0,}" Dec 16 12:47:54.548342 containerd[1623]: time="2025-12-16T12:47:54.548253026Z" level=info msg="connecting to shim e8fddd23fd190eb5269f532b8800325705e7912071a1f6d16a736e908274d42e" address="unix:///run/containerd/s/ccbefa3fa922f86384503257e3267b865df601a9790f285e3d50e188cb762455" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:47:54.592472 systemd[1]: Created slice kubepods-besteffort-pod370be85b_2fd6_4180_8bad_15f66b0dbf4c.slice - libcontainer container kubepods-besteffort-pod370be85b_2fd6_4180_8bad_15f66b0dbf4c.slice. Dec 16 12:47:54.603240 kubelet[2796]: I1216 12:47:54.603211 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/370be85b-2fd6-4180-8bad-15f66b0dbf4c-var-lib-calico\") pod \"tigera-operator-7dcd859c48-58brr\" (UID: \"370be85b-2fd6-4180-8bad-15f66b0dbf4c\") " pod="tigera-operator/tigera-operator-7dcd859c48-58brr" Dec 16 12:47:54.604137 kubelet[2796]: I1216 12:47:54.603248 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x669\" (UniqueName: \"kubernetes.io/projected/370be85b-2fd6-4180-8bad-15f66b0dbf4c-kube-api-access-4x669\") pod \"tigera-operator-7dcd859c48-58brr\" (UID: \"370be85b-2fd6-4180-8bad-15f66b0dbf4c\") " pod="tigera-operator/tigera-operator-7dcd859c48-58brr" Dec 16 12:47:54.604299 systemd[1]: Started cri-containerd-e8fddd23fd190eb5269f532b8800325705e7912071a1f6d16a736e908274d42e.scope - libcontainer container e8fddd23fd190eb5269f532b8800325705e7912071a1f6d16a736e908274d42e. Dec 16 12:47:54.618063 kernel: kauditd_printk_skb: 32 callbacks suppressed Dec 16 12:47:54.618135 kernel: audit: type=1334 audit(1765889274.614:443): prog-id=133 op=LOAD Dec 16 12:47:54.614000 audit: BPF prog-id=133 op=LOAD Dec 16 12:47:54.615000 audit: BPF prog-id=134 op=LOAD Dec 16 12:47:54.621635 kernel: audit: type=1334 audit(1765889274.615:444): prog-id=134 op=LOAD Dec 16 12:47:54.621686 kernel: audit: type=1300 audit(1765889274.615:444): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=2848 pid=2859 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.615000 audit[2859]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=2848 pid=2859 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.615000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538666464643233666431393065623532363966353332623838303033 Dec 16 12:47:54.629219 kernel: audit: type=1327 audit(1765889274.615:444): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538666464643233666431393065623532363966353332623838303033 Dec 16 12:47:54.615000 audit: BPF prog-id=134 op=UNLOAD Dec 16 12:47:54.615000 audit[2859]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2848 pid=2859 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.638639 kernel: audit: type=1334 audit(1765889274.615:445): prog-id=134 op=UNLOAD Dec 16 12:47:54.638736 kernel: audit: type=1300 audit(1765889274.615:445): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2848 pid=2859 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.615000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538666464643233666431393065623532363966353332623838303033 Dec 16 12:47:54.645360 kernel: audit: type=1327 audit(1765889274.615:445): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538666464643233666431393065623532363966353332623838303033 Dec 16 12:47:54.615000 audit: BPF prog-id=135 op=LOAD Dec 16 12:47:54.652387 kernel: audit: type=1334 audit(1765889274.615:446): prog-id=135 op=LOAD Dec 16 12:47:54.615000 audit[2859]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=2848 pid=2859 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.654198 containerd[1623]: time="2025-12-16T12:47:54.654168059Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-sfv8g,Uid:397a3e9e-f1b9-4b91-b7c0-8067a7d61aa1,Namespace:kube-system,Attempt:0,} returns sandbox id \"e8fddd23fd190eb5269f532b8800325705e7912071a1f6d16a736e908274d42e\"" Dec 16 12:47:54.654434 kernel: audit: type=1300 audit(1765889274.615:446): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=2848 pid=2859 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.658112 containerd[1623]: time="2025-12-16T12:47:54.658090624Z" level=info msg="CreateContainer within sandbox \"e8fddd23fd190eb5269f532b8800325705e7912071a1f6d16a736e908274d42e\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Dec 16 12:47:54.615000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538666464643233666431393065623532363966353332623838303033 Dec 16 12:47:54.668349 kernel: audit: type=1327 audit(1765889274.615:446): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538666464643233666431393065623532363966353332623838303033 Dec 16 12:47:54.615000 audit: BPF prog-id=136 op=LOAD Dec 16 12:47:54.615000 audit[2859]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=2848 pid=2859 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.615000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538666464643233666431393065623532363966353332623838303033 Dec 16 12:47:54.615000 audit: BPF prog-id=136 op=UNLOAD Dec 16 12:47:54.615000 audit[2859]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2848 pid=2859 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.615000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538666464643233666431393065623532363966353332623838303033 Dec 16 12:47:54.615000 audit: BPF prog-id=135 op=UNLOAD Dec 16 12:47:54.615000 audit[2859]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2848 pid=2859 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.615000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538666464643233666431393065623532363966353332623838303033 Dec 16 12:47:54.615000 audit: BPF prog-id=137 op=LOAD Dec 16 12:47:54.615000 audit[2859]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=2848 pid=2859 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.615000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6538666464643233666431393065623532363966353332623838303033 Dec 16 12:47:54.673730 containerd[1623]: time="2025-12-16T12:47:54.673622513Z" level=info msg="Container ce1dacf9caf0297b1f2c40254374d5a68e4f0eb551f00544c1a86865e60cee25: CDI devices from CRI Config.CDIDevices: []" Dec 16 12:47:54.682012 containerd[1623]: time="2025-12-16T12:47:54.681962110Z" level=info msg="CreateContainer within sandbox \"e8fddd23fd190eb5269f532b8800325705e7912071a1f6d16a736e908274d42e\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"ce1dacf9caf0297b1f2c40254374d5a68e4f0eb551f00544c1a86865e60cee25\"" Dec 16 12:47:54.683979 containerd[1623]: time="2025-12-16T12:47:54.682610411Z" level=info msg="StartContainer for \"ce1dacf9caf0297b1f2c40254374d5a68e4f0eb551f00544c1a86865e60cee25\"" Dec 16 12:47:54.684898 containerd[1623]: time="2025-12-16T12:47:54.684791100Z" level=info msg="connecting to shim ce1dacf9caf0297b1f2c40254374d5a68e4f0eb551f00544c1a86865e60cee25" address="unix:///run/containerd/s/ccbefa3fa922f86384503257e3267b865df601a9790f285e3d50e188cb762455" protocol=ttrpc version=3 Dec 16 12:47:54.702335 systemd[1]: Started cri-containerd-ce1dacf9caf0297b1f2c40254374d5a68e4f0eb551f00544c1a86865e60cee25.scope - libcontainer container ce1dacf9caf0297b1f2c40254374d5a68e4f0eb551f00544c1a86865e60cee25. Dec 16 12:47:54.745000 audit: BPF prog-id=138 op=LOAD Dec 16 12:47:54.745000 audit[2885]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=2848 pid=2885 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.745000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6365316461636639636166303239376231663263343032353433373464 Dec 16 12:47:54.746000 audit: BPF prog-id=139 op=LOAD Dec 16 12:47:54.746000 audit[2885]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=2848 pid=2885 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.746000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6365316461636639636166303239376231663263343032353433373464 Dec 16 12:47:54.746000 audit: BPF prog-id=139 op=UNLOAD Dec 16 12:47:54.746000 audit[2885]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2848 pid=2885 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.746000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6365316461636639636166303239376231663263343032353433373464 Dec 16 12:47:54.746000 audit: BPF prog-id=138 op=UNLOAD Dec 16 12:47:54.746000 audit[2885]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2848 pid=2885 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.746000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6365316461636639636166303239376231663263343032353433373464 Dec 16 12:47:54.746000 audit: BPF prog-id=140 op=LOAD Dec 16 12:47:54.746000 audit[2885]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=2848 pid=2885 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:54.746000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6365316461636639636166303239376231663263343032353433373464 Dec 16 12:47:54.768277 containerd[1623]: time="2025-12-16T12:47:54.768227983Z" level=info msg="StartContainer for \"ce1dacf9caf0297b1f2c40254374d5a68e4f0eb551f00544c1a86865e60cee25\" returns successfully" Dec 16 12:47:54.902081 containerd[1623]: time="2025-12-16T12:47:54.900748022Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-58brr,Uid:370be85b-2fd6-4180-8bad-15f66b0dbf4c,Namespace:tigera-operator,Attempt:0,}" Dec 16 12:47:54.932162 containerd[1623]: time="2025-12-16T12:47:54.931963730Z" level=info msg="connecting to shim 02ac242cef836574fb5d0643d85fa43d58200df961b34aef937073f2bcd94794" address="unix:///run/containerd/s/d13c141a964ed7933bf8b9828d04ff61c824a1d935f623eb0bedd7e1afc4c513" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:47:54.989462 systemd[1]: Started cri-containerd-02ac242cef836574fb5d0643d85fa43d58200df961b34aef937073f2bcd94794.scope - libcontainer container 02ac242cef836574fb5d0643d85fa43d58200df961b34aef937073f2bcd94794. Dec 16 12:47:55.024000 audit: BPF prog-id=141 op=LOAD Dec 16 12:47:55.025000 audit: BPF prog-id=142 op=LOAD Dec 16 12:47:55.025000 audit[2935]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=2924 pid=2935 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.025000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3032616332343263656638333635373466623564303634336438356661 Dec 16 12:47:55.025000 audit: BPF prog-id=142 op=UNLOAD Dec 16 12:47:55.025000 audit[2935]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2924 pid=2935 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.025000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3032616332343263656638333635373466623564303634336438356661 Dec 16 12:47:55.025000 audit: BPF prog-id=143 op=LOAD Dec 16 12:47:55.025000 audit[2935]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=2924 pid=2935 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.025000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3032616332343263656638333635373466623564303634336438356661 Dec 16 12:47:55.025000 audit: BPF prog-id=144 op=LOAD Dec 16 12:47:55.025000 audit[2935]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=2924 pid=2935 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.025000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3032616332343263656638333635373466623564303634336438356661 Dec 16 12:47:55.026000 audit: BPF prog-id=144 op=UNLOAD Dec 16 12:47:55.026000 audit[2935]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2924 pid=2935 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.026000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3032616332343263656638333635373466623564303634336438356661 Dec 16 12:47:55.026000 audit: BPF prog-id=143 op=UNLOAD Dec 16 12:47:55.026000 audit[2935]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2924 pid=2935 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.026000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3032616332343263656638333635373466623564303634336438356661 Dec 16 12:47:55.026000 audit: BPF prog-id=145 op=LOAD Dec 16 12:47:55.026000 audit[2935]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=2924 pid=2935 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.026000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3032616332343263656638333635373466623564303634336438356661 Dec 16 12:47:55.081580 containerd[1623]: time="2025-12-16T12:47:55.081466590Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-58brr,Uid:370be85b-2fd6-4180-8bad-15f66b0dbf4c,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"02ac242cef836574fb5d0643d85fa43d58200df961b34aef937073f2bcd94794\"" Dec 16 12:47:55.084782 containerd[1623]: time="2025-12-16T12:47:55.084738358Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\"" Dec 16 12:47:55.244000 audit[2993]: NETFILTER_CFG table=mangle:54 family=2 entries=1 op=nft_register_chain pid=2993 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.244000 audit[2993]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffdbd69bbb0 a2=0 a3=7ffdbd69bb9c items=0 ppid=2898 pid=2993 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.244000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Dec 16 12:47:55.253000 audit[2995]: NETFILTER_CFG table=mangle:55 family=10 entries=1 op=nft_register_chain pid=2995 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.253000 audit[2995]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fff6e311f90 a2=0 a3=7fff6e311f7c items=0 ppid=2898 pid=2995 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.253000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Dec 16 12:47:55.258000 audit[2996]: NETFILTER_CFG table=nat:56 family=2 entries=1 op=nft_register_chain pid=2996 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.258000 audit[2996]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffffce46830 a2=0 a3=7ffffce4681c items=0 ppid=2898 pid=2996 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.258000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Dec 16 12:47:55.265000 audit[2997]: NETFILTER_CFG table=nat:57 family=10 entries=1 op=nft_register_chain pid=2997 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.265000 audit[2997]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc44d1b340 a2=0 a3=7ffc44d1b32c items=0 ppid=2898 pid=2997 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.265000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Dec 16 12:47:55.269000 audit[2998]: NETFILTER_CFG table=filter:58 family=2 entries=1 op=nft_register_chain pid=2998 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.269000 audit[2998]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe04048bd0 a2=0 a3=7ffe04048bbc items=0 ppid=2898 pid=2998 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.269000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Dec 16 12:47:55.271000 audit[2999]: NETFILTER_CFG table=filter:59 family=10 entries=1 op=nft_register_chain pid=2999 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.271000 audit[2999]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe5a7049e0 a2=0 a3=7ffe5a7049cc items=0 ppid=2898 pid=2999 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.271000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Dec 16 12:47:55.336136 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount278511461.mount: Deactivated successfully. Dec 16 12:47:55.375000 audit[3000]: NETFILTER_CFG table=filter:60 family=2 entries=1 op=nft_register_chain pid=3000 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.375000 audit[3000]: SYSCALL arch=c000003e syscall=46 success=yes exit=108 a0=3 a1=7ffc886b0350 a2=0 a3=7ffc886b033c items=0 ppid=2898 pid=3000 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.375000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 Dec 16 12:47:55.379000 audit[3002]: NETFILTER_CFG table=filter:61 family=2 entries=1 op=nft_register_rule pid=3002 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.379000 audit[3002]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7fff9517fe40 a2=0 a3=7fff9517fe2c items=0 ppid=2898 pid=3002 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.379000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C652073657276696365 Dec 16 12:47:55.382000 audit[3005]: NETFILTER_CFG table=filter:62 family=2 entries=1 op=nft_register_rule pid=3005 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.382000 audit[3005]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffd876942a0 a2=0 a3=7ffd8769428c items=0 ppid=2898 pid=3005 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.382000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669 Dec 16 12:47:55.383000 audit[3006]: NETFILTER_CFG table=filter:63 family=2 entries=1 op=nft_register_chain pid=3006 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.383000 audit[3006]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffde15293a0 a2=0 a3=7ffde152938c items=0 ppid=2898 pid=3006 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.383000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4E4F4445504F525453002D740066696C746572 Dec 16 12:47:55.386000 audit[3008]: NETFILTER_CFG table=filter:64 family=2 entries=1 op=nft_register_rule pid=3008 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.386000 audit[3008]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffdb3083900 a2=0 a3=7ffdb30838ec items=0 ppid=2898 pid=3008 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.386000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 Dec 16 12:47:55.387000 audit[3009]: NETFILTER_CFG table=filter:65 family=2 entries=1 op=nft_register_chain pid=3009 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.387000 audit[3009]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd1ead4fc0 a2=0 a3=7ffd1ead4fac items=0 ppid=2898 pid=3009 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.387000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D740066696C746572 Dec 16 12:47:55.389000 audit[3011]: NETFILTER_CFG table=filter:66 family=2 entries=1 op=nft_register_rule pid=3011 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.389000 audit[3011]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7fff8bb54640 a2=0 a3=7fff8bb5462c items=0 ppid=2898 pid=3011 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.389000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D Dec 16 12:47:55.393000 audit[3014]: NETFILTER_CFG table=filter:67 family=2 entries=1 op=nft_register_rule pid=3014 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.393000 audit[3014]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7fff1d164d90 a2=0 a3=7fff1d164d7c items=0 ppid=2898 pid=3014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.393000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D53 Dec 16 12:47:55.394000 audit[3015]: NETFILTER_CFG table=filter:68 family=2 entries=1 op=nft_register_chain pid=3015 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.394000 audit[3015]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffeb7385990 a2=0 a3=7ffeb738597c items=0 ppid=2898 pid=3015 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.394000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D464F5257415244002D740066696C746572 Dec 16 12:47:55.396000 audit[3017]: NETFILTER_CFG table=filter:69 family=2 entries=1 op=nft_register_rule pid=3017 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.396000 audit[3017]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffc50b15d30 a2=0 a3=7ffc50b15d1c items=0 ppid=2898 pid=3017 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.396000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 Dec 16 12:47:55.397000 audit[3018]: NETFILTER_CFG table=filter:70 family=2 entries=1 op=nft_register_chain pid=3018 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.397000 audit[3018]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffc43598c50 a2=0 a3=7ffc43598c3c items=0 ppid=2898 pid=3018 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.397000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 Dec 16 12:47:55.401000 audit[3020]: NETFILTER_CFG table=filter:71 family=2 entries=1 op=nft_register_rule pid=3020 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.401000 audit[3020]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffc126cb490 a2=0 a3=7ffc126cb47c items=0 ppid=2898 pid=3020 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.401000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Dec 16 12:47:55.407000 audit[3023]: NETFILTER_CFG table=filter:72 family=2 entries=1 op=nft_register_rule pid=3023 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.407000 audit[3023]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffea8ef27e0 a2=0 a3=7ffea8ef27cc items=0 ppid=2898 pid=3023 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.407000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Dec 16 12:47:55.413000 audit[3026]: NETFILTER_CFG table=filter:73 family=2 entries=1 op=nft_register_rule pid=3026 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.413000 audit[3026]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffc99382de0 a2=0 a3=7ffc99382dcc items=0 ppid=2898 pid=3026 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.413000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D Dec 16 12:47:55.415000 audit[3027]: NETFILTER_CFG table=nat:74 family=2 entries=1 op=nft_register_chain pid=3027 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.415000 audit[3027]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffcf2789860 a2=0 a3=7ffcf278984c items=0 ppid=2898 pid=3027 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.415000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D74006E6174 Dec 16 12:47:55.419000 audit[3029]: NETFILTER_CFG table=nat:75 family=2 entries=1 op=nft_register_rule pid=3029 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.419000 audit[3029]: SYSCALL arch=c000003e syscall=46 success=yes exit=524 a0=3 a1=7ffe653360a0 a2=0 a3=7ffe6533608c items=0 ppid=2898 pid=3029 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.419000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Dec 16 12:47:55.425000 audit[3032]: NETFILTER_CFG table=nat:76 family=2 entries=1 op=nft_register_rule pid=3032 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.425000 audit[3032]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffc5fbf8f00 a2=0 a3=7ffc5fbf8eec items=0 ppid=2898 pid=3032 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.425000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Dec 16 12:47:55.427000 audit[3033]: NETFILTER_CFG table=nat:77 family=2 entries=1 op=nft_register_chain pid=3033 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.427000 audit[3033]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe50297bc0 a2=0 a3=7ffe50297bac items=0 ppid=2898 pid=3033 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.427000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Dec 16 12:47:55.432000 audit[3035]: NETFILTER_CFG table=nat:78 family=2 entries=1 op=nft_register_rule pid=3035 subj=system_u:system_r:kernel_t:s0 comm="iptables" Dec 16 12:47:55.432000 audit[3035]: SYSCALL arch=c000003e syscall=46 success=yes exit=532 a0=3 a1=7ffe166cf6a0 a2=0 a3=7ffe166cf68c items=0 ppid=2898 pid=3035 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.432000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 Dec 16 12:47:55.472000 audit[3041]: NETFILTER_CFG table=filter:79 family=2 entries=8 op=nft_register_rule pid=3041 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:47:55.472000 audit[3041]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7fff60e7bd70 a2=0 a3=7fff60e7bd5c items=0 ppid=2898 pid=3041 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.472000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:47:55.482000 audit[3041]: NETFILTER_CFG table=nat:80 family=2 entries=14 op=nft_register_chain pid=3041 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:47:55.482000 audit[3041]: SYSCALL arch=c000003e syscall=46 success=yes exit=5508 a0=3 a1=7fff60e7bd70 a2=0 a3=7fff60e7bd5c items=0 ppid=2898 pid=3041 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.482000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:47:55.485000 audit[3046]: NETFILTER_CFG table=filter:81 family=10 entries=1 op=nft_register_chain pid=3046 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.485000 audit[3046]: SYSCALL arch=c000003e syscall=46 success=yes exit=108 a0=3 a1=7ffe72f4b770 a2=0 a3=7ffe72f4b75c items=0 ppid=2898 pid=3046 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.485000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 Dec 16 12:47:55.490000 audit[3048]: NETFILTER_CFG table=filter:82 family=10 entries=2 op=nft_register_chain pid=3048 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.490000 audit[3048]: SYSCALL arch=c000003e syscall=46 success=yes exit=836 a0=3 a1=7fff469482c0 a2=0 a3=7fff469482ac items=0 ppid=2898 pid=3048 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.490000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C6520736572766963 Dec 16 12:47:55.500000 audit[3051]: NETFILTER_CFG table=filter:83 family=10 entries=1 op=nft_register_rule pid=3051 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.500000 audit[3051]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffdfb049f70 a2=0 a3=7ffdfb049f5c items=0 ppid=2898 pid=3051 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.500000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C652073657276 Dec 16 12:47:55.502000 audit[3052]: NETFILTER_CFG table=filter:84 family=10 entries=1 op=nft_register_chain pid=3052 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.502000 audit[3052]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffce3422f30 a2=0 a3=7ffce3422f1c items=0 ppid=2898 pid=3052 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.502000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4E4F4445504F525453002D740066696C746572 Dec 16 12:47:55.506000 audit[3054]: NETFILTER_CFG table=filter:85 family=10 entries=1 op=nft_register_rule pid=3054 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.506000 audit[3054]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffcbd1c4f30 a2=0 a3=7ffcbd1c4f1c items=0 ppid=2898 pid=3054 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.506000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 Dec 16 12:47:55.508000 audit[3055]: NETFILTER_CFG table=filter:86 family=10 entries=1 op=nft_register_chain pid=3055 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.508000 audit[3055]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff0b4b4a20 a2=0 a3=7fff0b4b4a0c items=0 ppid=2898 pid=3055 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.508000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D740066696C746572 Dec 16 12:47:55.513000 audit[3057]: NETFILTER_CFG table=filter:87 family=10 entries=1 op=nft_register_rule pid=3057 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.513000 audit[3057]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffd77fe21b0 a2=0 a3=7ffd77fe219c items=0 ppid=2898 pid=3057 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.513000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B554245 Dec 16 12:47:55.519000 audit[3060]: NETFILTER_CFG table=filter:88 family=10 entries=2 op=nft_register_chain pid=3060 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.519000 audit[3060]: SYSCALL arch=c000003e syscall=46 success=yes exit=828 a0=3 a1=7fffc9099d90 a2=0 a3=7fffc9099d7c items=0 ppid=2898 pid=3060 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.519000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D Dec 16 12:47:55.522000 audit[3061]: NETFILTER_CFG table=filter:89 family=10 entries=1 op=nft_register_chain pid=3061 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.522000 audit[3061]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc68017a70 a2=0 a3=7ffc68017a5c items=0 ppid=2898 pid=3061 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.522000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D464F5257415244002D740066696C746572 Dec 16 12:47:55.526000 audit[3063]: NETFILTER_CFG table=filter:90 family=10 entries=1 op=nft_register_rule pid=3063 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.526000 audit[3063]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffcf001ebd0 a2=0 a3=7ffcf001ebbc items=0 ppid=2898 pid=3063 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.526000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 Dec 16 12:47:55.529000 audit[3064]: NETFILTER_CFG table=filter:91 family=10 entries=1 op=nft_register_chain pid=3064 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.529000 audit[3064]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffd8a7408c0 a2=0 a3=7ffd8a7408ac items=0 ppid=2898 pid=3064 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.529000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 Dec 16 12:47:55.534000 audit[3066]: NETFILTER_CFG table=filter:92 family=10 entries=1 op=nft_register_rule pid=3066 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.534000 audit[3066]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffc1c2f6010 a2=0 a3=7ffc1c2f5ffc items=0 ppid=2898 pid=3066 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.534000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Dec 16 12:47:55.540000 audit[3069]: NETFILTER_CFG table=filter:93 family=10 entries=1 op=nft_register_rule pid=3069 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.540000 audit[3069]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7fffb0bbe140 a2=0 a3=7fffb0bbe12c items=0 ppid=2898 pid=3069 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.540000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D Dec 16 12:47:55.547000 audit[3072]: NETFILTER_CFG table=filter:94 family=10 entries=1 op=nft_register_rule pid=3072 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.547000 audit[3072]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7fff83b96a30 a2=0 a3=7fff83b96a1c items=0 ppid=2898 pid=3072 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.547000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C Dec 16 12:47:55.549000 audit[3073]: NETFILTER_CFG table=nat:95 family=10 entries=1 op=nft_register_chain pid=3073 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.549000 audit[3073]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7fff94bf66d0 a2=0 a3=7fff94bf66bc items=0 ppid=2898 pid=3073 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.549000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D74006E6174 Dec 16 12:47:55.554000 audit[3075]: NETFILTER_CFG table=nat:96 family=10 entries=1 op=nft_register_rule pid=3075 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.554000 audit[3075]: SYSCALL arch=c000003e syscall=46 success=yes exit=524 a0=3 a1=7ffea8d6bda0 a2=0 a3=7ffea8d6bd8c items=0 ppid=2898 pid=3075 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.554000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Dec 16 12:47:55.559000 audit[3078]: NETFILTER_CFG table=nat:97 family=10 entries=1 op=nft_register_rule pid=3078 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.559000 audit[3078]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffd816c7130 a2=0 a3=7ffd816c711c items=0 ppid=2898 pid=3078 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.559000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Dec 16 12:47:55.562000 audit[3079]: NETFILTER_CFG table=nat:98 family=10 entries=1 op=nft_register_chain pid=3079 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.562000 audit[3079]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffeda920170 a2=0 a3=7ffeda92015c items=0 ppid=2898 pid=3079 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.562000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Dec 16 12:47:55.569000 audit[3081]: NETFILTER_CFG table=nat:99 family=10 entries=2 op=nft_register_chain pid=3081 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.569000 audit[3081]: SYSCALL arch=c000003e syscall=46 success=yes exit=612 a0=3 a1=7ffc74711a40 a2=0 a3=7ffc74711a2c items=0 ppid=2898 pid=3081 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.569000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 Dec 16 12:47:55.571000 audit[3082]: NETFILTER_CFG table=filter:100 family=10 entries=1 op=nft_register_chain pid=3082 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.571000 audit[3082]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff5f756570 a2=0 a3=7fff5f75655c items=0 ppid=2898 pid=3082 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.571000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 Dec 16 12:47:55.575000 audit[3084]: NETFILTER_CFG table=filter:101 family=10 entries=1 op=nft_register_rule pid=3084 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.575000 audit[3084]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7ffc10ee8890 a2=0 a3=7ffc10ee887c items=0 ppid=2898 pid=3084 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.575000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Dec 16 12:47:55.582000 audit[3087]: NETFILTER_CFG table=filter:102 family=10 entries=1 op=nft_register_rule pid=3087 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Dec 16 12:47:55.582000 audit[3087]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7ffd5495d6c0 a2=0 a3=7ffd5495d6ac items=0 ppid=2898 pid=3087 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.582000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Dec 16 12:47:55.588000 audit[3089]: NETFILTER_CFG table=filter:103 family=10 entries=3 op=nft_register_rule pid=3089 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" Dec 16 12:47:55.588000 audit[3089]: SYSCALL arch=c000003e syscall=46 success=yes exit=2088 a0=3 a1=7ffc40cb5bb0 a2=0 a3=7ffc40cb5b9c items=0 ppid=2898 pid=3089 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.588000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:47:55.589000 audit[3089]: NETFILTER_CFG table=nat:104 family=10 entries=7 op=nft_register_chain pid=3089 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" Dec 16 12:47:55.589000 audit[3089]: SYSCALL arch=c000003e syscall=46 success=yes exit=2056 a0=3 a1=7ffc40cb5bb0 a2=0 a3=7ffc40cb5b9c items=0 ppid=2898 pid=3089 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:55.589000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:47:57.062882 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount247506945.mount: Deactivated successfully. Dec 16 12:47:57.493093 containerd[1623]: time="2025-12-16T12:47:57.492983502Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:57.494140 containerd[1623]: time="2025-12-16T12:47:57.494118848Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.7: active requests=0, bytes read=23558205" Dec 16 12:47:57.494995 containerd[1623]: time="2025-12-16T12:47:57.494960761Z" level=info msg="ImageCreate event name:\"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:57.496406 containerd[1623]: time="2025-12-16T12:47:57.496358859Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:47:57.497100 containerd[1623]: time="2025-12-16T12:47:57.496787561Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.7\" with image id \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\", repo tag \"quay.io/tigera/operator:v1.38.7\", repo digest \"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\", size \"25057686\" in 2.411830078s" Dec 16 12:47:57.497100 containerd[1623]: time="2025-12-16T12:47:57.496812004Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\" returns image reference \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\"" Dec 16 12:47:57.498987 containerd[1623]: time="2025-12-16T12:47:57.498968048Z" level=info msg="CreateContainer within sandbox \"02ac242cef836574fb5d0643d85fa43d58200df961b34aef937073f2bcd94794\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Dec 16 12:47:57.506032 containerd[1623]: time="2025-12-16T12:47:57.504368820Z" level=info msg="Container 3cf9cfb2aba45a3e247a26c87c419524f478a8503621e1fc8baea6bcda29d055: CDI devices from CRI Config.CDIDevices: []" Dec 16 12:47:57.507856 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3686740248.mount: Deactivated successfully. Dec 16 12:47:57.514841 containerd[1623]: time="2025-12-16T12:47:57.514810640Z" level=info msg="CreateContainer within sandbox \"02ac242cef836574fb5d0643d85fa43d58200df961b34aef937073f2bcd94794\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"3cf9cfb2aba45a3e247a26c87c419524f478a8503621e1fc8baea6bcda29d055\"" Dec 16 12:47:57.516157 containerd[1623]: time="2025-12-16T12:47:57.515292349Z" level=info msg="StartContainer for \"3cf9cfb2aba45a3e247a26c87c419524f478a8503621e1fc8baea6bcda29d055\"" Dec 16 12:47:57.516157 containerd[1623]: time="2025-12-16T12:47:57.516039958Z" level=info msg="connecting to shim 3cf9cfb2aba45a3e247a26c87c419524f478a8503621e1fc8baea6bcda29d055" address="unix:///run/containerd/s/d13c141a964ed7933bf8b9828d04ff61c824a1d935f623eb0bedd7e1afc4c513" protocol=ttrpc version=3 Dec 16 12:47:57.539244 systemd[1]: Started cri-containerd-3cf9cfb2aba45a3e247a26c87c419524f478a8503621e1fc8baea6bcda29d055.scope - libcontainer container 3cf9cfb2aba45a3e247a26c87c419524f478a8503621e1fc8baea6bcda29d055. Dec 16 12:47:57.549000 audit: BPF prog-id=146 op=LOAD Dec 16 12:47:57.549000 audit: BPF prog-id=147 op=LOAD Dec 16 12:47:57.549000 audit[3098]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8238 a2=98 a3=0 items=0 ppid=2924 pid=3098 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:57.549000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3363663963666232616261343561336532343761323663383763343139 Dec 16 12:47:57.549000 audit: BPF prog-id=147 op=UNLOAD Dec 16 12:47:57.549000 audit[3098]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2924 pid=3098 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:57.549000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3363663963666232616261343561336532343761323663383763343139 Dec 16 12:47:57.549000 audit: BPF prog-id=148 op=LOAD Dec 16 12:47:57.549000 audit[3098]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8488 a2=98 a3=0 items=0 ppid=2924 pid=3098 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:57.549000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3363663963666232616261343561336532343761323663383763343139 Dec 16 12:47:57.549000 audit: BPF prog-id=149 op=LOAD Dec 16 12:47:57.549000 audit[3098]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a8218 a2=98 a3=0 items=0 ppid=2924 pid=3098 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:57.549000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3363663963666232616261343561336532343761323663383763343139 Dec 16 12:47:57.549000 audit: BPF prog-id=149 op=UNLOAD Dec 16 12:47:57.549000 audit[3098]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2924 pid=3098 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:57.549000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3363663963666232616261343561336532343761323663383763343139 Dec 16 12:47:57.549000 audit: BPF prog-id=148 op=UNLOAD Dec 16 12:47:57.549000 audit[3098]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2924 pid=3098 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:57.549000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3363663963666232616261343561336532343761323663383763343139 Dec 16 12:47:57.549000 audit: BPF prog-id=150 op=LOAD Dec 16 12:47:57.549000 audit[3098]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a86e8 a2=98 a3=0 items=0 ppid=2924 pid=3098 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:47:57.549000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3363663963666232616261343561336532343761323663383763343139 Dec 16 12:47:57.563598 containerd[1623]: time="2025-12-16T12:47:57.563526439Z" level=info msg="StartContainer for \"3cf9cfb2aba45a3e247a26c87c419524f478a8503621e1fc8baea6bcda29d055\" returns successfully" Dec 16 12:47:58.209862 update_engine[1602]: I20251216 12:47:58.209753 1602 update_attempter.cc:509] Updating boot flags... Dec 16 12:47:58.253203 kubelet[2796]: I1216 12:47:58.252646 2796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-sfv8g" podStartSLOduration=4.252622446 podStartE2EDuration="4.252622446s" podCreationTimestamp="2025-12-16 12:47:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-16 12:47:55.278247649 +0000 UTC m=+8.211975677" watchObservedRunningTime="2025-12-16 12:47:58.252622446 +0000 UTC m=+11.186350484" Dec 16 12:47:58.253203 kubelet[2796]: I1216 12:47:58.252988 2796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7dcd859c48-58brr" podStartSLOduration=1.8382788190000001 podStartE2EDuration="4.252973829s" podCreationTimestamp="2025-12-16 12:47:54 +0000 UTC" firstStartedPulling="2025-12-16 12:47:55.083269795 +0000 UTC m=+8.016997802" lastFinishedPulling="2025-12-16 12:47:57.497964806 +0000 UTC m=+10.431692812" observedRunningTime="2025-12-16 12:47:58.252195154 +0000 UTC m=+11.185923192" watchObservedRunningTime="2025-12-16 12:47:58.252973829 +0000 UTC m=+11.186701866" Dec 16 12:48:01.394209 sudo[1855]: pam_unix(sudo:session): session closed for user root Dec 16 12:48:01.401032 kernel: kauditd_printk_skb: 224 callbacks suppressed Dec 16 12:48:01.401110 kernel: audit: type=1106 audit(1765889281.393:523): pid=1855 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Dec 16 12:48:01.393000 audit[1855]: USER_END pid=1855 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Dec 16 12:48:01.393000 audit[1855]: CRED_DISP pid=1855 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Dec 16 12:48:01.408049 kernel: audit: type=1104 audit(1765889281.393:524): pid=1855 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Dec 16 12:48:01.574108 sshd[1854]: Connection closed by 147.75.109.163 port 33338 Dec 16 12:48:01.574311 sshd-session[1850]: pam_unix(sshd:session): session closed for user core Dec 16 12:48:01.576000 audit[1850]: USER_END pid=1850 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:48:01.587497 kernel: audit: type=1106 audit(1765889281.576:525): pid=1850 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:48:01.585868 systemd[1]: sshd@6-77.42.19.41:22-147.75.109.163:33338.service: Deactivated successfully. Dec 16 12:48:01.587265 systemd-logind[1601]: Session 8 logged out. Waiting for processes to exit. Dec 16 12:48:01.588307 systemd[1]: session-8.scope: Deactivated successfully. Dec 16 12:48:01.588532 systemd[1]: session-8.scope: Consumed 4.393s CPU time, 158.1M memory peak. Dec 16 12:48:01.576000 audit[1850]: CRED_DISP pid=1850 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:48:01.596661 systemd-logind[1601]: Removed session 8. Dec 16 12:48:01.597072 kernel: audit: type=1104 audit(1765889281.576:526): pid=1850 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:48:01.585000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-77.42.19.41:22-147.75.109.163:33338 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:48:01.604138 kernel: audit: type=1131 audit(1765889281.585:527): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-77.42.19.41:22-147.75.109.163:33338 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:48:02.019000 audit[3204]: NETFILTER_CFG table=filter:105 family=2 entries=15 op=nft_register_rule pid=3204 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:02.031605 kernel: audit: type=1325 audit(1765889282.019:528): table=filter:105 family=2 entries=15 op=nft_register_rule pid=3204 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:02.031681 kernel: audit: type=1300 audit(1765889282.019:528): arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7fff028955f0 a2=0 a3=7fff028955dc items=0 ppid=2898 pid=3204 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:02.019000 audit[3204]: SYSCALL arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7fff028955f0 a2=0 a3=7fff028955dc items=0 ppid=2898 pid=3204 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:02.019000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:02.041633 kernel: audit: type=1327 audit(1765889282.019:528): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:02.041693 kernel: audit: type=1325 audit(1765889282.026:529): table=nat:106 family=2 entries=12 op=nft_register_rule pid=3204 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:02.026000 audit[3204]: NETFILTER_CFG table=nat:106 family=2 entries=12 op=nft_register_rule pid=3204 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:02.048508 kernel: audit: type=1300 audit(1765889282.026:529): arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7fff028955f0 a2=0 a3=0 items=0 ppid=2898 pid=3204 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:02.026000 audit[3204]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7fff028955f0 a2=0 a3=0 items=0 ppid=2898 pid=3204 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:02.026000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:02.063000 audit[3206]: NETFILTER_CFG table=filter:107 family=2 entries=16 op=nft_register_rule pid=3206 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:02.063000 audit[3206]: SYSCALL arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7ffc27dcbed0 a2=0 a3=7ffc27dcbebc items=0 ppid=2898 pid=3206 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:02.063000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:02.069000 audit[3206]: NETFILTER_CFG table=nat:108 family=2 entries=12 op=nft_register_rule pid=3206 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:02.069000 audit[3206]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffc27dcbed0 a2=0 a3=0 items=0 ppid=2898 pid=3206 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:02.069000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:03.864000 audit[3208]: NETFILTER_CFG table=filter:109 family=2 entries=17 op=nft_register_rule pid=3208 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:03.864000 audit[3208]: SYSCALL arch=c000003e syscall=46 success=yes exit=6736 a0=3 a1=7ffc43abd1d0 a2=0 a3=7ffc43abd1bc items=0 ppid=2898 pid=3208 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:03.864000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:03.870000 audit[3208]: NETFILTER_CFG table=nat:110 family=2 entries=12 op=nft_register_rule pid=3208 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:03.870000 audit[3208]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffc43abd1d0 a2=0 a3=0 items=0 ppid=2898 pid=3208 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:03.870000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:04.037000 audit[3210]: NETFILTER_CFG table=filter:111 family=2 entries=19 op=nft_register_rule pid=3210 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:04.037000 audit[3210]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffc78eb2340 a2=0 a3=7ffc78eb232c items=0 ppid=2898 pid=3210 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:04.037000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:04.041000 audit[3210]: NETFILTER_CFG table=nat:112 family=2 entries=12 op=nft_register_rule pid=3210 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:04.041000 audit[3210]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffc78eb2340 a2=0 a3=0 items=0 ppid=2898 pid=3210 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:04.041000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:05.056000 audit[3212]: NETFILTER_CFG table=filter:113 family=2 entries=20 op=nft_register_rule pid=3212 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:05.056000 audit[3212]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffc67a46870 a2=0 a3=7ffc67a4685c items=0 ppid=2898 pid=3212 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:05.056000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:05.060000 audit[3212]: NETFILTER_CFG table=nat:114 family=2 entries=12 op=nft_register_rule pid=3212 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:05.060000 audit[3212]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffc67a46870 a2=0 a3=0 items=0 ppid=2898 pid=3212 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:05.060000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:05.770275 systemd[1]: Created slice kubepods-besteffort-podbb069726_46da_40dc_936f_a923527dcad0.slice - libcontainer container kubepods-besteffort-podbb069726_46da_40dc_936f_a923527dcad0.slice. Dec 16 12:48:05.776687 kubelet[2796]: I1216 12:48:05.776665 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb069726-46da-40dc-936f-a923527dcad0-tigera-ca-bundle\") pod \"calico-typha-58cf6cb8d8-qvj4d\" (UID: \"bb069726-46da-40dc-936f-a923527dcad0\") " pod="calico-system/calico-typha-58cf6cb8d8-qvj4d" Dec 16 12:48:05.777537 kubelet[2796]: I1216 12:48:05.777491 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/bb069726-46da-40dc-936f-a923527dcad0-typha-certs\") pod \"calico-typha-58cf6cb8d8-qvj4d\" (UID: \"bb069726-46da-40dc-936f-a923527dcad0\") " pod="calico-system/calico-typha-58cf6cb8d8-qvj4d" Dec 16 12:48:05.777678 kubelet[2796]: I1216 12:48:05.777612 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv8qh\" (UniqueName: \"kubernetes.io/projected/bb069726-46da-40dc-936f-a923527dcad0-kube-api-access-pv8qh\") pod \"calico-typha-58cf6cb8d8-qvj4d\" (UID: \"bb069726-46da-40dc-936f-a923527dcad0\") " pod="calico-system/calico-typha-58cf6cb8d8-qvj4d" Dec 16 12:48:05.988098 systemd[1]: Created slice kubepods-besteffort-pod10e2eba4_b4e6_41d6_b45f_fa8e3b9fe83b.slice - libcontainer container kubepods-besteffort-pod10e2eba4_b4e6_41d6_b45f_fa8e3b9fe83b.slice. Dec 16 12:48:06.076000 audit[3218]: NETFILTER_CFG table=filter:115 family=2 entries=21 op=nft_register_rule pid=3218 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:06.076000 audit[3218]: SYSCALL arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7ffd7caa7530 a2=0 a3=7ffd7caa751c items=0 ppid=2898 pid=3218 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.076000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:06.079010 kubelet[2796]: I1216 12:48:06.078616 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b-var-lib-calico\") pod \"calico-node-kddxn\" (UID: \"10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b\") " pod="calico-system/calico-node-kddxn" Dec 16 12:48:06.079010 kubelet[2796]: I1216 12:48:06.078657 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b-cni-net-dir\") pod \"calico-node-kddxn\" (UID: \"10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b\") " pod="calico-system/calico-node-kddxn" Dec 16 12:48:06.079010 kubelet[2796]: I1216 12:48:06.078674 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b-cni-bin-dir\") pod \"calico-node-kddxn\" (UID: \"10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b\") " pod="calico-system/calico-node-kddxn" Dec 16 12:48:06.079010 kubelet[2796]: I1216 12:48:06.078690 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b-flexvol-driver-host\") pod \"calico-node-kddxn\" (UID: \"10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b\") " pod="calico-system/calico-node-kddxn" Dec 16 12:48:06.079010 kubelet[2796]: I1216 12:48:06.078704 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b-xtables-lock\") pod \"calico-node-kddxn\" (UID: \"10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b\") " pod="calico-system/calico-node-kddxn" Dec 16 12:48:06.079215 kubelet[2796]: I1216 12:48:06.078719 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b-lib-modules\") pod \"calico-node-kddxn\" (UID: \"10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b\") " pod="calico-system/calico-node-kddxn" Dec 16 12:48:06.079215 kubelet[2796]: I1216 12:48:06.078733 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b-tigera-ca-bundle\") pod \"calico-node-kddxn\" (UID: \"10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b\") " pod="calico-system/calico-node-kddxn" Dec 16 12:48:06.079215 kubelet[2796]: I1216 12:48:06.078747 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x9dx\" (UniqueName: \"kubernetes.io/projected/10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b-kube-api-access-2x9dx\") pod \"calico-node-kddxn\" (UID: \"10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b\") " pod="calico-system/calico-node-kddxn" Dec 16 12:48:06.079215 kubelet[2796]: I1216 12:48:06.078763 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b-node-certs\") pod \"calico-node-kddxn\" (UID: \"10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b\") " pod="calico-system/calico-node-kddxn" Dec 16 12:48:06.079215 kubelet[2796]: I1216 12:48:06.078786 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b-cni-log-dir\") pod \"calico-node-kddxn\" (UID: \"10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b\") " pod="calico-system/calico-node-kddxn" Dec 16 12:48:06.079339 kubelet[2796]: I1216 12:48:06.078801 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b-policysync\") pod \"calico-node-kddxn\" (UID: \"10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b\") " pod="calico-system/calico-node-kddxn" Dec 16 12:48:06.079339 kubelet[2796]: I1216 12:48:06.078844 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b-var-run-calico\") pod \"calico-node-kddxn\" (UID: \"10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b\") " pod="calico-system/calico-node-kddxn" Dec 16 12:48:06.080421 containerd[1623]: time="2025-12-16T12:48:06.080363713Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-58cf6cb8d8-qvj4d,Uid:bb069726-46da-40dc-936f-a923527dcad0,Namespace:calico-system,Attempt:0,}" Dec 16 12:48:06.081000 audit[3218]: NETFILTER_CFG table=nat:116 family=2 entries=12 op=nft_register_rule pid=3218 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:06.081000 audit[3218]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd7caa7530 a2=0 a3=0 items=0 ppid=2898 pid=3218 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.081000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:06.123083 containerd[1623]: time="2025-12-16T12:48:06.122821515Z" level=info msg="connecting to shim f493bf30acc3ff193cf2991a2f4b5dfe4c58919a06232e9e68e7cf1140fd7484" address="unix:///run/containerd/s/118132e030fad33a655a9cac733eb4e13d7db8654e75ce8185154114c5f5146a" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:48:06.171379 systemd[1]: Started cri-containerd-f493bf30acc3ff193cf2991a2f4b5dfe4c58919a06232e9e68e7cf1140fd7484.scope - libcontainer container f493bf30acc3ff193cf2991a2f4b5dfe4c58919a06232e9e68e7cf1140fd7484. Dec 16 12:48:06.174609 kubelet[2796]: E1216 12:48:06.174357 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:48:06.190037 kubelet[2796]: E1216 12:48:06.189994 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.190088 kubelet[2796]: W1216 12:48:06.190044 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.190674 kubelet[2796]: E1216 12:48:06.190647 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.197152 kubelet[2796]: E1216 12:48:06.197130 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.197152 kubelet[2796]: W1216 12:48:06.197146 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.197285 kubelet[2796]: E1216 12:48:06.197160 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.224000 audit: BPF prog-id=151 op=LOAD Dec 16 12:48:06.225000 audit: BPF prog-id=152 op=LOAD Dec 16 12:48:06.225000 audit[3239]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0238 a2=98 a3=0 items=0 ppid=3228 pid=3239 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.225000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6634393362663330616363336666313933636632393931613266346235 Dec 16 12:48:06.225000 audit: BPF prog-id=152 op=UNLOAD Dec 16 12:48:06.225000 audit[3239]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3228 pid=3239 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.225000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6634393362663330616363336666313933636632393931613266346235 Dec 16 12:48:06.225000 audit: BPF prog-id=153 op=LOAD Dec 16 12:48:06.225000 audit[3239]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a0488 a2=98 a3=0 items=0 ppid=3228 pid=3239 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.225000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6634393362663330616363336666313933636632393931613266346235 Dec 16 12:48:06.225000 audit: BPF prog-id=154 op=LOAD Dec 16 12:48:06.225000 audit[3239]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a0218 a2=98 a3=0 items=0 ppid=3228 pid=3239 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.225000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6634393362663330616363336666313933636632393931613266346235 Dec 16 12:48:06.225000 audit: BPF prog-id=154 op=UNLOAD Dec 16 12:48:06.225000 audit[3239]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3228 pid=3239 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.225000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6634393362663330616363336666313933636632393931613266346235 Dec 16 12:48:06.225000 audit: BPF prog-id=153 op=UNLOAD Dec 16 12:48:06.225000 audit[3239]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3228 pid=3239 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.225000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6634393362663330616363336666313933636632393931613266346235 Dec 16 12:48:06.225000 audit: BPF prog-id=155 op=LOAD Dec 16 12:48:06.225000 audit[3239]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a06e8 a2=98 a3=0 items=0 ppid=3228 pid=3239 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.225000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6634393362663330616363336666313933636632393931613266346235 Dec 16 12:48:06.260177 containerd[1623]: time="2025-12-16T12:48:06.260139386Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-58cf6cb8d8-qvj4d,Uid:bb069726-46da-40dc-936f-a923527dcad0,Namespace:calico-system,Attempt:0,} returns sandbox id \"f493bf30acc3ff193cf2991a2f4b5dfe4c58919a06232e9e68e7cf1140fd7484\"" Dec 16 12:48:06.262220 containerd[1623]: time="2025-12-16T12:48:06.262200403Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\"" Dec 16 12:48:06.270264 kubelet[2796]: E1216 12:48:06.270213 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.270264 kubelet[2796]: W1216 12:48:06.270261 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.270367 kubelet[2796]: E1216 12:48:06.270278 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.271001 kubelet[2796]: E1216 12:48:06.270714 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.271001 kubelet[2796]: W1216 12:48:06.270829 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.271001 kubelet[2796]: E1216 12:48:06.270859 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.271182 kubelet[2796]: E1216 12:48:06.271170 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.271332 kubelet[2796]: W1216 12:48:06.271237 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.271429 kubelet[2796]: E1216 12:48:06.271419 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.271660 kubelet[2796]: E1216 12:48:06.271651 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.271733 kubelet[2796]: W1216 12:48:06.271715 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.272308 kubelet[2796]: E1216 12:48:06.272295 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.272558 kubelet[2796]: E1216 12:48:06.272513 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.272558 kubelet[2796]: W1216 12:48:06.272523 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.272558 kubelet[2796]: E1216 12:48:06.272531 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.273035 kubelet[2796]: E1216 12:48:06.272799 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.273035 kubelet[2796]: W1216 12:48:06.272809 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.273035 kubelet[2796]: E1216 12:48:06.272816 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.273413 kubelet[2796]: E1216 12:48:06.273284 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.273413 kubelet[2796]: W1216 12:48:06.273294 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.273413 kubelet[2796]: E1216 12:48:06.273302 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.274695 kubelet[2796]: E1216 12:48:06.274608 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.274695 kubelet[2796]: W1216 12:48:06.274618 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.274695 kubelet[2796]: E1216 12:48:06.274626 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.274930 kubelet[2796]: E1216 12:48:06.274855 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.274930 kubelet[2796]: W1216 12:48:06.274882 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.274930 kubelet[2796]: E1216 12:48:06.274891 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.275246 kubelet[2796]: E1216 12:48:06.275202 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.275246 kubelet[2796]: W1216 12:48:06.275211 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.275246 kubelet[2796]: E1216 12:48:06.275219 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.275516 kubelet[2796]: E1216 12:48:06.275471 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.275516 kubelet[2796]: W1216 12:48:06.275480 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.275516 kubelet[2796]: E1216 12:48:06.275488 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.275848 kubelet[2796]: E1216 12:48:06.275797 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.275848 kubelet[2796]: W1216 12:48:06.275806 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.275848 kubelet[2796]: E1216 12:48:06.275814 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.276412 kubelet[2796]: E1216 12:48:06.276272 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.276412 kubelet[2796]: W1216 12:48:06.276334 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.276412 kubelet[2796]: E1216 12:48:06.276344 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.278342 kubelet[2796]: E1216 12:48:06.278295 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.278342 kubelet[2796]: W1216 12:48:06.278305 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.278342 kubelet[2796]: E1216 12:48:06.278314 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.278628 kubelet[2796]: E1216 12:48:06.278580 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.278628 kubelet[2796]: W1216 12:48:06.278590 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.278628 kubelet[2796]: E1216 12:48:06.278598 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.278841 kubelet[2796]: E1216 12:48:06.278831 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.278935 kubelet[2796]: W1216 12:48:06.278887 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.278935 kubelet[2796]: E1216 12:48:06.278898 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.279173 kubelet[2796]: E1216 12:48:06.279128 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.279173 kubelet[2796]: W1216 12:48:06.279139 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.279173 kubelet[2796]: E1216 12:48:06.279147 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.279435 kubelet[2796]: E1216 12:48:06.279424 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.279569 kubelet[2796]: W1216 12:48:06.279496 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.279569 kubelet[2796]: E1216 12:48:06.279508 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.279965 kubelet[2796]: E1216 12:48:06.279949 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.280340 kubelet[2796]: W1216 12:48:06.280117 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.280340 kubelet[2796]: E1216 12:48:06.280128 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.280752 kubelet[2796]: E1216 12:48:06.280561 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.280752 kubelet[2796]: W1216 12:48:06.280569 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.280752 kubelet[2796]: E1216 12:48:06.280577 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.281194 kubelet[2796]: E1216 12:48:06.281172 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.281194 kubelet[2796]: W1216 12:48:06.281191 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.281517 kubelet[2796]: E1216 12:48:06.281206 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.281517 kubelet[2796]: I1216 12:48:06.281233 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/2053803c-11b9-48f6-949c-d0d885185bfe-varrun\") pod \"csi-node-driver-m8p7d\" (UID: \"2053803c-11b9-48f6-949c-d0d885185bfe\") " pod="calico-system/csi-node-driver-m8p7d" Dec 16 12:48:06.281517 kubelet[2796]: E1216 12:48:06.281349 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.281517 kubelet[2796]: W1216 12:48:06.281357 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.281517 kubelet[2796]: E1216 12:48:06.281364 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.281517 kubelet[2796]: I1216 12:48:06.281376 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2053803c-11b9-48f6-949c-d0d885185bfe-registration-dir\") pod \"csi-node-driver-m8p7d\" (UID: \"2053803c-11b9-48f6-949c-d0d885185bfe\") " pod="calico-system/csi-node-driver-m8p7d" Dec 16 12:48:06.281517 kubelet[2796]: E1216 12:48:06.281519 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.281647 kubelet[2796]: W1216 12:48:06.281528 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.281647 kubelet[2796]: E1216 12:48:06.281537 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.281647 kubelet[2796]: I1216 12:48:06.281549 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2053803c-11b9-48f6-949c-d0d885185bfe-socket-dir\") pod \"csi-node-driver-m8p7d\" (UID: \"2053803c-11b9-48f6-949c-d0d885185bfe\") " pod="calico-system/csi-node-driver-m8p7d" Dec 16 12:48:06.282058 kubelet[2796]: E1216 12:48:06.281807 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.282058 kubelet[2796]: W1216 12:48:06.281823 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.282058 kubelet[2796]: E1216 12:48:06.281831 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.282058 kubelet[2796]: I1216 12:48:06.281843 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k7vq\" (UniqueName: \"kubernetes.io/projected/2053803c-11b9-48f6-949c-d0d885185bfe-kube-api-access-6k7vq\") pod \"csi-node-driver-m8p7d\" (UID: \"2053803c-11b9-48f6-949c-d0d885185bfe\") " pod="calico-system/csi-node-driver-m8p7d" Dec 16 12:48:06.283332 kubelet[2796]: E1216 12:48:06.283237 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.283332 kubelet[2796]: W1216 12:48:06.283249 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.283332 kubelet[2796]: E1216 12:48:06.283274 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.283332 kubelet[2796]: I1216 12:48:06.283296 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2053803c-11b9-48f6-949c-d0d885185bfe-kubelet-dir\") pod \"csi-node-driver-m8p7d\" (UID: \"2053803c-11b9-48f6-949c-d0d885185bfe\") " pod="calico-system/csi-node-driver-m8p7d" Dec 16 12:48:06.283582 kubelet[2796]: E1216 12:48:06.283562 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.283582 kubelet[2796]: W1216 12:48:06.283571 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.283682 kubelet[2796]: E1216 12:48:06.283652 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.283853 kubelet[2796]: E1216 12:48:06.283824 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.283853 kubelet[2796]: W1216 12:48:06.283836 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.286039 kubelet[2796]: E1216 12:48:06.284033 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.286194 kubelet[2796]: E1216 12:48:06.286171 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.286194 kubelet[2796]: W1216 12:48:06.286182 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.286370 kubelet[2796]: E1216 12:48:06.286354 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.286467 kubelet[2796]: E1216 12:48:06.286459 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.286552 kubelet[2796]: W1216 12:48:06.286508 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.286623 kubelet[2796]: E1216 12:48:06.286595 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.286751 kubelet[2796]: E1216 12:48:06.286733 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.286751 kubelet[2796]: W1216 12:48:06.286742 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.286850 kubelet[2796]: E1216 12:48:06.286835 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.286981 kubelet[2796]: E1216 12:48:06.286963 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.286981 kubelet[2796]: W1216 12:48:06.286972 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.287225 kubelet[2796]: E1216 12:48:06.287208 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.287332 kubelet[2796]: E1216 12:48:06.287324 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.287390 kubelet[2796]: W1216 12:48:06.287381 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.287495 kubelet[2796]: E1216 12:48:06.287446 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.287696 kubelet[2796]: E1216 12:48:06.287687 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.287743 kubelet[2796]: W1216 12:48:06.287735 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.287863 kubelet[2796]: E1216 12:48:06.287813 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.288058 kubelet[2796]: E1216 12:48:06.288010 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.288058 kubelet[2796]: W1216 12:48:06.288038 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.288058 kubelet[2796]: E1216 12:48:06.288046 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.288297 kubelet[2796]: E1216 12:48:06.288269 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.288297 kubelet[2796]: W1216 12:48:06.288277 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.288297 kubelet[2796]: E1216 12:48:06.288285 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.294197 containerd[1623]: time="2025-12-16T12:48:06.294155106Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-kddxn,Uid:10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b,Namespace:calico-system,Attempt:0,}" Dec 16 12:48:06.317060 containerd[1623]: time="2025-12-16T12:48:06.316940954Z" level=info msg="connecting to shim 31a2c3fb2359d25620ab9b707e5cae784a39bde50f5c6db376d2523e06a8681d" address="unix:///run/containerd/s/698b86754e3d1d4bc1220c168741fa4038a4fc0135498ef4de6454cf03124774" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:48:06.341258 systemd[1]: Started cri-containerd-31a2c3fb2359d25620ab9b707e5cae784a39bde50f5c6db376d2523e06a8681d.scope - libcontainer container 31a2c3fb2359d25620ab9b707e5cae784a39bde50f5c6db376d2523e06a8681d. Dec 16 12:48:06.352000 audit: BPF prog-id=156 op=LOAD Dec 16 12:48:06.353000 audit: BPF prog-id=157 op=LOAD Dec 16 12:48:06.353000 audit[3332]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=3320 pid=3332 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.353000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3331613263336662323335396432353632306162396237303765356361 Dec 16 12:48:06.353000 audit: BPF prog-id=157 op=UNLOAD Dec 16 12:48:06.353000 audit[3332]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3320 pid=3332 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.353000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3331613263336662323335396432353632306162396237303765356361 Dec 16 12:48:06.353000 audit: BPF prog-id=158 op=LOAD Dec 16 12:48:06.353000 audit[3332]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3320 pid=3332 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.353000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3331613263336662323335396432353632306162396237303765356361 Dec 16 12:48:06.353000 audit: BPF prog-id=159 op=LOAD Dec 16 12:48:06.353000 audit[3332]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3320 pid=3332 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.353000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3331613263336662323335396432353632306162396237303765356361 Dec 16 12:48:06.353000 audit: BPF prog-id=159 op=UNLOAD Dec 16 12:48:06.353000 audit[3332]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3320 pid=3332 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.353000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3331613263336662323335396432353632306162396237303765356361 Dec 16 12:48:06.353000 audit: BPF prog-id=158 op=UNLOAD Dec 16 12:48:06.353000 audit[3332]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3320 pid=3332 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.353000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3331613263336662323335396432353632306162396237303765356361 Dec 16 12:48:06.353000 audit: BPF prog-id=160 op=LOAD Dec 16 12:48:06.353000 audit[3332]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=3320 pid=3332 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:06.353000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3331613263336662323335396432353632306162396237303765356361 Dec 16 12:48:06.377806 containerd[1623]: time="2025-12-16T12:48:06.377742593Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-kddxn,Uid:10e2eba4-b4e6-41d6-b45f-fa8e3b9fe83b,Namespace:calico-system,Attempt:0,} returns sandbox id \"31a2c3fb2359d25620ab9b707e5cae784a39bde50f5c6db376d2523e06a8681d\"" Dec 16 12:48:06.383965 kubelet[2796]: E1216 12:48:06.383825 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.383965 kubelet[2796]: W1216 12:48:06.383841 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.383965 kubelet[2796]: E1216 12:48:06.383856 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.384153 kubelet[2796]: E1216 12:48:06.384143 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.384290 kubelet[2796]: W1216 12:48:06.384194 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.384290 kubelet[2796]: E1216 12:48:06.384213 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.384380 kubelet[2796]: E1216 12:48:06.384371 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.384432 kubelet[2796]: W1216 12:48:06.384423 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.384498 kubelet[2796]: E1216 12:48:06.384489 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.384753 kubelet[2796]: E1216 12:48:06.384723 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.384753 kubelet[2796]: W1216 12:48:06.384743 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.384830 kubelet[2796]: E1216 12:48:06.384787 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.384997 kubelet[2796]: E1216 12:48:06.384975 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.384997 kubelet[2796]: W1216 12:48:06.384988 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.385072 kubelet[2796]: E1216 12:48:06.385005 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.385276 kubelet[2796]: E1216 12:48:06.385254 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.385276 kubelet[2796]: W1216 12:48:06.385268 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.385326 kubelet[2796]: E1216 12:48:06.385284 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.385491 kubelet[2796]: E1216 12:48:06.385460 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.385491 kubelet[2796]: W1216 12:48:06.385482 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.385617 kubelet[2796]: E1216 12:48:06.385573 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.385681 kubelet[2796]: E1216 12:48:06.385660 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.385681 kubelet[2796]: W1216 12:48:06.385672 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.385681 kubelet[2796]: E1216 12:48:06.385681 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.385897 kubelet[2796]: E1216 12:48:06.385875 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.385897 kubelet[2796]: W1216 12:48:06.385887 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.385952 kubelet[2796]: E1216 12:48:06.385948 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.386122 kubelet[2796]: E1216 12:48:06.386108 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.386122 kubelet[2796]: W1216 12:48:06.386119 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.386174 kubelet[2796]: E1216 12:48:06.386159 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.386386 kubelet[2796]: E1216 12:48:06.386346 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.386386 kubelet[2796]: W1216 12:48:06.386358 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.386460 kubelet[2796]: E1216 12:48:06.386437 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.386569 kubelet[2796]: E1216 12:48:06.386549 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.386591 kubelet[2796]: W1216 12:48:06.386576 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.386612 kubelet[2796]: E1216 12:48:06.386592 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.386810 kubelet[2796]: E1216 12:48:06.386789 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.386810 kubelet[2796]: W1216 12:48:06.386801 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.386810 kubelet[2796]: E1216 12:48:06.386809 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.387006 kubelet[2796]: E1216 12:48:06.386984 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.387066 kubelet[2796]: W1216 12:48:06.387006 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.387066 kubelet[2796]: E1216 12:48:06.387023 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.387260 kubelet[2796]: E1216 12:48:06.387237 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.387260 kubelet[2796]: W1216 12:48:06.387250 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.387349 kubelet[2796]: E1216 12:48:06.387328 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.387503 kubelet[2796]: E1216 12:48:06.387481 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.387503 kubelet[2796]: W1216 12:48:06.387492 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.387597 kubelet[2796]: E1216 12:48:06.387576 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.387756 kubelet[2796]: E1216 12:48:06.387735 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.387756 kubelet[2796]: W1216 12:48:06.387747 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.387809 kubelet[2796]: E1216 12:48:06.387759 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.387964 kubelet[2796]: E1216 12:48:06.387943 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.387964 kubelet[2796]: W1216 12:48:06.387955 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.387964 kubelet[2796]: E1216 12:48:06.387964 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.388180 kubelet[2796]: E1216 12:48:06.388158 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.388180 kubelet[2796]: W1216 12:48:06.388170 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.388248 kubelet[2796]: E1216 12:48:06.388206 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.388375 kubelet[2796]: E1216 12:48:06.388345 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.388375 kubelet[2796]: W1216 12:48:06.388356 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.388424 kubelet[2796]: E1216 12:48:06.388416 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.388633 kubelet[2796]: E1216 12:48:06.388611 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.388633 kubelet[2796]: W1216 12:48:06.388623 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.388712 kubelet[2796]: E1216 12:48:06.388694 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.388831 kubelet[2796]: E1216 12:48:06.388810 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.388831 kubelet[2796]: W1216 12:48:06.388830 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.388961 kubelet[2796]: E1216 12:48:06.388919 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.389034 kubelet[2796]: E1216 12:48:06.388993 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.389034 kubelet[2796]: W1216 12:48:06.389006 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.389085 kubelet[2796]: E1216 12:48:06.389058 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.389318 kubelet[2796]: E1216 12:48:06.389294 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.389318 kubelet[2796]: W1216 12:48:06.389307 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.389318 kubelet[2796]: E1216 12:48:06.389313 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.390216 kubelet[2796]: E1216 12:48:06.390182 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.390216 kubelet[2796]: W1216 12:48:06.390197 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.390216 kubelet[2796]: E1216 12:48:06.390205 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:06.395849 kubelet[2796]: E1216 12:48:06.395822 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:06.395849 kubelet[2796]: W1216 12:48:06.395837 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:06.395849 kubelet[2796]: E1216 12:48:06.395848 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:08.046708 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4049672633.mount: Deactivated successfully. Dec 16 12:48:08.154291 kubelet[2796]: E1216 12:48:08.154199 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:48:08.476940 containerd[1623]: time="2025-12-16T12:48:08.476823496Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:48:08.478173 containerd[1623]: time="2025-12-16T12:48:08.478042202Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.4: active requests=0, bytes read=33735893" Dec 16 12:48:08.479347 containerd[1623]: time="2025-12-16T12:48:08.479323058Z" level=info msg="ImageCreate event name:\"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:48:08.481462 containerd[1623]: time="2025-12-16T12:48:08.481429906Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:48:08.482007 containerd[1623]: time="2025-12-16T12:48:08.481984549Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.4\" with image id \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\", size \"35234482\" in 2.219756938s" Dec 16 12:48:08.482096 containerd[1623]: time="2025-12-16T12:48:08.482082179Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\" returns image reference \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\"" Dec 16 12:48:08.483117 containerd[1623]: time="2025-12-16T12:48:08.483098958Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\"" Dec 16 12:48:08.494834 containerd[1623]: time="2025-12-16T12:48:08.494789196Z" level=info msg="CreateContainer within sandbox \"f493bf30acc3ff193cf2991a2f4b5dfe4c58919a06232e9e68e7cf1140fd7484\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Dec 16 12:48:08.504460 containerd[1623]: time="2025-12-16T12:48:08.504433737Z" level=info msg="Container 238897f00b3e4c45fd7e38d99fdb8c6f70cfa0b30c8497a1dd91cf094373e2cf: CDI devices from CRI Config.CDIDevices: []" Dec 16 12:48:08.511825 containerd[1623]: time="2025-12-16T12:48:08.511787115Z" level=info msg="CreateContainer within sandbox \"f493bf30acc3ff193cf2991a2f4b5dfe4c58919a06232e9e68e7cf1140fd7484\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"238897f00b3e4c45fd7e38d99fdb8c6f70cfa0b30c8497a1dd91cf094373e2cf\"" Dec 16 12:48:08.513329 containerd[1623]: time="2025-12-16T12:48:08.513220117Z" level=info msg="StartContainer for \"238897f00b3e4c45fd7e38d99fdb8c6f70cfa0b30c8497a1dd91cf094373e2cf\"" Dec 16 12:48:08.514469 containerd[1623]: time="2025-12-16T12:48:08.514424988Z" level=info msg="connecting to shim 238897f00b3e4c45fd7e38d99fdb8c6f70cfa0b30c8497a1dd91cf094373e2cf" address="unix:///run/containerd/s/118132e030fad33a655a9cac733eb4e13d7db8654e75ce8185154114c5f5146a" protocol=ttrpc version=3 Dec 16 12:48:08.554187 systemd[1]: Started cri-containerd-238897f00b3e4c45fd7e38d99fdb8c6f70cfa0b30c8497a1dd91cf094373e2cf.scope - libcontainer container 238897f00b3e4c45fd7e38d99fdb8c6f70cfa0b30c8497a1dd91cf094373e2cf. Dec 16 12:48:08.570049 kernel: kauditd_printk_skb: 75 callbacks suppressed Dec 16 12:48:08.570119 kernel: audit: type=1334 audit(1765889288.566:556): prog-id=161 op=LOAD Dec 16 12:48:08.566000 audit: BPF prog-id=161 op=LOAD Dec 16 12:48:08.572102 kernel: audit: type=1334 audit(1765889288.569:557): prog-id=162 op=LOAD Dec 16 12:48:08.569000 audit: BPF prog-id=162 op=LOAD Dec 16 12:48:08.580140 kernel: audit: type=1300 audit(1765889288.569:557): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=3228 pid=3395 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:08.569000 audit[3395]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=3228 pid=3395 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:08.587521 kernel: audit: type=1327 audit(1765889288.569:557): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3233383839376630306233653463343566643765333864393966646238 Dec 16 12:48:08.569000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3233383839376630306233653463343566643765333864393966646238 Dec 16 12:48:08.589846 kernel: audit: type=1334 audit(1765889288.569:558): prog-id=162 op=UNLOAD Dec 16 12:48:08.569000 audit: BPF prog-id=162 op=UNLOAD Dec 16 12:48:08.597095 kernel: audit: type=1300 audit(1765889288.569:558): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3228 pid=3395 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:08.569000 audit[3395]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3228 pid=3395 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:08.569000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3233383839376630306233653463343566643765333864393966646238 Dec 16 12:48:08.569000 audit: BPF prog-id=163 op=LOAD Dec 16 12:48:08.605639 kernel: audit: type=1327 audit(1765889288.569:558): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3233383839376630306233653463343566643765333864393966646238 Dec 16 12:48:08.605796 kernel: audit: type=1334 audit(1765889288.569:559): prog-id=163 op=LOAD Dec 16 12:48:08.607198 kernel: audit: type=1300 audit(1765889288.569:559): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3228 pid=3395 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:08.569000 audit[3395]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3228 pid=3395 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:08.569000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3233383839376630306233653463343566643765333864393966646238 Dec 16 12:48:08.616878 kernel: audit: type=1327 audit(1765889288.569:559): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3233383839376630306233653463343566643765333864393966646238 Dec 16 12:48:08.569000 audit: BPF prog-id=164 op=LOAD Dec 16 12:48:08.569000 audit[3395]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3228 pid=3395 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:08.569000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3233383839376630306233653463343566643765333864393966646238 Dec 16 12:48:08.569000 audit: BPF prog-id=164 op=UNLOAD Dec 16 12:48:08.569000 audit[3395]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3228 pid=3395 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:08.569000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3233383839376630306233653463343566643765333864393966646238 Dec 16 12:48:08.569000 audit: BPF prog-id=163 op=UNLOAD Dec 16 12:48:08.569000 audit[3395]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3228 pid=3395 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:08.569000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3233383839376630306233653463343566643765333864393966646238 Dec 16 12:48:08.569000 audit: BPF prog-id=165 op=LOAD Dec 16 12:48:08.569000 audit[3395]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=3228 pid=3395 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:08.569000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3233383839376630306233653463343566643765333864393966646238 Dec 16 12:48:08.626081 containerd[1623]: time="2025-12-16T12:48:08.626055610Z" level=info msg="StartContainer for \"238897f00b3e4c45fd7e38d99fdb8c6f70cfa0b30c8497a1dd91cf094373e2cf\" returns successfully" Dec 16 12:48:09.302165 kubelet[2796]: E1216 12:48:09.302127 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.302165 kubelet[2796]: W1216 12:48:09.302149 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.304989 kubelet[2796]: E1216 12:48:09.304959 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.305196 kubelet[2796]: E1216 12:48:09.305167 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.305196 kubelet[2796]: W1216 12:48:09.305185 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.305196 kubelet[2796]: E1216 12:48:09.305200 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.306284 kubelet[2796]: E1216 12:48:09.306254 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.306284 kubelet[2796]: W1216 12:48:09.306269 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.306284 kubelet[2796]: E1216 12:48:09.306279 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.306480 kubelet[2796]: E1216 12:48:09.306443 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.306480 kubelet[2796]: W1216 12:48:09.306451 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.306480 kubelet[2796]: E1216 12:48:09.306459 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.306781 kubelet[2796]: E1216 12:48:09.306657 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.306781 kubelet[2796]: W1216 12:48:09.306677 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.306781 kubelet[2796]: E1216 12:48:09.306693 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.307155 kubelet[2796]: E1216 12:48:09.307139 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.307267 kubelet[2796]: W1216 12:48:09.307248 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.307366 kubelet[2796]: E1216 12:48:09.307348 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.307793 kubelet[2796]: E1216 12:48:09.307756 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.307976 kubelet[2796]: W1216 12:48:09.307873 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.307976 kubelet[2796]: E1216 12:48:09.307896 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.308388 kubelet[2796]: E1216 12:48:09.308269 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.308388 kubelet[2796]: W1216 12:48:09.308282 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.308388 kubelet[2796]: E1216 12:48:09.308291 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.308598 kubelet[2796]: E1216 12:48:09.308586 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.308779 kubelet[2796]: W1216 12:48:09.308679 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.308779 kubelet[2796]: E1216 12:48:09.308702 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.309083 kubelet[2796]: E1216 12:48:09.308988 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.309083 kubelet[2796]: W1216 12:48:09.309004 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.309326 kubelet[2796]: E1216 12:48:09.309264 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.309615 kubelet[2796]: E1216 12:48:09.309523 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.309615 kubelet[2796]: W1216 12:48:09.309539 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.309615 kubelet[2796]: E1216 12:48:09.309552 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.309903 kubelet[2796]: E1216 12:48:09.309853 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.309903 kubelet[2796]: W1216 12:48:09.309867 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.309903 kubelet[2796]: E1216 12:48:09.309881 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.310433 kubelet[2796]: E1216 12:48:09.310361 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.310433 kubelet[2796]: W1216 12:48:09.310377 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.310433 kubelet[2796]: E1216 12:48:09.310391 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.310762 kubelet[2796]: E1216 12:48:09.310704 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.310762 kubelet[2796]: W1216 12:48:09.310716 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.310762 kubelet[2796]: E1216 12:48:09.310725 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.311132 kubelet[2796]: E1216 12:48:09.311082 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.311132 kubelet[2796]: W1216 12:48:09.311098 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.311132 kubelet[2796]: E1216 12:48:09.311112 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.318288 kubelet[2796]: E1216 12:48:09.318179 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.318288 kubelet[2796]: W1216 12:48:09.318196 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.318288 kubelet[2796]: E1216 12:48:09.318227 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.319280 kubelet[2796]: E1216 12:48:09.319110 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.319280 kubelet[2796]: W1216 12:48:09.319123 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.319280 kubelet[2796]: E1216 12:48:09.319134 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.319903 kubelet[2796]: E1216 12:48:09.319890 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.319994 kubelet[2796]: W1216 12:48:09.319956 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.319994 kubelet[2796]: E1216 12:48:09.319976 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.321177 kubelet[2796]: I1216 12:48:09.320997 2796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-58cf6cb8d8-qvj4d" podStartSLOduration=2.099995422 podStartE2EDuration="4.32098611s" podCreationTimestamp="2025-12-16 12:48:05 +0000 UTC" firstStartedPulling="2025-12-16 12:48:06.261785587 +0000 UTC m=+19.195513594" lastFinishedPulling="2025-12-16 12:48:08.482776276 +0000 UTC m=+21.416504282" observedRunningTime="2025-12-16 12:48:09.31538119 +0000 UTC m=+22.249109217" watchObservedRunningTime="2025-12-16 12:48:09.32098611 +0000 UTC m=+22.254714116" Dec 16 12:48:09.321177 kubelet[2796]: E1216 12:48:09.321085 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.321177 kubelet[2796]: W1216 12:48:09.321096 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.321177 kubelet[2796]: E1216 12:48:09.321107 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.321509 kubelet[2796]: E1216 12:48:09.321421 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.321509 kubelet[2796]: W1216 12:48:09.321433 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.321509 kubelet[2796]: E1216 12:48:09.321443 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.321660 kubelet[2796]: E1216 12:48:09.321631 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.321660 kubelet[2796]: W1216 12:48:09.321645 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.321660 kubelet[2796]: E1216 12:48:09.321654 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.321836 kubelet[2796]: E1216 12:48:09.321808 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.321836 kubelet[2796]: W1216 12:48:09.321822 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.321836 kubelet[2796]: E1216 12:48:09.321831 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.322858 kubelet[2796]: E1216 12:48:09.322110 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.322858 kubelet[2796]: W1216 12:48:09.322139 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.322858 kubelet[2796]: E1216 12:48:09.322148 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.323151 kubelet[2796]: E1216 12:48:09.323125 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.323151 kubelet[2796]: W1216 12:48:09.323141 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.323151 kubelet[2796]: E1216 12:48:09.323150 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.323338 kubelet[2796]: E1216 12:48:09.323307 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.323338 kubelet[2796]: W1216 12:48:09.323320 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.323464 kubelet[2796]: E1216 12:48:09.323436 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.323543 kubelet[2796]: E1216 12:48:09.323530 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.323578 kubelet[2796]: W1216 12:48:09.323543 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.323578 kubelet[2796]: E1216 12:48:09.323565 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.323771 kubelet[2796]: E1216 12:48:09.323756 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.323771 kubelet[2796]: W1216 12:48:09.323768 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.323966 kubelet[2796]: E1216 12:48:09.323783 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.323966 kubelet[2796]: E1216 12:48:09.323901 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.323966 kubelet[2796]: W1216 12:48:09.323908 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.323966 kubelet[2796]: E1216 12:48:09.323915 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.324534 kubelet[2796]: E1216 12:48:09.324194 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.324534 kubelet[2796]: W1216 12:48:09.324202 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.324534 kubelet[2796]: E1216 12:48:09.324229 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.324806 kubelet[2796]: E1216 12:48:09.324779 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.324806 kubelet[2796]: W1216 12:48:09.324800 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.324876 kubelet[2796]: E1216 12:48:09.324825 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.325140 kubelet[2796]: E1216 12:48:09.325106 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.325140 kubelet[2796]: W1216 12:48:09.325126 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.325238 kubelet[2796]: E1216 12:48:09.325179 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.325429 kubelet[2796]: E1216 12:48:09.325404 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.325429 kubelet[2796]: W1216 12:48:09.325425 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.325800 kubelet[2796]: E1216 12:48:09.325440 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:09.325894 kubelet[2796]: E1216 12:48:09.325870 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:09.325894 kubelet[2796]: W1216 12:48:09.325884 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:09.325987 kubelet[2796]: E1216 12:48:09.325896 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.154209 kubelet[2796]: E1216 12:48:10.154110 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:48:10.162064 containerd[1623]: time="2025-12-16T12:48:10.161769880Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:48:10.162873 containerd[1623]: time="2025-12-16T12:48:10.162849873Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:10.164048 containerd[1623]: time="2025-12-16T12:48:10.163974772Z" level=info msg="ImageCreate event name:\"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:48:10.165918 containerd[1623]: time="2025-12-16T12:48:10.165877941Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:48:10.167131 containerd[1623]: time="2025-12-16T12:48:10.167081992Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" with image id \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\", size \"5941314\" in 1.683876401s" Dec 16 12:48:10.167861 containerd[1623]: time="2025-12-16T12:48:10.167836995Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" returns image reference \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\"" Dec 16 12:48:10.170891 containerd[1623]: time="2025-12-16T12:48:10.170813610Z" level=info msg="CreateContainer within sandbox \"31a2c3fb2359d25620ab9b707e5cae784a39bde50f5c6db376d2523e06a8681d\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Dec 16 12:48:10.179525 containerd[1623]: time="2025-12-16T12:48:10.179504983Z" level=info msg="Container 3d18b43e5573d40395b41632fdec5b8c9406c9daeabd3a7aa6d87ddc981d41b4: CDI devices from CRI Config.CDIDevices: []" Dec 16 12:48:10.203487 containerd[1623]: time="2025-12-16T12:48:10.203443308Z" level=info msg="CreateContainer within sandbox \"31a2c3fb2359d25620ab9b707e5cae784a39bde50f5c6db376d2523e06a8681d\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"3d18b43e5573d40395b41632fdec5b8c9406c9daeabd3a7aa6d87ddc981d41b4\"" Dec 16 12:48:10.205369 containerd[1623]: time="2025-12-16T12:48:10.205247159Z" level=info msg="StartContainer for \"3d18b43e5573d40395b41632fdec5b8c9406c9daeabd3a7aa6d87ddc981d41b4\"" Dec 16 12:48:10.207988 containerd[1623]: time="2025-12-16T12:48:10.207938188Z" level=info msg="connecting to shim 3d18b43e5573d40395b41632fdec5b8c9406c9daeabd3a7aa6d87ddc981d41b4" address="unix:///run/containerd/s/698b86754e3d1d4bc1220c168741fa4038a4fc0135498ef4de6454cf03124774" protocol=ttrpc version=3 Dec 16 12:48:10.252216 systemd[1]: Started cri-containerd-3d18b43e5573d40395b41632fdec5b8c9406c9daeabd3a7aa6d87ddc981d41b4.scope - libcontainer container 3d18b43e5573d40395b41632fdec5b8c9406c9daeabd3a7aa6d87ddc981d41b4. Dec 16 12:48:10.255000 audit[3487]: NETFILTER_CFG table=filter:117 family=2 entries=21 op=nft_register_rule pid=3487 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:10.255000 audit[3487]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7fff2f8e83b0 a2=0 a3=7fff2f8e839c items=0 ppid=2898 pid=3487 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:10.255000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:10.257000 audit[3487]: NETFILTER_CFG table=nat:118 family=2 entries=19 op=nft_register_chain pid=3487 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:10.257000 audit[3487]: SYSCALL arch=c000003e syscall=46 success=yes exit=6276 a0=3 a1=7fff2f8e83b0 a2=0 a3=7fff2f8e839c items=0 ppid=2898 pid=3487 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:10.257000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:10.302000 audit: BPF prog-id=166 op=LOAD Dec 16 12:48:10.302000 audit[3472]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=3320 pid=3472 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:10.302000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364313862343365353537336434303339356234313633326664656335 Dec 16 12:48:10.302000 audit: BPF prog-id=167 op=LOAD Dec 16 12:48:10.302000 audit[3472]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=3320 pid=3472 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:10.302000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364313862343365353537336434303339356234313633326664656335 Dec 16 12:48:10.302000 audit: BPF prog-id=167 op=UNLOAD Dec 16 12:48:10.302000 audit[3472]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3320 pid=3472 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:10.302000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364313862343365353537336434303339356234313633326664656335 Dec 16 12:48:10.302000 audit: BPF prog-id=166 op=UNLOAD Dec 16 12:48:10.302000 audit[3472]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3320 pid=3472 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:10.302000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364313862343365353537336434303339356234313633326664656335 Dec 16 12:48:10.302000 audit: BPF prog-id=168 op=LOAD Dec 16 12:48:10.302000 audit[3472]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=3320 pid=3472 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:10.302000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3364313862343365353537336434303339356234313633326664656335 Dec 16 12:48:10.326507 kubelet[2796]: E1216 12:48:10.326458 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.326507 kubelet[2796]: W1216 12:48:10.326481 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.326507 kubelet[2796]: E1216 12:48:10.326511 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.327317 kubelet[2796]: E1216 12:48:10.326693 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.327317 kubelet[2796]: W1216 12:48:10.326702 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.327317 kubelet[2796]: E1216 12:48:10.326718 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.327317 kubelet[2796]: E1216 12:48:10.326875 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.327317 kubelet[2796]: W1216 12:48:10.326883 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.327317 kubelet[2796]: E1216 12:48:10.326891 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.327317 kubelet[2796]: E1216 12:48:10.327139 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.327317 kubelet[2796]: W1216 12:48:10.327148 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.327317 kubelet[2796]: E1216 12:48:10.327156 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.327595 kubelet[2796]: E1216 12:48:10.327331 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.327595 kubelet[2796]: W1216 12:48:10.327343 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.327595 kubelet[2796]: E1216 12:48:10.327351 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.327595 kubelet[2796]: E1216 12:48:10.327494 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.327595 kubelet[2796]: W1216 12:48:10.327502 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.327595 kubelet[2796]: E1216 12:48:10.327509 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.327800 kubelet[2796]: E1216 12:48:10.327656 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.327800 kubelet[2796]: W1216 12:48:10.327669 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.327800 kubelet[2796]: E1216 12:48:10.327690 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.327893 kubelet[2796]: E1216 12:48:10.327847 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.327893 kubelet[2796]: W1216 12:48:10.327863 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.327893 kubelet[2796]: E1216 12:48:10.327871 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.328982 kubelet[2796]: E1216 12:48:10.328003 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.328982 kubelet[2796]: W1216 12:48:10.328074 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.328982 kubelet[2796]: E1216 12:48:10.328084 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.328982 kubelet[2796]: E1216 12:48:10.328216 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.328982 kubelet[2796]: W1216 12:48:10.328224 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.328982 kubelet[2796]: E1216 12:48:10.328232 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.328982 kubelet[2796]: E1216 12:48:10.328343 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.328982 kubelet[2796]: W1216 12:48:10.328350 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.328982 kubelet[2796]: E1216 12:48:10.328358 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.328982 kubelet[2796]: E1216 12:48:10.328474 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.330549 kubelet[2796]: W1216 12:48:10.328481 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.330549 kubelet[2796]: E1216 12:48:10.328488 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.330549 kubelet[2796]: E1216 12:48:10.328731 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.330549 kubelet[2796]: W1216 12:48:10.328740 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.330549 kubelet[2796]: E1216 12:48:10.328747 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.330549 kubelet[2796]: E1216 12:48:10.328887 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.330549 kubelet[2796]: W1216 12:48:10.328895 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.330549 kubelet[2796]: E1216 12:48:10.328902 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.330549 kubelet[2796]: E1216 12:48:10.329070 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.330549 kubelet[2796]: W1216 12:48:10.329079 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.330859 kubelet[2796]: E1216 12:48:10.329086 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.330859 kubelet[2796]: E1216 12:48:10.329314 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.330859 kubelet[2796]: W1216 12:48:10.329324 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.330859 kubelet[2796]: E1216 12:48:10.329332 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.330859 kubelet[2796]: E1216 12:48:10.329478 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.330859 kubelet[2796]: W1216 12:48:10.329486 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.330859 kubelet[2796]: E1216 12:48:10.329505 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.330859 kubelet[2796]: E1216 12:48:10.329740 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.330859 kubelet[2796]: W1216 12:48:10.329769 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.330859 kubelet[2796]: E1216 12:48:10.329791 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.332217 kubelet[2796]: E1216 12:48:10.330110 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.332217 kubelet[2796]: W1216 12:48:10.330133 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.332217 kubelet[2796]: E1216 12:48:10.330168 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.332217 kubelet[2796]: E1216 12:48:10.332147 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.332217 kubelet[2796]: W1216 12:48:10.332164 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.332217 kubelet[2796]: E1216 12:48:10.332191 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.332495 kubelet[2796]: E1216 12:48:10.332440 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.332495 kubelet[2796]: W1216 12:48:10.332450 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.332565 kubelet[2796]: E1216 12:48:10.332536 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.332649 kubelet[2796]: E1216 12:48:10.332627 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.332649 kubelet[2796]: W1216 12:48:10.332644 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.332749 kubelet[2796]: E1216 12:48:10.332729 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.332902 kubelet[2796]: E1216 12:48:10.332829 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.332902 kubelet[2796]: W1216 12:48:10.332840 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.332985 kubelet[2796]: E1216 12:48:10.332929 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.333158 kubelet[2796]: E1216 12:48:10.333097 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.333158 kubelet[2796]: W1216 12:48:10.333110 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.333158 kubelet[2796]: E1216 12:48:10.333130 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.333482 kubelet[2796]: E1216 12:48:10.333423 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.333482 kubelet[2796]: W1216 12:48:10.333436 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.333482 kubelet[2796]: E1216 12:48:10.333458 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.333833 kubelet[2796]: E1216 12:48:10.333663 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.333833 kubelet[2796]: W1216 12:48:10.333670 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.333833 kubelet[2796]: E1216 12:48:10.333679 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.334098 kubelet[2796]: E1216 12:48:10.333861 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.334098 kubelet[2796]: W1216 12:48:10.333869 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.334098 kubelet[2796]: E1216 12:48:10.333948 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.334098 kubelet[2796]: E1216 12:48:10.334067 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.334098 kubelet[2796]: W1216 12:48:10.334075 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.334098 kubelet[2796]: E1216 12:48:10.334085 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.334743 kubelet[2796]: E1216 12:48:10.334289 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.334743 kubelet[2796]: W1216 12:48:10.334298 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.334743 kubelet[2796]: E1216 12:48:10.334318 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.334743 kubelet[2796]: E1216 12:48:10.334546 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.334743 kubelet[2796]: W1216 12:48:10.334561 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.334743 kubelet[2796]: E1216 12:48:10.334594 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.335867 kubelet[2796]: E1216 12:48:10.335327 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.335867 kubelet[2796]: W1216 12:48:10.335341 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.335867 kubelet[2796]: E1216 12:48:10.335366 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.336272 kubelet[2796]: E1216 12:48:10.336235 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.336272 kubelet[2796]: W1216 12:48:10.336252 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.336272 kubelet[2796]: E1216 12:48:10.336266 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.336941 kubelet[2796]: E1216 12:48:10.336909 2796 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Dec 16 12:48:10.336941 kubelet[2796]: W1216 12:48:10.336931 2796 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Dec 16 12:48:10.337042 kubelet[2796]: E1216 12:48:10.336945 2796 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Dec 16 12:48:10.344288 containerd[1623]: time="2025-12-16T12:48:10.344237619Z" level=info msg="StartContainer for \"3d18b43e5573d40395b41632fdec5b8c9406c9daeabd3a7aa6d87ddc981d41b4\" returns successfully" Dec 16 12:48:10.357761 systemd[1]: cri-containerd-3d18b43e5573d40395b41632fdec5b8c9406c9daeabd3a7aa6d87ddc981d41b4.scope: Deactivated successfully. Dec 16 12:48:10.361000 audit: BPF prog-id=168 op=UNLOAD Dec 16 12:48:10.391580 containerd[1623]: time="2025-12-16T12:48:10.391530527Z" level=info msg="received container exit event container_id:\"3d18b43e5573d40395b41632fdec5b8c9406c9daeabd3a7aa6d87ddc981d41b4\" id:\"3d18b43e5573d40395b41632fdec5b8c9406c9daeabd3a7aa6d87ddc981d41b4\" pid:3489 exited_at:{seconds:1765889290 nanos:363951723}" Dec 16 12:48:10.413938 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3d18b43e5573d40395b41632fdec5b8c9406c9daeabd3a7aa6d87ddc981d41b4-rootfs.mount: Deactivated successfully. Dec 16 12:48:11.294737 containerd[1623]: time="2025-12-16T12:48:11.294675737Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\"" Dec 16 12:48:12.153972 kubelet[2796]: E1216 12:48:12.153679 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:48:13.893774 containerd[1623]: time="2025-12-16T12:48:13.893720698Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:48:13.894753 containerd[1623]: time="2025-12-16T12:48:13.894674338Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.4: active requests=0, bytes read=70442291" Dec 16 12:48:13.896069 containerd[1623]: time="2025-12-16T12:48:13.895528647Z" level=info msg="ImageCreate event name:\"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:48:13.897223 containerd[1623]: time="2025-12-16T12:48:13.897183759Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:48:13.898067 containerd[1623]: time="2025-12-16T12:48:13.897594978Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.4\" with image id \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\", size \"71941459\" in 2.602883025s" Dec 16 12:48:13.898067 containerd[1623]: time="2025-12-16T12:48:13.897637326Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\" returns image reference \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\"" Dec 16 12:48:13.901089 containerd[1623]: time="2025-12-16T12:48:13.900376832Z" level=info msg="CreateContainer within sandbox \"31a2c3fb2359d25620ab9b707e5cae784a39bde50f5c6db376d2523e06a8681d\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Dec 16 12:48:13.914042 containerd[1623]: time="2025-12-16T12:48:13.913982715Z" level=info msg="Container 14219fa0c502e4015c002312a6f20bcf58a51ac217bf319f631928d695ca8743: CDI devices from CRI Config.CDIDevices: []" Dec 16 12:48:13.922613 containerd[1623]: time="2025-12-16T12:48:13.922575338Z" level=info msg="CreateContainer within sandbox \"31a2c3fb2359d25620ab9b707e5cae784a39bde50f5c6db376d2523e06a8681d\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"14219fa0c502e4015c002312a6f20bcf58a51ac217bf319f631928d695ca8743\"" Dec 16 12:48:13.928934 containerd[1623]: time="2025-12-16T12:48:13.928899109Z" level=info msg="StartContainer for \"14219fa0c502e4015c002312a6f20bcf58a51ac217bf319f631928d695ca8743\"" Dec 16 12:48:13.930236 containerd[1623]: time="2025-12-16T12:48:13.930205370Z" level=info msg="connecting to shim 14219fa0c502e4015c002312a6f20bcf58a51ac217bf319f631928d695ca8743" address="unix:///run/containerd/s/698b86754e3d1d4bc1220c168741fa4038a4fc0135498ef4de6454cf03124774" protocol=ttrpc version=3 Dec 16 12:48:13.952603 systemd[1]: Started cri-containerd-14219fa0c502e4015c002312a6f20bcf58a51ac217bf319f631928d695ca8743.scope - libcontainer container 14219fa0c502e4015c002312a6f20bcf58a51ac217bf319f631928d695ca8743. Dec 16 12:48:13.996000 audit: BPF prog-id=169 op=LOAD Dec 16 12:48:14.003275 kernel: kauditd_printk_skb: 34 callbacks suppressed Dec 16 12:48:14.003382 kernel: audit: type=1334 audit(1765889293.996:572): prog-id=169 op=LOAD Dec 16 12:48:14.014066 kernel: audit: type=1300 audit(1765889293.996:572): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001b0488 a2=98 a3=0 items=0 ppid=3320 pid=3571 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:13.996000 audit[3571]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001b0488 a2=98 a3=0 items=0 ppid=3320 pid=3571 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:14.015824 kernel: audit: type=1327 audit(1765889293.996:572): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134323139666130633530326534303135633030323331326136663230 Dec 16 12:48:13.996000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134323139666130633530326534303135633030323331326136663230 Dec 16 12:48:13.996000 audit: BPF prog-id=170 op=LOAD Dec 16 12:48:14.031053 kernel: audit: type=1334 audit(1765889293.996:573): prog-id=170 op=LOAD Dec 16 12:48:13.996000 audit[3571]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001b0218 a2=98 a3=0 items=0 ppid=3320 pid=3571 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:13.996000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134323139666130633530326534303135633030323331326136663230 Dec 16 12:48:14.046301 kernel: audit: type=1300 audit(1765889293.996:573): arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c0001b0218 a2=98 a3=0 items=0 ppid=3320 pid=3571 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:14.046358 kernel: audit: type=1327 audit(1765889293.996:573): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134323139666130633530326534303135633030323331326136663230 Dec 16 12:48:14.054942 kernel: audit: type=1334 audit(1765889293.996:574): prog-id=170 op=UNLOAD Dec 16 12:48:13.996000 audit: BPF prog-id=170 op=UNLOAD Dec 16 12:48:13.996000 audit[3571]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3320 pid=3571 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:14.062604 containerd[1623]: time="2025-12-16T12:48:14.062473494Z" level=info msg="StartContainer for \"14219fa0c502e4015c002312a6f20bcf58a51ac217bf319f631928d695ca8743\" returns successfully" Dec 16 12:48:14.070065 kernel: audit: type=1300 audit(1765889293.996:574): arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3320 pid=3571 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:13.996000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134323139666130633530326534303135633030323331326136663230 Dec 16 12:48:14.092165 kernel: audit: type=1327 audit(1765889293.996:574): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134323139666130633530326534303135633030323331326136663230 Dec 16 12:48:14.092338 kernel: audit: type=1334 audit(1765889293.996:575): prog-id=169 op=UNLOAD Dec 16 12:48:13.996000 audit: BPF prog-id=169 op=UNLOAD Dec 16 12:48:13.996000 audit[3571]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3320 pid=3571 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:13.996000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134323139666130633530326534303135633030323331326136663230 Dec 16 12:48:13.996000 audit: BPF prog-id=171 op=LOAD Dec 16 12:48:13.996000 audit[3571]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001b06e8 a2=98 a3=0 items=0 ppid=3320 pid=3571 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:13.996000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3134323139666130633530326534303135633030323331326136663230 Dec 16 12:48:14.158176 kubelet[2796]: E1216 12:48:14.157345 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:48:14.661573 systemd[1]: cri-containerd-14219fa0c502e4015c002312a6f20bcf58a51ac217bf319f631928d695ca8743.scope: Deactivated successfully. Dec 16 12:48:14.663260 systemd[1]: cri-containerd-14219fa0c502e4015c002312a6f20bcf58a51ac217bf319f631928d695ca8743.scope: Consumed 528ms CPU time, 164.6M memory peak, 10.8M read from disk, 171.3M written to disk. Dec 16 12:48:14.665000 audit: BPF prog-id=171 op=UNLOAD Dec 16 12:48:14.672722 containerd[1623]: time="2025-12-16T12:48:14.672560470Z" level=info msg="received container exit event container_id:\"14219fa0c502e4015c002312a6f20bcf58a51ac217bf319f631928d695ca8743\" id:\"14219fa0c502e4015c002312a6f20bcf58a51ac217bf319f631928d695ca8743\" pid:3583 exited_at:{seconds:1765889294 nanos:672241775}" Dec 16 12:48:14.716483 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-14219fa0c502e4015c002312a6f20bcf58a51ac217bf319f631928d695ca8743-rootfs.mount: Deactivated successfully. Dec 16 12:48:14.746888 kubelet[2796]: I1216 12:48:14.746862 2796 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Dec 16 12:48:14.801904 systemd[1]: Created slice kubepods-burstable-podd74a60d1_a1a7_4b5b_ad38_2c2ebbe76fd5.slice - libcontainer container kubepods-burstable-podd74a60d1_a1a7_4b5b_ad38_2c2ebbe76fd5.slice. Dec 16 12:48:14.820460 systemd[1]: Created slice kubepods-burstable-podb556e31a_1d4e_41c7_bc84_fbb790df238f.slice - libcontainer container kubepods-burstable-podb556e31a_1d4e_41c7_bc84_fbb790df238f.slice. Dec 16 12:48:14.826813 systemd[1]: Created slice kubepods-besteffort-pod6200e57e_82e7_4316_9456_f3714e3f66da.slice - libcontainer container kubepods-besteffort-pod6200e57e_82e7_4316_9456_f3714e3f66da.slice. Dec 16 12:48:14.833709 systemd[1]: Created slice kubepods-besteffort-poddd8550e8_a28d_4154_8fdc_fe7811fd3803.slice - libcontainer container kubepods-besteffort-poddd8550e8_a28d_4154_8fdc_fe7811fd3803.slice. Dec 16 12:48:14.841759 systemd[1]: Created slice kubepods-besteffort-pod0e649fa6_ddd7_43a4_b265_cc0585bd2a9e.slice - libcontainer container kubepods-besteffort-pod0e649fa6_ddd7_43a4_b265_cc0585bd2a9e.slice. Dec 16 12:48:14.852178 systemd[1]: Created slice kubepods-besteffort-pod9e36b6d0_206d_4ded_90fa_9c24767dcdf8.slice - libcontainer container kubepods-besteffort-pod9e36b6d0_206d_4ded_90fa_9c24767dcdf8.slice. Dec 16 12:48:14.858567 systemd[1]: Created slice kubepods-besteffort-podf4651243_c791_471b_a364_a0f9a5a677b8.slice - libcontainer container kubepods-besteffort-podf4651243_c791_471b_a364_a0f9a5a677b8.slice. Dec 16 12:48:14.864573 kubelet[2796]: I1216 12:48:14.864546 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlcq8\" (UniqueName: \"kubernetes.io/projected/b556e31a-1d4e-41c7-bc84-fbb790df238f-kube-api-access-tlcq8\") pod \"coredns-668d6bf9bc-48cft\" (UID: \"b556e31a-1d4e-41c7-bc84-fbb790df238f\") " pod="kube-system/coredns-668d6bf9bc-48cft" Dec 16 12:48:14.864663 kubelet[2796]: I1216 12:48:14.864580 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcpvr\" (UniqueName: \"kubernetes.io/projected/6200e57e-82e7-4316-9456-f3714e3f66da-kube-api-access-gcpvr\") pod \"calico-kube-controllers-98457786-pp6vg\" (UID: \"6200e57e-82e7-4316-9456-f3714e3f66da\") " pod="calico-system/calico-kube-controllers-98457786-pp6vg" Dec 16 12:48:14.864663 kubelet[2796]: I1216 12:48:14.864600 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp8j9\" (UniqueName: \"kubernetes.io/projected/dd8550e8-a28d-4154-8fdc-fe7811fd3803-kube-api-access-sp8j9\") pod \"calico-apiserver-5668559856-lv4tb\" (UID: \"dd8550e8-a28d-4154-8fdc-fe7811fd3803\") " pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" Dec 16 12:48:14.864663 kubelet[2796]: I1216 12:48:14.864621 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gsv7\" (UniqueName: \"kubernetes.io/projected/f4651243-c791-471b-a364-a0f9a5a677b8-kube-api-access-9gsv7\") pod \"calico-apiserver-5668559856-xl9w8\" (UID: \"f4651243-c791-471b-a364-a0f9a5a677b8\") " pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" Dec 16 12:48:14.864663 kubelet[2796]: I1216 12:48:14.864650 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/0e649fa6-ddd7-43a4-b265-cc0585bd2a9e-goldmane-key-pair\") pod \"goldmane-666569f655-5hslq\" (UID: \"0e649fa6-ddd7-43a4-b265-cc0585bd2a9e\") " pod="calico-system/goldmane-666569f655-5hslq" Dec 16 12:48:14.864747 kubelet[2796]: I1216 12:48:14.864670 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/dd8550e8-a28d-4154-8fdc-fe7811fd3803-calico-apiserver-certs\") pod \"calico-apiserver-5668559856-lv4tb\" (UID: \"dd8550e8-a28d-4154-8fdc-fe7811fd3803\") " pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" Dec 16 12:48:14.864747 kubelet[2796]: I1216 12:48:14.864685 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d74a60d1-a1a7-4b5b-ad38-2c2ebbe76fd5-config-volume\") pod \"coredns-668d6bf9bc-srwr6\" (UID: \"d74a60d1-a1a7-4b5b-ad38-2c2ebbe76fd5\") " pod="kube-system/coredns-668d6bf9bc-srwr6" Dec 16 12:48:14.864747 kubelet[2796]: I1216 12:48:14.864697 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e649fa6-ddd7-43a4-b265-cc0585bd2a9e-goldmane-ca-bundle\") pod \"goldmane-666569f655-5hslq\" (UID: \"0e649fa6-ddd7-43a4-b265-cc0585bd2a9e\") " pod="calico-system/goldmane-666569f655-5hslq" Dec 16 12:48:14.864747 kubelet[2796]: I1216 12:48:14.864714 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e649fa6-ddd7-43a4-b265-cc0585bd2a9e-config\") pod \"goldmane-666569f655-5hslq\" (UID: \"0e649fa6-ddd7-43a4-b265-cc0585bd2a9e\") " pod="calico-system/goldmane-666569f655-5hslq" Dec 16 12:48:14.864747 kubelet[2796]: I1216 12:48:14.864728 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw75k\" (UniqueName: \"kubernetes.io/projected/0e649fa6-ddd7-43a4-b265-cc0585bd2a9e-kube-api-access-kw75k\") pod \"goldmane-666569f655-5hslq\" (UID: \"0e649fa6-ddd7-43a4-b265-cc0585bd2a9e\") " pod="calico-system/goldmane-666569f655-5hslq" Dec 16 12:48:14.864834 kubelet[2796]: I1216 12:48:14.864744 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b556e31a-1d4e-41c7-bc84-fbb790df238f-config-volume\") pod \"coredns-668d6bf9bc-48cft\" (UID: \"b556e31a-1d4e-41c7-bc84-fbb790df238f\") " pod="kube-system/coredns-668d6bf9bc-48cft" Dec 16 12:48:14.864834 kubelet[2796]: I1216 12:48:14.864759 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6200e57e-82e7-4316-9456-f3714e3f66da-tigera-ca-bundle\") pod \"calico-kube-controllers-98457786-pp6vg\" (UID: \"6200e57e-82e7-4316-9456-f3714e3f66da\") " pod="calico-system/calico-kube-controllers-98457786-pp6vg" Dec 16 12:48:14.864834 kubelet[2796]: I1216 12:48:14.864772 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/9e36b6d0-206d-4ded-90fa-9c24767dcdf8-whisker-backend-key-pair\") pod \"whisker-6c49df4495-r24st\" (UID: \"9e36b6d0-206d-4ded-90fa-9c24767dcdf8\") " pod="calico-system/whisker-6c49df4495-r24st" Dec 16 12:48:14.864834 kubelet[2796]: I1216 12:48:14.864786 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cztt7\" (UniqueName: \"kubernetes.io/projected/9e36b6d0-206d-4ded-90fa-9c24767dcdf8-kube-api-access-cztt7\") pod \"whisker-6c49df4495-r24st\" (UID: \"9e36b6d0-206d-4ded-90fa-9c24767dcdf8\") " pod="calico-system/whisker-6c49df4495-r24st" Dec 16 12:48:14.864834 kubelet[2796]: I1216 12:48:14.864804 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6jq9\" (UniqueName: \"kubernetes.io/projected/d74a60d1-a1a7-4b5b-ad38-2c2ebbe76fd5-kube-api-access-f6jq9\") pod \"coredns-668d6bf9bc-srwr6\" (UID: \"d74a60d1-a1a7-4b5b-ad38-2c2ebbe76fd5\") " pod="kube-system/coredns-668d6bf9bc-srwr6" Dec 16 12:48:14.864936 kubelet[2796]: I1216 12:48:14.864820 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/f4651243-c791-471b-a364-a0f9a5a677b8-calico-apiserver-certs\") pod \"calico-apiserver-5668559856-xl9w8\" (UID: \"f4651243-c791-471b-a364-a0f9a5a677b8\") " pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" Dec 16 12:48:14.864936 kubelet[2796]: I1216 12:48:14.864834 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9e36b6d0-206d-4ded-90fa-9c24767dcdf8-whisker-ca-bundle\") pod \"whisker-6c49df4495-r24st\" (UID: \"9e36b6d0-206d-4ded-90fa-9c24767dcdf8\") " pod="calico-system/whisker-6c49df4495-r24st" Dec 16 12:48:15.109994 containerd[1623]: time="2025-12-16T12:48:15.109929043Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-srwr6,Uid:d74a60d1-a1a7-4b5b-ad38-2c2ebbe76fd5,Namespace:kube-system,Attempt:0,}" Dec 16 12:48:15.125945 containerd[1623]: time="2025-12-16T12:48:15.125865845Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-48cft,Uid:b556e31a-1d4e-41c7-bc84-fbb790df238f,Namespace:kube-system,Attempt:0,}" Dec 16 12:48:15.140350 containerd[1623]: time="2025-12-16T12:48:15.138141916Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-98457786-pp6vg,Uid:6200e57e-82e7-4316-9456-f3714e3f66da,Namespace:calico-system,Attempt:0,}" Dec 16 12:48:15.140572 containerd[1623]: time="2025-12-16T12:48:15.140511043Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5668559856-lv4tb,Uid:dd8550e8-a28d-4154-8fdc-fe7811fd3803,Namespace:calico-apiserver,Attempt:0,}" Dec 16 12:48:15.177657 containerd[1623]: time="2025-12-16T12:48:15.177482285Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-5hslq,Uid:0e649fa6-ddd7-43a4-b265-cc0585bd2a9e,Namespace:calico-system,Attempt:0,}" Dec 16 12:48:15.179665 containerd[1623]: time="2025-12-16T12:48:15.179606131Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6c49df4495-r24st,Uid:9e36b6d0-206d-4ded-90fa-9c24767dcdf8,Namespace:calico-system,Attempt:0,}" Dec 16 12:48:15.190883 containerd[1623]: time="2025-12-16T12:48:15.190824474Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5668559856-xl9w8,Uid:f4651243-c791-471b-a364-a0f9a5a677b8,Namespace:calico-apiserver,Attempt:0,}" Dec 16 12:48:15.329168 containerd[1623]: time="2025-12-16T12:48:15.329140436Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\"" Dec 16 12:48:15.412353 containerd[1623]: time="2025-12-16T12:48:15.412249469Z" level=error msg="Failed to destroy network for sandbox \"e3432a303706d1148df0e5144316d555e85a96f9d21ecd27565474c92b62d4e3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.415355 containerd[1623]: time="2025-12-16T12:48:15.415323057Z" level=error msg="Failed to destroy network for sandbox \"9081519a19a929587fcb4010e5ee523cbe4860b50c4dcf4e5ee1f672a6cd3ac0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.415667 containerd[1623]: time="2025-12-16T12:48:15.415644522Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6c49df4495-r24st,Uid:9e36b6d0-206d-4ded-90fa-9c24767dcdf8,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3432a303706d1148df0e5144316d555e85a96f9d21ecd27565474c92b62d4e3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.417237 kubelet[2796]: E1216 12:48:15.417201 2796 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3432a303706d1148df0e5144316d555e85a96f9d21ecd27565474c92b62d4e3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.417580 kubelet[2796]: E1216 12:48:15.417562 2796 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3432a303706d1148df0e5144316d555e85a96f9d21ecd27565474c92b62d4e3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-6c49df4495-r24st" Dec 16 12:48:15.418809 kubelet[2796]: E1216 12:48:15.418551 2796 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3432a303706d1148df0e5144316d555e85a96f9d21ecd27565474c92b62d4e3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-6c49df4495-r24st" Dec 16 12:48:15.418809 kubelet[2796]: E1216 12:48:15.418625 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-6c49df4495-r24st_calico-system(9e36b6d0-206d-4ded-90fa-9c24767dcdf8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-6c49df4495-r24st_calico-system(9e36b6d0-206d-4ded-90fa-9c24767dcdf8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e3432a303706d1148df0e5144316d555e85a96f9d21ecd27565474c92b62d4e3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-6c49df4495-r24st" podUID="9e36b6d0-206d-4ded-90fa-9c24767dcdf8" Dec 16 12:48:15.421481 containerd[1623]: time="2025-12-16T12:48:15.421448453Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-5hslq,Uid:0e649fa6-ddd7-43a4-b265-cc0585bd2a9e,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"9081519a19a929587fcb4010e5ee523cbe4860b50c4dcf4e5ee1f672a6cd3ac0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.431204 containerd[1623]: time="2025-12-16T12:48:15.431162797Z" level=error msg="Failed to destroy network for sandbox \"61239515a226e9a772abf7ca02897278b3895f396b607f6c2c52d48380040165\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.440125 kubelet[2796]: E1216 12:48:15.439243 2796 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9081519a19a929587fcb4010e5ee523cbe4860b50c4dcf4e5ee1f672a6cd3ac0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.440125 kubelet[2796]: E1216 12:48:15.439295 2796 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9081519a19a929587fcb4010e5ee523cbe4860b50c4dcf4e5ee1f672a6cd3ac0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-5hslq" Dec 16 12:48:15.440125 kubelet[2796]: E1216 12:48:15.439314 2796 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9081519a19a929587fcb4010e5ee523cbe4860b50c4dcf4e5ee1f672a6cd3ac0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-5hslq" Dec 16 12:48:15.440773 kubelet[2796]: E1216 12:48:15.439348 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-5hslq_calico-system(0e649fa6-ddd7-43a4-b265-cc0585bd2a9e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-5hslq_calico-system(0e649fa6-ddd7-43a4-b265-cc0585bd2a9e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9081519a19a929587fcb4010e5ee523cbe4860b50c4dcf4e5ee1f672a6cd3ac0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-5hslq" podUID="0e649fa6-ddd7-43a4-b265-cc0585bd2a9e" Dec 16 12:48:15.440852 containerd[1623]: time="2025-12-16T12:48:15.440218244Z" level=error msg="Failed to destroy network for sandbox \"efc8b82039441527d4b9cc8472f6dcdaf6298f570e0ba3d27de7cb62ff193c11\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.445656 containerd[1623]: time="2025-12-16T12:48:15.445617781Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5668559856-xl9w8,Uid:f4651243-c791-471b-a364-a0f9a5a677b8,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"efc8b82039441527d4b9cc8472f6dcdaf6298f570e0ba3d27de7cb62ff193c11\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.446045 containerd[1623]: time="2025-12-16T12:48:15.445863252Z" level=error msg="Failed to destroy network for sandbox \"49c03226ae11a9ac5f4934f7ed67f8582606d24c7c53e3e09b44398da4ac5f45\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.446219 kubelet[2796]: E1216 12:48:15.445997 2796 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"efc8b82039441527d4b9cc8472f6dcdaf6298f570e0ba3d27de7cb62ff193c11\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.446219 kubelet[2796]: E1216 12:48:15.446140 2796 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"efc8b82039441527d4b9cc8472f6dcdaf6298f570e0ba3d27de7cb62ff193c11\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" Dec 16 12:48:15.446219 kubelet[2796]: E1216 12:48:15.446159 2796 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"efc8b82039441527d4b9cc8472f6dcdaf6298f570e0ba3d27de7cb62ff193c11\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" Dec 16 12:48:15.446625 kubelet[2796]: E1216 12:48:15.446196 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-5668559856-xl9w8_calico-apiserver(f4651243-c791-471b-a364-a0f9a5a677b8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-5668559856-xl9w8_calico-apiserver(f4651243-c791-471b-a364-a0f9a5a677b8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"efc8b82039441527d4b9cc8472f6dcdaf6298f570e0ba3d27de7cb62ff193c11\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" podUID="f4651243-c791-471b-a364-a0f9a5a677b8" Dec 16 12:48:15.448583 containerd[1623]: time="2025-12-16T12:48:15.448497476Z" level=error msg="Failed to destroy network for sandbox \"631a0dd6ae6716e9f09e55975af8a598df04a88f6ed7f5d1986663c8f3d261bf\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.448935 containerd[1623]: time="2025-12-16T12:48:15.448881634Z" level=error msg="Failed to destroy network for sandbox \"88eb002235eb0e0502eb465f4b3c97737b922e313cbfe4fbfb9f1bae180ddd38\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.449622 containerd[1623]: time="2025-12-16T12:48:15.449543657Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-srwr6,Uid:d74a60d1-a1a7-4b5b-ad38-2c2ebbe76fd5,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"49c03226ae11a9ac5f4934f7ed67f8582606d24c7c53e3e09b44398da4ac5f45\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.449818 kubelet[2796]: E1216 12:48:15.449786 2796 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"49c03226ae11a9ac5f4934f7ed67f8582606d24c7c53e3e09b44398da4ac5f45\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.449818 kubelet[2796]: E1216 12:48:15.449817 2796 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"49c03226ae11a9ac5f4934f7ed67f8582606d24c7c53e3e09b44398da4ac5f45\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-srwr6" Dec 16 12:48:15.449895 kubelet[2796]: E1216 12:48:15.449832 2796 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"49c03226ae11a9ac5f4934f7ed67f8582606d24c7c53e3e09b44398da4ac5f45\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-srwr6" Dec 16 12:48:15.449895 kubelet[2796]: E1216 12:48:15.449883 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-srwr6_kube-system(d74a60d1-a1a7-4b5b-ad38-2c2ebbe76fd5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-srwr6_kube-system(d74a60d1-a1a7-4b5b-ad38-2c2ebbe76fd5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"49c03226ae11a9ac5f4934f7ed67f8582606d24c7c53e3e09b44398da4ac5f45\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-srwr6" podUID="d74a60d1-a1a7-4b5b-ad38-2c2ebbe76fd5" Dec 16 12:48:15.451230 containerd[1623]: time="2025-12-16T12:48:15.451151273Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5668559856-lv4tb,Uid:dd8550e8-a28d-4154-8fdc-fe7811fd3803,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"88eb002235eb0e0502eb465f4b3c97737b922e313cbfe4fbfb9f1bae180ddd38\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.451555 kubelet[2796]: E1216 12:48:15.451456 2796 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"88eb002235eb0e0502eb465f4b3c97737b922e313cbfe4fbfb9f1bae180ddd38\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.451555 kubelet[2796]: E1216 12:48:15.451549 2796 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"88eb002235eb0e0502eb465f4b3c97737b922e313cbfe4fbfb9f1bae180ddd38\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" Dec 16 12:48:15.451624 kubelet[2796]: E1216 12:48:15.451583 2796 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"88eb002235eb0e0502eb465f4b3c97737b922e313cbfe4fbfb9f1bae180ddd38\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" Dec 16 12:48:15.451624 kubelet[2796]: E1216 12:48:15.451612 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-5668559856-lv4tb_calico-apiserver(dd8550e8-a28d-4154-8fdc-fe7811fd3803)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-5668559856-lv4tb_calico-apiserver(dd8550e8-a28d-4154-8fdc-fe7811fd3803)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"88eb002235eb0e0502eb465f4b3c97737b922e313cbfe4fbfb9f1bae180ddd38\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" podUID="dd8550e8-a28d-4154-8fdc-fe7811fd3803" Dec 16 12:48:15.452077 containerd[1623]: time="2025-12-16T12:48:15.451989521Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-48cft,Uid:b556e31a-1d4e-41c7-bc84-fbb790df238f,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"61239515a226e9a772abf7ca02897278b3895f396b607f6c2c52d48380040165\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.452266 kubelet[2796]: E1216 12:48:15.452250 2796 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"61239515a226e9a772abf7ca02897278b3895f396b607f6c2c52d48380040165\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.452384 kubelet[2796]: E1216 12:48:15.452369 2796 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"61239515a226e9a772abf7ca02897278b3895f396b607f6c2c52d48380040165\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-48cft" Dec 16 12:48:15.452722 kubelet[2796]: E1216 12:48:15.452689 2796 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"61239515a226e9a772abf7ca02897278b3895f396b607f6c2c52d48380040165\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-48cft" Dec 16 12:48:15.452834 kubelet[2796]: E1216 12:48:15.452807 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-48cft_kube-system(b556e31a-1d4e-41c7-bc84-fbb790df238f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-48cft_kube-system(b556e31a-1d4e-41c7-bc84-fbb790df238f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"61239515a226e9a772abf7ca02897278b3895f396b607f6c2c52d48380040165\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-48cft" podUID="b556e31a-1d4e-41c7-bc84-fbb790df238f" Dec 16 12:48:15.453414 containerd[1623]: time="2025-12-16T12:48:15.453383258Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-98457786-pp6vg,Uid:6200e57e-82e7-4316-9456-f3714e3f66da,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"631a0dd6ae6716e9f09e55975af8a598df04a88f6ed7f5d1986663c8f3d261bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.453681 kubelet[2796]: E1216 12:48:15.453651 2796 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"631a0dd6ae6716e9f09e55975af8a598df04a88f6ed7f5d1986663c8f3d261bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:15.453720 kubelet[2796]: E1216 12:48:15.453680 2796 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"631a0dd6ae6716e9f09e55975af8a598df04a88f6ed7f5d1986663c8f3d261bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-98457786-pp6vg" Dec 16 12:48:15.453720 kubelet[2796]: E1216 12:48:15.453693 2796 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"631a0dd6ae6716e9f09e55975af8a598df04a88f6ed7f5d1986663c8f3d261bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-98457786-pp6vg" Dec 16 12:48:15.453761 kubelet[2796]: E1216 12:48:15.453728 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-98457786-pp6vg_calico-system(6200e57e-82e7-4316-9456-f3714e3f66da)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-98457786-pp6vg_calico-system(6200e57e-82e7-4316-9456-f3714e3f66da)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"631a0dd6ae6716e9f09e55975af8a598df04a88f6ed7f5d1986663c8f3d261bf\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-98457786-pp6vg" podUID="6200e57e-82e7-4316-9456-f3714e3f66da" Dec 16 12:48:16.163820 systemd[1]: Created slice kubepods-besteffort-pod2053803c_11b9_48f6_949c_d0d885185bfe.slice - libcontainer container kubepods-besteffort-pod2053803c_11b9_48f6_949c_d0d885185bfe.slice. Dec 16 12:48:16.171529 containerd[1623]: time="2025-12-16T12:48:16.171462589Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-m8p7d,Uid:2053803c-11b9-48f6-949c-d0d885185bfe,Namespace:calico-system,Attempt:0,}" Dec 16 12:48:16.268043 containerd[1623]: time="2025-12-16T12:48:16.267891196Z" level=error msg="Failed to destroy network for sandbox \"296ce5af9f3b6065a0c9a895a0c9c3732f513ecc58171f89f3e25e918c709e78\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:16.271840 containerd[1623]: time="2025-12-16T12:48:16.271677013Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-m8p7d,Uid:2053803c-11b9-48f6-949c-d0d885185bfe,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"296ce5af9f3b6065a0c9a895a0c9c3732f513ecc58171f89f3e25e918c709e78\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:16.273001 kubelet[2796]: E1216 12:48:16.272842 2796 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"296ce5af9f3b6065a0c9a895a0c9c3732f513ecc58171f89f3e25e918c709e78\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Dec 16 12:48:16.273337 kubelet[2796]: E1216 12:48:16.273138 2796 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"296ce5af9f3b6065a0c9a895a0c9c3732f513ecc58171f89f3e25e918c709e78\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-m8p7d" Dec 16 12:48:16.273747 kubelet[2796]: E1216 12:48:16.273587 2796 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"296ce5af9f3b6065a0c9a895a0c9c3732f513ecc58171f89f3e25e918c709e78\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-m8p7d" Dec 16 12:48:16.274063 kubelet[2796]: E1216 12:48:16.273918 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-m8p7d_calico-system(2053803c-11b9-48f6-949c-d0d885185bfe)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-m8p7d_calico-system(2053803c-11b9-48f6-949c-d0d885185bfe)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"296ce5af9f3b6065a0c9a895a0c9c3732f513ecc58171f89f3e25e918c709e78\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:48:16.278290 systemd[1]: run-netns-cni\x2dfbcaf71a\x2d56c3\x2db085\x2decdf\x2de35bda3da8b3.mount: Deactivated successfully. Dec 16 12:48:19.603394 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3698901062.mount: Deactivated successfully. Dec 16 12:48:19.730645 containerd[1623]: time="2025-12-16T12:48:19.710619282Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.4: active requests=0, bytes read=156880025" Dec 16 12:48:19.733392 containerd[1623]: time="2025-12-16T12:48:19.732419348Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:48:19.804205 containerd[1623]: time="2025-12-16T12:48:19.802434145Z" level=info msg="ImageCreate event name:\"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:48:19.804205 containerd[1623]: time="2025-12-16T12:48:19.803879983Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Dec 16 12:48:19.805866 containerd[1623]: time="2025-12-16T12:48:19.805474401Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.4\" with image id \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\", size \"156883537\" in 4.476066627s" Dec 16 12:48:19.805866 containerd[1623]: time="2025-12-16T12:48:19.805514620Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\" returns image reference \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\"" Dec 16 12:48:19.887154 containerd[1623]: time="2025-12-16T12:48:19.886917342Z" level=info msg="CreateContainer within sandbox \"31a2c3fb2359d25620ab9b707e5cae784a39bde50f5c6db376d2523e06a8681d\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Dec 16 12:48:19.953108 containerd[1623]: time="2025-12-16T12:48:19.952959215Z" level=info msg="Container 6fb83315affcf4396650a73db90ec1c1d6fbb15dc1ef9c14efc213958aea1608: CDI devices from CRI Config.CDIDevices: []" Dec 16 12:48:19.954340 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1211364971.mount: Deactivated successfully. Dec 16 12:48:19.993245 containerd[1623]: time="2025-12-16T12:48:19.993198750Z" level=info msg="CreateContainer within sandbox \"31a2c3fb2359d25620ab9b707e5cae784a39bde50f5c6db376d2523e06a8681d\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"6fb83315affcf4396650a73db90ec1c1d6fbb15dc1ef9c14efc213958aea1608\"" Dec 16 12:48:19.998229 containerd[1623]: time="2025-12-16T12:48:19.998205760Z" level=info msg="StartContainer for \"6fb83315affcf4396650a73db90ec1c1d6fbb15dc1ef9c14efc213958aea1608\"" Dec 16 12:48:20.002872 containerd[1623]: time="2025-12-16T12:48:20.002831967Z" level=info msg="connecting to shim 6fb83315affcf4396650a73db90ec1c1d6fbb15dc1ef9c14efc213958aea1608" address="unix:///run/containerd/s/698b86754e3d1d4bc1220c168741fa4038a4fc0135498ef4de6454cf03124774" protocol=ttrpc version=3 Dec 16 12:48:20.150841 systemd[1]: Started cri-containerd-6fb83315affcf4396650a73db90ec1c1d6fbb15dc1ef9c14efc213958aea1608.scope - libcontainer container 6fb83315affcf4396650a73db90ec1c1d6fbb15dc1ef9c14efc213958aea1608. Dec 16 12:48:20.221495 kernel: kauditd_printk_skb: 6 callbacks suppressed Dec 16 12:48:20.221617 kernel: audit: type=1334 audit(1765889300.218:578): prog-id=172 op=LOAD Dec 16 12:48:20.218000 audit: BPF prog-id=172 op=LOAD Dec 16 12:48:20.231758 kernel: audit: type=1300 audit(1765889300.218:578): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3320 pid=3846 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:20.218000 audit[3846]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3320 pid=3846 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:20.218000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3666623833333135616666636634333936363530613733646239306563 Dec 16 12:48:20.248682 kernel: audit: type=1327 audit(1765889300.218:578): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3666623833333135616666636634333936363530613733646239306563 Dec 16 12:48:20.255955 kernel: audit: type=1334 audit(1765889300.218:579): prog-id=173 op=LOAD Dec 16 12:48:20.256001 kernel: audit: type=1300 audit(1765889300.218:579): arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3320 pid=3846 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:20.262824 kernel: audit: type=1327 audit(1765889300.218:579): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3666623833333135616666636634333936363530613733646239306563 Dec 16 12:48:20.218000 audit: BPF prog-id=173 op=LOAD Dec 16 12:48:20.218000 audit[3846]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3320 pid=3846 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:20.218000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3666623833333135616666636634333936363530613733646239306563 Dec 16 12:48:20.220000 audit: BPF prog-id=173 op=UNLOAD Dec 16 12:48:20.271740 kernel: audit: type=1334 audit(1765889300.220:580): prog-id=173 op=UNLOAD Dec 16 12:48:20.271789 kernel: audit: type=1300 audit(1765889300.220:580): arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3320 pid=3846 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:20.220000 audit[3846]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3320 pid=3846 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:20.276899 kernel: audit: type=1327 audit(1765889300.220:580): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3666623833333135616666636634333936363530613733646239306563 Dec 16 12:48:20.220000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3666623833333135616666636634333936363530613733646239306563 Dec 16 12:48:20.279074 kernel: audit: type=1334 audit(1765889300.220:581): prog-id=172 op=UNLOAD Dec 16 12:48:20.220000 audit: BPF prog-id=172 op=UNLOAD Dec 16 12:48:20.220000 audit[3846]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3320 pid=3846 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:20.220000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3666623833333135616666636634333936363530613733646239306563 Dec 16 12:48:20.220000 audit: BPF prog-id=174 op=LOAD Dec 16 12:48:20.220000 audit[3846]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=3320 pid=3846 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:20.220000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3666623833333135616666636634333936363530613733646239306563 Dec 16 12:48:20.323404 containerd[1623]: time="2025-12-16T12:48:20.323371857Z" level=info msg="StartContainer for \"6fb83315affcf4396650a73db90ec1c1d6fbb15dc1ef9c14efc213958aea1608\" returns successfully" Dec 16 12:48:20.639229 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Dec 16 12:48:20.639328 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Dec 16 12:48:20.892394 kubelet[2796]: I1216 12:48:20.879914 2796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-kddxn" podStartSLOduration=2.4536463 podStartE2EDuration="15.879894839s" podCreationTimestamp="2025-12-16 12:48:05 +0000 UTC" firstStartedPulling="2025-12-16 12:48:06.380935124 +0000 UTC m=+19.314663131" lastFinishedPulling="2025-12-16 12:48:19.807183653 +0000 UTC m=+32.740911670" observedRunningTime="2025-12-16 12:48:20.396078396 +0000 UTC m=+33.329806403" watchObservedRunningTime="2025-12-16 12:48:20.879894839 +0000 UTC m=+33.813622846" Dec 16 12:48:21.020647 kubelet[2796]: I1216 12:48:21.020597 2796 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cztt7\" (UniqueName: \"kubernetes.io/projected/9e36b6d0-206d-4ded-90fa-9c24767dcdf8-kube-api-access-cztt7\") pod \"9e36b6d0-206d-4ded-90fa-9c24767dcdf8\" (UID: \"9e36b6d0-206d-4ded-90fa-9c24767dcdf8\") " Dec 16 12:48:21.020647 kubelet[2796]: I1216 12:48:21.020658 2796 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9e36b6d0-206d-4ded-90fa-9c24767dcdf8-whisker-ca-bundle\") pod \"9e36b6d0-206d-4ded-90fa-9c24767dcdf8\" (UID: \"9e36b6d0-206d-4ded-90fa-9c24767dcdf8\") " Dec 16 12:48:21.020938 kubelet[2796]: I1216 12:48:21.020681 2796 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/9e36b6d0-206d-4ded-90fa-9c24767dcdf8-whisker-backend-key-pair\") pod \"9e36b6d0-206d-4ded-90fa-9c24767dcdf8\" (UID: \"9e36b6d0-206d-4ded-90fa-9c24767dcdf8\") " Dec 16 12:48:21.054250 kubelet[2796]: I1216 12:48:21.037497 2796 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e36b6d0-206d-4ded-90fa-9c24767dcdf8-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "9e36b6d0-206d-4ded-90fa-9c24767dcdf8" (UID: "9e36b6d0-206d-4ded-90fa-9c24767dcdf8"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 16 12:48:21.057169 systemd[1]: var-lib-kubelet-pods-9e36b6d0\x2d206d\x2d4ded\x2d90fa\x2d9c24767dcdf8-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dcztt7.mount: Deactivated successfully. Dec 16 12:48:21.061912 kubelet[2796]: I1216 12:48:21.061846 2796 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e36b6d0-206d-4ded-90fa-9c24767dcdf8-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "9e36b6d0-206d-4ded-90fa-9c24767dcdf8" (UID: "9e36b6d0-206d-4ded-90fa-9c24767dcdf8"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 16 12:48:21.063180 systemd[1]: var-lib-kubelet-pods-9e36b6d0\x2d206d\x2d4ded\x2d90fa\x2d9c24767dcdf8-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Dec 16 12:48:21.068700 kubelet[2796]: I1216 12:48:21.068645 2796 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e36b6d0-206d-4ded-90fa-9c24767dcdf8-kube-api-access-cztt7" (OuterVolumeSpecName: "kube-api-access-cztt7") pod "9e36b6d0-206d-4ded-90fa-9c24767dcdf8" (UID: "9e36b6d0-206d-4ded-90fa-9c24767dcdf8"). InnerVolumeSpecName "kube-api-access-cztt7". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 16 12:48:21.124713 kubelet[2796]: I1216 12:48:21.124665 2796 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9e36b6d0-206d-4ded-90fa-9c24767dcdf8-whisker-ca-bundle\") on node \"ci-4547-0-0-6-b8319fed62\" DevicePath \"\"" Dec 16 12:48:21.124713 kubelet[2796]: I1216 12:48:21.124709 2796 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/9e36b6d0-206d-4ded-90fa-9c24767dcdf8-whisker-backend-key-pair\") on node \"ci-4547-0-0-6-b8319fed62\" DevicePath \"\"" Dec 16 12:48:21.124713 kubelet[2796]: I1216 12:48:21.124719 2796 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-cztt7\" (UniqueName: \"kubernetes.io/projected/9e36b6d0-206d-4ded-90fa-9c24767dcdf8-kube-api-access-cztt7\") on node \"ci-4547-0-0-6-b8319fed62\" DevicePath \"\"" Dec 16 12:48:21.199273 systemd[1]: Removed slice kubepods-besteffort-pod9e36b6d0_206d_4ded_90fa_9c24767dcdf8.slice - libcontainer container kubepods-besteffort-pod9e36b6d0_206d_4ded_90fa_9c24767dcdf8.slice. Dec 16 12:48:21.522006 systemd[1]: Created slice kubepods-besteffort-poded1cdd8a_3d38_40e0_80a5_f1884fc6e6a1.slice - libcontainer container kubepods-besteffort-poded1cdd8a_3d38_40e0_80a5_f1884fc6e6a1.slice. Dec 16 12:48:21.628521 kubelet[2796]: I1216 12:48:21.628435 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79n4c\" (UniqueName: \"kubernetes.io/projected/ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1-kube-api-access-79n4c\") pod \"whisker-f48cdf5c5-pnl6x\" (UID: \"ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1\") " pod="calico-system/whisker-f48cdf5c5-pnl6x" Dec 16 12:48:21.628521 kubelet[2796]: I1216 12:48:21.628517 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1-whisker-ca-bundle\") pod \"whisker-f48cdf5c5-pnl6x\" (UID: \"ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1\") " pod="calico-system/whisker-f48cdf5c5-pnl6x" Dec 16 12:48:21.628895 kubelet[2796]: I1216 12:48:21.628551 2796 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1-whisker-backend-key-pair\") pod \"whisker-f48cdf5c5-pnl6x\" (UID: \"ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1\") " pod="calico-system/whisker-f48cdf5c5-pnl6x" Dec 16 12:48:21.825784 containerd[1623]: time="2025-12-16T12:48:21.825609530Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-f48cdf5c5-pnl6x,Uid:ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1,Namespace:calico-system,Attempt:0,}" Dec 16 12:48:22.252506 systemd-networkd[1527]: cali01bd9653f07: Link UP Dec 16 12:48:22.253200 systemd-networkd[1527]: cali01bd9653f07: Gained carrier Dec 16 12:48:22.293142 containerd[1623]: 2025-12-16 12:48:21.884 [INFO][3962] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Dec 16 12:48:22.293142 containerd[1623]: 2025-12-16 12:48:21.922 [INFO][3962] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4547--0--0--6--b8319fed62-k8s-whisker--f48cdf5c5--pnl6x-eth0 whisker-f48cdf5c5- calico-system ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1 860 0 2025-12-16 12:48:21 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:f48cdf5c5 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4547-0-0-6-b8319fed62 whisker-f48cdf5c5-pnl6x eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali01bd9653f07 [] [] }} ContainerID="92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" Namespace="calico-system" Pod="whisker-f48cdf5c5-pnl6x" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-whisker--f48cdf5c5--pnl6x-" Dec 16 12:48:22.293142 containerd[1623]: 2025-12-16 12:48:21.922 [INFO][3962] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" Namespace="calico-system" Pod="whisker-f48cdf5c5-pnl6x" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-whisker--f48cdf5c5--pnl6x-eth0" Dec 16 12:48:22.293142 containerd[1623]: 2025-12-16 12:48:22.157 [INFO][3973] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" HandleID="k8s-pod-network.92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" Workload="ci--4547--0--0--6--b8319fed62-k8s-whisker--f48cdf5c5--pnl6x-eth0" Dec 16 12:48:22.293394 containerd[1623]: 2025-12-16 12:48:22.161 [INFO][3973] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" HandleID="k8s-pod-network.92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" Workload="ci--4547--0--0--6--b8319fed62-k8s-whisker--f48cdf5c5--pnl6x-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003124c0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4547-0-0-6-b8319fed62", "pod":"whisker-f48cdf5c5-pnl6x", "timestamp":"2025-12-16 12:48:22.157073861 +0000 UTC"}, Hostname:"ci-4547-0-0-6-b8319fed62", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 16 12:48:22.293394 containerd[1623]: 2025-12-16 12:48:22.161 [INFO][3973] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 16 12:48:22.293394 containerd[1623]: 2025-12-16 12:48:22.162 [INFO][3973] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 16 12:48:22.293394 containerd[1623]: 2025-12-16 12:48:22.163 [INFO][3973] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4547-0-0-6-b8319fed62' Dec 16 12:48:22.293394 containerd[1623]: 2025-12-16 12:48:22.186 [INFO][3973] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:22.293394 containerd[1623]: 2025-12-16 12:48:22.211 [INFO][3973] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:22.293394 containerd[1623]: 2025-12-16 12:48:22.216 [INFO][3973] ipam/ipam.go 511: Trying affinity for 192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:22.293394 containerd[1623]: 2025-12-16 12:48:22.218 [INFO][3973] ipam/ipam.go 158: Attempting to load block cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:22.293394 containerd[1623]: 2025-12-16 12:48:22.219 [INFO][3973] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:22.294105 containerd[1623]: 2025-12-16 12:48:22.220 [INFO][3973] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.75.64/26 handle="k8s-pod-network.92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:22.294105 containerd[1623]: 2025-12-16 12:48:22.223 [INFO][3973] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78 Dec 16 12:48:22.294105 containerd[1623]: 2025-12-16 12:48:22.230 [INFO][3973] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.75.64/26 handle="k8s-pod-network.92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:22.294105 containerd[1623]: 2025-12-16 12:48:22.235 [INFO][3973] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.75.65/26] block=192.168.75.64/26 handle="k8s-pod-network.92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:22.294105 containerd[1623]: 2025-12-16 12:48:22.235 [INFO][3973] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.75.65/26] handle="k8s-pod-network.92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:22.294105 containerd[1623]: 2025-12-16 12:48:22.235 [INFO][3973] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 16 12:48:22.294105 containerd[1623]: 2025-12-16 12:48:22.235 [INFO][3973] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.75.65/26] IPv6=[] ContainerID="92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" HandleID="k8s-pod-network.92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" Workload="ci--4547--0--0--6--b8319fed62-k8s-whisker--f48cdf5c5--pnl6x-eth0" Dec 16 12:48:22.294221 containerd[1623]: 2025-12-16 12:48:22.238 [INFO][3962] cni-plugin/k8s.go 418: Populated endpoint ContainerID="92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" Namespace="calico-system" Pod="whisker-f48cdf5c5-pnl6x" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-whisker--f48cdf5c5--pnl6x-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-whisker--f48cdf5c5--pnl6x-eth0", GenerateName:"whisker-f48cdf5c5-", Namespace:"calico-system", SelfLink:"", UID:"ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1", ResourceVersion:"860", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 48, 21, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"f48cdf5c5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"", Pod:"whisker-f48cdf5c5-pnl6x", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.75.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali01bd9653f07", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:22.294221 containerd[1623]: 2025-12-16 12:48:22.238 [INFO][3962] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.75.65/32] ContainerID="92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" Namespace="calico-system" Pod="whisker-f48cdf5c5-pnl6x" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-whisker--f48cdf5c5--pnl6x-eth0" Dec 16 12:48:22.294290 containerd[1623]: 2025-12-16 12:48:22.238 [INFO][3962] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali01bd9653f07 ContainerID="92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" Namespace="calico-system" Pod="whisker-f48cdf5c5-pnl6x" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-whisker--f48cdf5c5--pnl6x-eth0" Dec 16 12:48:22.294290 containerd[1623]: 2025-12-16 12:48:22.251 [INFO][3962] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" Namespace="calico-system" Pod="whisker-f48cdf5c5-pnl6x" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-whisker--f48cdf5c5--pnl6x-eth0" Dec 16 12:48:22.294324 containerd[1623]: 2025-12-16 12:48:22.251 [INFO][3962] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" Namespace="calico-system" Pod="whisker-f48cdf5c5-pnl6x" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-whisker--f48cdf5c5--pnl6x-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-whisker--f48cdf5c5--pnl6x-eth0", GenerateName:"whisker-f48cdf5c5-", Namespace:"calico-system", SelfLink:"", UID:"ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1", ResourceVersion:"860", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 48, 21, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"f48cdf5c5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78", Pod:"whisker-f48cdf5c5-pnl6x", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.75.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali01bd9653f07", MAC:"72:ff:df:eb:0b:87", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:22.294370 containerd[1623]: 2025-12-16 12:48:22.288 [INFO][3962] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" Namespace="calico-system" Pod="whisker-f48cdf5c5-pnl6x" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-whisker--f48cdf5c5--pnl6x-eth0" Dec 16 12:48:22.496343 containerd[1623]: time="2025-12-16T12:48:22.496299612Z" level=info msg="connecting to shim 92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78" address="unix:///run/containerd/s/c736c60568890e6278384a3afabc642ae64306f2e9bbe951c18f0b7355318488" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:48:22.537690 systemd[1]: Started cri-containerd-92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78.scope - libcontainer container 92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78. Dec 16 12:48:22.556000 audit: BPF prog-id=175 op=LOAD Dec 16 12:48:22.557000 audit: BPF prog-id=176 op=LOAD Dec 16 12:48:22.557000 audit[4115]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=4100 pid=4115 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.557000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3932653231393632626436646338363363393933626239306330653834 Dec 16 12:48:22.557000 audit: BPF prog-id=176 op=UNLOAD Dec 16 12:48:22.557000 audit[4115]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4100 pid=4115 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.557000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3932653231393632626436646338363363393933626239306330653834 Dec 16 12:48:22.557000 audit: BPF prog-id=177 op=LOAD Dec 16 12:48:22.557000 audit[4115]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=4100 pid=4115 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.557000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3932653231393632626436646338363363393933626239306330653834 Dec 16 12:48:22.557000 audit: BPF prog-id=178 op=LOAD Dec 16 12:48:22.557000 audit[4115]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=4100 pid=4115 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.557000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3932653231393632626436646338363363393933626239306330653834 Dec 16 12:48:22.558000 audit: BPF prog-id=178 op=UNLOAD Dec 16 12:48:22.558000 audit[4115]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=4100 pid=4115 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.558000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3932653231393632626436646338363363393933626239306330653834 Dec 16 12:48:22.558000 audit: BPF prog-id=177 op=UNLOAD Dec 16 12:48:22.558000 audit[4115]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4100 pid=4115 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.558000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3932653231393632626436646338363363393933626239306330653834 Dec 16 12:48:22.558000 audit: BPF prog-id=179 op=LOAD Dec 16 12:48:22.558000 audit[4115]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=4100 pid=4115 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.558000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3932653231393632626436646338363363393933626239306330653834 Dec 16 12:48:22.644103 containerd[1623]: time="2025-12-16T12:48:22.644053042Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-f48cdf5c5-pnl6x,Uid:ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1,Namespace:calico-system,Attempt:0,} returns sandbox id \"92e21962bd6dc863c993bb90c0e84247341e5fe56c43115890e6ae8e28582b78\"" Dec 16 12:48:22.656662 containerd[1623]: time="2025-12-16T12:48:22.656389692Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Dec 16 12:48:22.781000 audit: BPF prog-id=180 op=LOAD Dec 16 12:48:22.781000 audit[4176]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffce68b0a40 a2=98 a3=1fffffffffffffff items=0 ppid=4004 pid=4176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.781000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Dec 16 12:48:22.781000 audit: BPF prog-id=180 op=UNLOAD Dec 16 12:48:22.781000 audit[4176]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffce68b0a10 a3=0 items=0 ppid=4004 pid=4176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.781000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Dec 16 12:48:22.782000 audit: BPF prog-id=181 op=LOAD Dec 16 12:48:22.782000 audit[4176]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffce68b0920 a2=94 a3=3 items=0 ppid=4004 pid=4176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.782000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Dec 16 12:48:22.782000 audit: BPF prog-id=181 op=UNLOAD Dec 16 12:48:22.782000 audit[4176]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffce68b0920 a2=94 a3=3 items=0 ppid=4004 pid=4176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.782000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Dec 16 12:48:22.782000 audit: BPF prog-id=182 op=LOAD Dec 16 12:48:22.782000 audit[4176]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffce68b0960 a2=94 a3=7ffce68b0b40 items=0 ppid=4004 pid=4176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.782000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Dec 16 12:48:22.782000 audit: BPF prog-id=182 op=UNLOAD Dec 16 12:48:22.782000 audit[4176]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffce68b0960 a2=94 a3=7ffce68b0b40 items=0 ppid=4004 pid=4176 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.782000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Dec 16 12:48:22.785000 audit: BPF prog-id=183 op=LOAD Dec 16 12:48:22.785000 audit[4177]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffea65a9280 a2=98 a3=3 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.785000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.786000 audit: BPF prog-id=183 op=UNLOAD Dec 16 12:48:22.786000 audit[4177]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffea65a9250 a3=0 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.786000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.787000 audit: BPF prog-id=184 op=LOAD Dec 16 12:48:22.787000 audit[4177]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffea65a9070 a2=94 a3=54428f items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.787000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.787000 audit: BPF prog-id=184 op=UNLOAD Dec 16 12:48:22.787000 audit[4177]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffea65a9070 a2=94 a3=54428f items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.787000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.787000 audit: BPF prog-id=185 op=LOAD Dec 16 12:48:22.787000 audit[4177]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffea65a90a0 a2=94 a3=2 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.787000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.787000 audit: BPF prog-id=185 op=UNLOAD Dec 16 12:48:22.787000 audit[4177]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffea65a90a0 a2=0 a3=2 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.787000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.944000 audit: BPF prog-id=186 op=LOAD Dec 16 12:48:22.944000 audit[4177]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffea65a8f60 a2=94 a3=1 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.944000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.944000 audit: BPF prog-id=186 op=UNLOAD Dec 16 12:48:22.944000 audit[4177]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffea65a8f60 a2=94 a3=1 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.944000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.952000 audit: BPF prog-id=187 op=LOAD Dec 16 12:48:22.952000 audit[4177]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffea65a8f50 a2=94 a3=4 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.952000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.952000 audit: BPF prog-id=187 op=UNLOAD Dec 16 12:48:22.952000 audit[4177]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffea65a8f50 a2=0 a3=4 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.952000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.953000 audit: BPF prog-id=188 op=LOAD Dec 16 12:48:22.953000 audit[4177]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffea65a8db0 a2=94 a3=5 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.953000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.953000 audit: BPF prog-id=188 op=UNLOAD Dec 16 12:48:22.953000 audit[4177]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffea65a8db0 a2=0 a3=5 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.953000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.953000 audit: BPF prog-id=189 op=LOAD Dec 16 12:48:22.953000 audit[4177]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffea65a8fd0 a2=94 a3=6 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.953000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.953000 audit: BPF prog-id=189 op=UNLOAD Dec 16 12:48:22.953000 audit[4177]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffea65a8fd0 a2=0 a3=6 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.953000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.953000 audit: BPF prog-id=190 op=LOAD Dec 16 12:48:22.953000 audit[4177]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffea65a8780 a2=94 a3=88 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.953000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.953000 audit: BPF prog-id=191 op=LOAD Dec 16 12:48:22.953000 audit[4177]: SYSCALL arch=c000003e syscall=321 success=yes exit=7 a0=5 a1=7ffea65a8600 a2=94 a3=2 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.953000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.953000 audit: BPF prog-id=191 op=UNLOAD Dec 16 12:48:22.953000 audit[4177]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=7 a1=7ffea65a8630 a2=0 a3=7ffea65a8730 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.953000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.954000 audit: BPF prog-id=190 op=UNLOAD Dec 16 12:48:22.954000 audit[4177]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=3928cd10 a2=0 a3=365dcf92263b1e86 items=0 ppid=4004 pid=4177 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.954000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Dec 16 12:48:22.971000 audit: BPF prog-id=192 op=LOAD Dec 16 12:48:22.971000 audit[4180]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd847b1280 a2=98 a3=1999999999999999 items=0 ppid=4004 pid=4180 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.971000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Dec 16 12:48:22.971000 audit: BPF prog-id=192 op=UNLOAD Dec 16 12:48:22.971000 audit[4180]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffd847b1250 a3=0 items=0 ppid=4004 pid=4180 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.971000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Dec 16 12:48:22.971000 audit: BPF prog-id=193 op=LOAD Dec 16 12:48:22.971000 audit[4180]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd847b1160 a2=94 a3=ffff items=0 ppid=4004 pid=4180 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.971000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Dec 16 12:48:22.971000 audit: BPF prog-id=193 op=UNLOAD Dec 16 12:48:22.971000 audit[4180]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffd847b1160 a2=94 a3=ffff items=0 ppid=4004 pid=4180 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.971000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Dec 16 12:48:22.971000 audit: BPF prog-id=194 op=LOAD Dec 16 12:48:22.971000 audit[4180]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd847b11a0 a2=94 a3=7ffd847b1380 items=0 ppid=4004 pid=4180 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.971000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Dec 16 12:48:22.972000 audit: BPF prog-id=194 op=UNLOAD Dec 16 12:48:22.972000 audit[4180]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffd847b11a0 a2=94 a3=7ffd847b1380 items=0 ppid=4004 pid=4180 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:22.972000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Dec 16 12:48:23.037753 systemd-networkd[1527]: vxlan.calico: Link UP Dec 16 12:48:23.037761 systemd-networkd[1527]: vxlan.calico: Gained carrier Dec 16 12:48:23.056000 audit: BPF prog-id=195 op=LOAD Dec 16 12:48:23.056000 audit[4205]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7fffe5cd8880 a2=98 a3=0 items=0 ppid=4004 pid=4205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.056000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Dec 16 12:48:23.056000 audit: BPF prog-id=195 op=UNLOAD Dec 16 12:48:23.056000 audit[4205]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7fffe5cd8850 a3=0 items=0 ppid=4004 pid=4205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.056000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Dec 16 12:48:23.060000 audit: BPF prog-id=196 op=LOAD Dec 16 12:48:23.060000 audit[4205]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7fffe5cd8690 a2=94 a3=54428f items=0 ppid=4004 pid=4205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.060000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Dec 16 12:48:23.060000 audit: BPF prog-id=196 op=UNLOAD Dec 16 12:48:23.060000 audit[4205]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7fffe5cd8690 a2=94 a3=54428f items=0 ppid=4004 pid=4205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.060000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Dec 16 12:48:23.060000 audit: BPF prog-id=197 op=LOAD Dec 16 12:48:23.060000 audit[4205]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7fffe5cd86c0 a2=94 a3=2 items=0 ppid=4004 pid=4205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.060000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Dec 16 12:48:23.060000 audit: BPF prog-id=197 op=UNLOAD Dec 16 12:48:23.060000 audit[4205]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7fffe5cd86c0 a2=0 a3=2 items=0 ppid=4004 pid=4205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.060000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Dec 16 12:48:23.060000 audit: BPF prog-id=198 op=LOAD Dec 16 12:48:23.060000 audit[4205]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fffe5cd8470 a2=94 a3=4 items=0 ppid=4004 pid=4205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.060000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Dec 16 12:48:23.060000 audit: BPF prog-id=198 op=UNLOAD Dec 16 12:48:23.060000 audit[4205]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7fffe5cd8470 a2=94 a3=4 items=0 ppid=4004 pid=4205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.060000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Dec 16 12:48:23.060000 audit: BPF prog-id=199 op=LOAD Dec 16 12:48:23.060000 audit[4205]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fffe5cd8570 a2=94 a3=7fffe5cd86f0 items=0 ppid=4004 pid=4205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.060000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Dec 16 12:48:23.060000 audit: BPF prog-id=199 op=UNLOAD Dec 16 12:48:23.060000 audit[4205]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7fffe5cd8570 a2=0 a3=7fffe5cd86f0 items=0 ppid=4004 pid=4205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.060000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Dec 16 12:48:23.061000 audit: BPF prog-id=200 op=LOAD Dec 16 12:48:23.061000 audit[4205]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fffe5cd7ca0 a2=94 a3=2 items=0 ppid=4004 pid=4205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.061000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Dec 16 12:48:23.061000 audit: BPF prog-id=200 op=UNLOAD Dec 16 12:48:23.061000 audit[4205]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7fffe5cd7ca0 a2=0 a3=2 items=0 ppid=4004 pid=4205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.061000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Dec 16 12:48:23.061000 audit: BPF prog-id=201 op=LOAD Dec 16 12:48:23.061000 audit[4205]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fffe5cd7da0 a2=94 a3=30 items=0 ppid=4004 pid=4205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.061000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Dec 16 12:48:23.070000 audit: BPF prog-id=202 op=LOAD Dec 16 12:48:23.070000 audit[4209]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffdb0b779d0 a2=98 a3=0 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.070000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.071000 audit: BPF prog-id=202 op=UNLOAD Dec 16 12:48:23.071000 audit[4209]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffdb0b779a0 a3=0 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.071000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.071000 audit: BPF prog-id=203 op=LOAD Dec 16 12:48:23.071000 audit[4209]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffdb0b777c0 a2=94 a3=54428f items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.071000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.071000 audit: BPF prog-id=203 op=UNLOAD Dec 16 12:48:23.071000 audit[4209]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffdb0b777c0 a2=94 a3=54428f items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.071000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.071000 audit: BPF prog-id=204 op=LOAD Dec 16 12:48:23.071000 audit[4209]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffdb0b777f0 a2=94 a3=2 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.071000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.071000 audit: BPF prog-id=204 op=UNLOAD Dec 16 12:48:23.071000 audit[4209]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffdb0b777f0 a2=0 a3=2 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.071000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.133349 containerd[1623]: time="2025-12-16T12:48:23.133264693Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:23.135674 containerd[1623]: time="2025-12-16T12:48:23.135510828Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Dec 16 12:48:23.135941 containerd[1623]: time="2025-12-16T12:48:23.135861287Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:23.159899 kubelet[2796]: I1216 12:48:23.159834 2796 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e36b6d0-206d-4ded-90fa-9c24767dcdf8" path="/var/lib/kubelet/pods/9e36b6d0-206d-4ded-90fa-9c24767dcdf8/volumes" Dec 16 12:48:23.164195 kubelet[2796]: E1216 12:48:23.164072 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 16 12:48:23.168069 kubelet[2796]: E1216 12:48:23.167837 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 16 12:48:23.199805 kubelet[2796]: E1216 12:48:23.199711 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:f9261a3a5fe8412497181a324a67db5c,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-79n4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-f48cdf5c5-pnl6x_calico-system(ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:23.202621 containerd[1623]: time="2025-12-16T12:48:23.202485198Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Dec 16 12:48:23.225000 audit: BPF prog-id=205 op=LOAD Dec 16 12:48:23.225000 audit[4209]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffdb0b776b0 a2=94 a3=1 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.225000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.225000 audit: BPF prog-id=205 op=UNLOAD Dec 16 12:48:23.225000 audit[4209]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffdb0b776b0 a2=94 a3=1 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.225000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.236000 audit: BPF prog-id=206 op=LOAD Dec 16 12:48:23.236000 audit[4209]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffdb0b776a0 a2=94 a3=4 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.236000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.236000 audit: BPF prog-id=206 op=UNLOAD Dec 16 12:48:23.236000 audit[4209]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffdb0b776a0 a2=0 a3=4 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.236000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.236000 audit: BPF prog-id=207 op=LOAD Dec 16 12:48:23.236000 audit[4209]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffdb0b77500 a2=94 a3=5 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.236000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.236000 audit: BPF prog-id=207 op=UNLOAD Dec 16 12:48:23.236000 audit[4209]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffdb0b77500 a2=0 a3=5 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.236000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.236000 audit: BPF prog-id=208 op=LOAD Dec 16 12:48:23.236000 audit[4209]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffdb0b77720 a2=94 a3=6 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.236000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.236000 audit: BPF prog-id=208 op=UNLOAD Dec 16 12:48:23.236000 audit[4209]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffdb0b77720 a2=0 a3=6 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.236000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.237000 audit: BPF prog-id=209 op=LOAD Dec 16 12:48:23.237000 audit[4209]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffdb0b76ed0 a2=94 a3=88 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.237000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.237000 audit: BPF prog-id=210 op=LOAD Dec 16 12:48:23.237000 audit[4209]: SYSCALL arch=c000003e syscall=321 success=yes exit=7 a0=5 a1=7ffdb0b76d50 a2=94 a3=2 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.237000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.237000 audit: BPF prog-id=210 op=UNLOAD Dec 16 12:48:23.237000 audit[4209]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=7 a1=7ffdb0b76d80 a2=0 a3=7ffdb0b76e80 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.237000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.237000 audit: BPF prog-id=209 op=UNLOAD Dec 16 12:48:23.237000 audit[4209]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=b904d10 a2=0 a3=9aac6554aa4ea6a4 items=0 ppid=4004 pid=4209 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.237000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Dec 16 12:48:23.252000 audit: BPF prog-id=201 op=UNLOAD Dec 16 12:48:23.252000 audit[4004]: SYSCALL arch=c000003e syscall=263 success=yes exit=0 a0=ffffffffffffff9c a1=c000dbc040 a2=0 a3=0 items=0 ppid=3981 pid=4004 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="calico-node" exe="/usr/bin/calico-node" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.252000 audit: PROCTITLE proctitle=63616C69636F2D6E6F6465002D66656C6978 Dec 16 12:48:23.308000 audit[4233]: NETFILTER_CFG table=nat:119 family=2 entries=15 op=nft_register_chain pid=4233 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Dec 16 12:48:23.308000 audit[4233]: SYSCALL arch=c000003e syscall=46 success=yes exit=5084 a0=3 a1=7ffd502d8f90 a2=0 a3=7ffd502d8f7c items=0 ppid=4004 pid=4233 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.308000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Dec 16 12:48:23.322000 audit[4236]: NETFILTER_CFG table=mangle:120 family=2 entries=16 op=nft_register_chain pid=4236 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Dec 16 12:48:23.322000 audit[4236]: SYSCALL arch=c000003e syscall=46 success=yes exit=6868 a0=3 a1=7ffe364c6120 a2=0 a3=7ffe364c610c items=0 ppid=4004 pid=4236 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.322000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Dec 16 12:48:23.334000 audit[4234]: NETFILTER_CFG table=raw:121 family=2 entries=21 op=nft_register_chain pid=4234 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Dec 16 12:48:23.334000 audit[4234]: SYSCALL arch=c000003e syscall=46 success=yes exit=8452 a0=3 a1=7fff3a808680 a2=0 a3=7fff3a80866c items=0 ppid=4004 pid=4234 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.334000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Dec 16 12:48:23.337000 audit[4235]: NETFILTER_CFG table=filter:122 family=2 entries=94 op=nft_register_chain pid=4235 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Dec 16 12:48:23.337000 audit[4235]: SYSCALL arch=c000003e syscall=46 success=yes exit=53116 a0=3 a1=7fff3479a0f0 a2=0 a3=5586ace6d000 items=0 ppid=4004 pid=4235 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:23.337000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Dec 16 12:48:23.535092 systemd-networkd[1527]: cali01bd9653f07: Gained IPv6LL Dec 16 12:48:23.660440 containerd[1623]: time="2025-12-16T12:48:23.660266773Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:23.661934 containerd[1623]: time="2025-12-16T12:48:23.661817411Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Dec 16 12:48:23.662191 containerd[1623]: time="2025-12-16T12:48:23.661949607Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:23.662289 kubelet[2796]: E1216 12:48:23.662137 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 16 12:48:23.662289 kubelet[2796]: E1216 12:48:23.662190 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 16 12:48:23.662570 kubelet[2796]: E1216 12:48:23.662342 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-79n4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-f48cdf5c5-pnl6x_calico-system(ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:23.673093 kubelet[2796]: E1216 12:48:23.672791 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-f48cdf5c5-pnl6x" podUID="ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1" Dec 16 12:48:24.369116 kubelet[2796]: E1216 12:48:24.368948 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-f48cdf5c5-pnl6x" podUID="ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1" Dec 16 12:48:24.429307 systemd-networkd[1527]: vxlan.calico: Gained IPv6LL Dec 16 12:48:24.475000 audit[4249]: NETFILTER_CFG table=filter:123 family=2 entries=20 op=nft_register_rule pid=4249 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:24.475000 audit[4249]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffef93bbf30 a2=0 a3=7ffef93bbf1c items=0 ppid=2898 pid=4249 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:24.475000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:24.479000 audit[4249]: NETFILTER_CFG table=nat:124 family=2 entries=14 op=nft_register_rule pid=4249 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:24.479000 audit[4249]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffef93bbf30 a2=0 a3=0 items=0 ppid=2898 pid=4249 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:24.479000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:26.155447 containerd[1623]: time="2025-12-16T12:48:26.154579963Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-48cft,Uid:b556e31a-1d4e-41c7-bc84-fbb790df238f,Namespace:kube-system,Attempt:0,}" Dec 16 12:48:26.348864 systemd-networkd[1527]: calidf1eb0e5316: Link UP Dec 16 12:48:26.349674 systemd-networkd[1527]: calidf1eb0e5316: Gained carrier Dec 16 12:48:26.376845 containerd[1623]: 2025-12-16 12:48:26.233 [INFO][4255] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--48cft-eth0 coredns-668d6bf9bc- kube-system b556e31a-1d4e-41c7-bc84-fbb790df238f 786 0 2025-12-16 12:47:54 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4547-0-0-6-b8319fed62 coredns-668d6bf9bc-48cft eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calidf1eb0e5316 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" Namespace="kube-system" Pod="coredns-668d6bf9bc-48cft" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--48cft-" Dec 16 12:48:26.376845 containerd[1623]: 2025-12-16 12:48:26.233 [INFO][4255] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" Namespace="kube-system" Pod="coredns-668d6bf9bc-48cft" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--48cft-eth0" Dec 16 12:48:26.376845 containerd[1623]: 2025-12-16 12:48:26.276 [INFO][4265] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" HandleID="k8s-pod-network.5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" Workload="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--48cft-eth0" Dec 16 12:48:26.377158 containerd[1623]: 2025-12-16 12:48:26.276 [INFO][4265] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" HandleID="k8s-pod-network.5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" Workload="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--48cft-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d56a0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4547-0-0-6-b8319fed62", "pod":"coredns-668d6bf9bc-48cft", "timestamp":"2025-12-16 12:48:26.276487483 +0000 UTC"}, Hostname:"ci-4547-0-0-6-b8319fed62", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 16 12:48:26.377158 containerd[1623]: 2025-12-16 12:48:26.276 [INFO][4265] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 16 12:48:26.377158 containerd[1623]: 2025-12-16 12:48:26.276 [INFO][4265] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 16 12:48:26.377158 containerd[1623]: 2025-12-16 12:48:26.276 [INFO][4265] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4547-0-0-6-b8319fed62' Dec 16 12:48:26.377158 containerd[1623]: 2025-12-16 12:48:26.294 [INFO][4265] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:26.377158 containerd[1623]: 2025-12-16 12:48:26.302 [INFO][4265] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:26.377158 containerd[1623]: 2025-12-16 12:48:26.309 [INFO][4265] ipam/ipam.go 511: Trying affinity for 192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:26.377158 containerd[1623]: 2025-12-16 12:48:26.312 [INFO][4265] ipam/ipam.go 158: Attempting to load block cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:26.377158 containerd[1623]: 2025-12-16 12:48:26.315 [INFO][4265] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:26.377423 containerd[1623]: 2025-12-16 12:48:26.315 [INFO][4265] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.75.64/26 handle="k8s-pod-network.5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:26.377423 containerd[1623]: 2025-12-16 12:48:26.317 [INFO][4265] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99 Dec 16 12:48:26.377423 containerd[1623]: 2025-12-16 12:48:26.323 [INFO][4265] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.75.64/26 handle="k8s-pod-network.5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:26.377423 containerd[1623]: 2025-12-16 12:48:26.333 [INFO][4265] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.75.66/26] block=192.168.75.64/26 handle="k8s-pod-network.5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:26.377423 containerd[1623]: 2025-12-16 12:48:26.333 [INFO][4265] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.75.66/26] handle="k8s-pod-network.5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:26.377423 containerd[1623]: 2025-12-16 12:48:26.333 [INFO][4265] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 16 12:48:26.377423 containerd[1623]: 2025-12-16 12:48:26.333 [INFO][4265] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.75.66/26] IPv6=[] ContainerID="5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" HandleID="k8s-pod-network.5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" Workload="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--48cft-eth0" Dec 16 12:48:26.378619 containerd[1623]: 2025-12-16 12:48:26.338 [INFO][4255] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" Namespace="kube-system" Pod="coredns-668d6bf9bc-48cft" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--48cft-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--48cft-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"b556e31a-1d4e-41c7-bc84-fbb790df238f", ResourceVersion:"786", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 47, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"", Pod:"coredns-668d6bf9bc-48cft", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.75.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calidf1eb0e5316", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:26.378619 containerd[1623]: 2025-12-16 12:48:26.339 [INFO][4255] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.75.66/32] ContainerID="5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" Namespace="kube-system" Pod="coredns-668d6bf9bc-48cft" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--48cft-eth0" Dec 16 12:48:26.378619 containerd[1623]: 2025-12-16 12:48:26.339 [INFO][4255] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calidf1eb0e5316 ContainerID="5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" Namespace="kube-system" Pod="coredns-668d6bf9bc-48cft" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--48cft-eth0" Dec 16 12:48:26.378619 containerd[1623]: 2025-12-16 12:48:26.350 [INFO][4255] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" Namespace="kube-system" Pod="coredns-668d6bf9bc-48cft" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--48cft-eth0" Dec 16 12:48:26.378619 containerd[1623]: 2025-12-16 12:48:26.351 [INFO][4255] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" Namespace="kube-system" Pod="coredns-668d6bf9bc-48cft" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--48cft-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--48cft-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"b556e31a-1d4e-41c7-bc84-fbb790df238f", ResourceVersion:"786", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 47, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99", Pod:"coredns-668d6bf9bc-48cft", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.75.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calidf1eb0e5316", MAC:"6e:08:ca:9d:6a:f7", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:26.378619 containerd[1623]: 2025-12-16 12:48:26.364 [INFO][4255] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" Namespace="kube-system" Pod="coredns-668d6bf9bc-48cft" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--48cft-eth0" Dec 16 12:48:26.415073 kernel: kauditd_printk_skb: 231 callbacks suppressed Dec 16 12:48:26.415195 kernel: audit: type=1325 audit(1765889306.410:659): table=filter:125 family=2 entries=42 op=nft_register_chain pid=4284 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Dec 16 12:48:26.410000 audit[4284]: NETFILTER_CFG table=filter:125 family=2 entries=42 op=nft_register_chain pid=4284 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Dec 16 12:48:26.422434 containerd[1623]: time="2025-12-16T12:48:26.422242758Z" level=info msg="connecting to shim 5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99" address="unix:///run/containerd/s/d5efd51f28995fe078510d880774003aeac85272486e579429277cd794c1ac91" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:48:26.410000 audit[4284]: SYSCALL arch=c000003e syscall=46 success=yes exit=22552 a0=3 a1=7ffe096b3d90 a2=0 a3=7ffe096b3d7c items=0 ppid=4004 pid=4284 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.435053 kernel: audit: type=1300 audit(1765889306.410:659): arch=c000003e syscall=46 success=yes exit=22552 a0=3 a1=7ffe096b3d90 a2=0 a3=7ffe096b3d7c items=0 ppid=4004 pid=4284 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.435121 kernel: audit: type=1327 audit(1765889306.410:659): proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Dec 16 12:48:26.410000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Dec 16 12:48:26.463163 systemd[1]: Started cri-containerd-5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99.scope - libcontainer container 5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99. Dec 16 12:48:26.470000 audit: BPF prog-id=211 op=LOAD Dec 16 12:48:26.475875 kernel: audit: type=1334 audit(1765889306.470:660): prog-id=211 op=LOAD Dec 16 12:48:26.475957 kernel: audit: type=1334 audit(1765889306.470:661): prog-id=212 op=LOAD Dec 16 12:48:26.470000 audit: BPF prog-id=212 op=LOAD Dec 16 12:48:26.470000 audit[4303]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=4291 pid=4303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.483063 kernel: audit: type=1300 audit(1765889306.470:661): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=4291 pid=4303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.483129 kernel: audit: type=1327 audit(1765889306.470:661): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562316232663262396162376363633232306462636365666138613836 Dec 16 12:48:26.470000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562316232663262396162376363633232306462636365666138613836 Dec 16 12:48:26.489369 kernel: audit: type=1334 audit(1765889306.470:662): prog-id=212 op=UNLOAD Dec 16 12:48:26.470000 audit: BPF prog-id=212 op=UNLOAD Dec 16 12:48:26.495086 kernel: audit: type=1300 audit(1765889306.470:662): arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4291 pid=4303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.470000 audit[4303]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4291 pid=4303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.470000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562316232663262396162376363633232306462636365666138613836 Dec 16 12:48:26.501079 kernel: audit: type=1327 audit(1765889306.470:662): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562316232663262396162376363633232306462636365666138613836 Dec 16 12:48:26.471000 audit: BPF prog-id=213 op=LOAD Dec 16 12:48:26.471000 audit[4303]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=4291 pid=4303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.471000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562316232663262396162376363633232306462636365666138613836 Dec 16 12:48:26.471000 audit: BPF prog-id=214 op=LOAD Dec 16 12:48:26.471000 audit[4303]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=4291 pid=4303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.471000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562316232663262396162376363633232306462636365666138613836 Dec 16 12:48:26.471000 audit: BPF prog-id=214 op=UNLOAD Dec 16 12:48:26.471000 audit[4303]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=4291 pid=4303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.471000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562316232663262396162376363633232306462636365666138613836 Dec 16 12:48:26.471000 audit: BPF prog-id=213 op=UNLOAD Dec 16 12:48:26.471000 audit[4303]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4291 pid=4303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.471000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562316232663262396162376363633232306462636365666138613836 Dec 16 12:48:26.471000 audit: BPF prog-id=215 op=LOAD Dec 16 12:48:26.471000 audit[4303]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=4291 pid=4303 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.471000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3562316232663262396162376363633232306462636365666138613836 Dec 16 12:48:26.519536 containerd[1623]: time="2025-12-16T12:48:26.519496089Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-48cft,Uid:b556e31a-1d4e-41c7-bc84-fbb790df238f,Namespace:kube-system,Attempt:0,} returns sandbox id \"5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99\"" Dec 16 12:48:26.524008 containerd[1623]: time="2025-12-16T12:48:26.523973918Z" level=info msg="CreateContainer within sandbox \"5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Dec 16 12:48:26.538343 containerd[1623]: time="2025-12-16T12:48:26.538313839Z" level=info msg="Container d48e6765479001dce34d2759c24f2e1ea8c98770d793ee4648b5f2a827143d99: CDI devices from CRI Config.CDIDevices: []" Dec 16 12:48:26.539385 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3748008630.mount: Deactivated successfully. Dec 16 12:48:26.543754 containerd[1623]: time="2025-12-16T12:48:26.543720891Z" level=info msg="CreateContainer within sandbox \"5b1b2f2b9ab7ccc220dbccefa8a864ce3e46cf9acb50cd40e4ef3ceef8d6bb99\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"d48e6765479001dce34d2759c24f2e1ea8c98770d793ee4648b5f2a827143d99\"" Dec 16 12:48:26.545295 containerd[1623]: time="2025-12-16T12:48:26.545265227Z" level=info msg="StartContainer for \"d48e6765479001dce34d2759c24f2e1ea8c98770d793ee4648b5f2a827143d99\"" Dec 16 12:48:26.546116 containerd[1623]: time="2025-12-16T12:48:26.546053820Z" level=info msg="connecting to shim d48e6765479001dce34d2759c24f2e1ea8c98770d793ee4648b5f2a827143d99" address="unix:///run/containerd/s/d5efd51f28995fe078510d880774003aeac85272486e579429277cd794c1ac91" protocol=ttrpc version=3 Dec 16 12:48:26.564165 systemd[1]: Started cri-containerd-d48e6765479001dce34d2759c24f2e1ea8c98770d793ee4648b5f2a827143d99.scope - libcontainer container d48e6765479001dce34d2759c24f2e1ea8c98770d793ee4648b5f2a827143d99. Dec 16 12:48:26.574000 audit: BPF prog-id=216 op=LOAD Dec 16 12:48:26.575000 audit: BPF prog-id=217 op=LOAD Dec 16 12:48:26.575000 audit[4328]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000186238 a2=98 a3=0 items=0 ppid=4291 pid=4328 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.575000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6434386536373635343739303031646365333464323735396332346632 Dec 16 12:48:26.575000 audit: BPF prog-id=217 op=UNLOAD Dec 16 12:48:26.575000 audit[4328]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4291 pid=4328 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.575000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6434386536373635343739303031646365333464323735396332346632 Dec 16 12:48:26.575000 audit: BPF prog-id=218 op=LOAD Dec 16 12:48:26.575000 audit[4328]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000186488 a2=98 a3=0 items=0 ppid=4291 pid=4328 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.575000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6434386536373635343739303031646365333464323735396332346632 Dec 16 12:48:26.575000 audit: BPF prog-id=219 op=LOAD Dec 16 12:48:26.575000 audit[4328]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000186218 a2=98 a3=0 items=0 ppid=4291 pid=4328 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.575000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6434386536373635343739303031646365333464323735396332346632 Dec 16 12:48:26.575000 audit: BPF prog-id=219 op=UNLOAD Dec 16 12:48:26.575000 audit[4328]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4291 pid=4328 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.575000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6434386536373635343739303031646365333464323735396332346632 Dec 16 12:48:26.575000 audit: BPF prog-id=218 op=UNLOAD Dec 16 12:48:26.575000 audit[4328]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4291 pid=4328 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.575000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6434386536373635343739303031646365333464323735396332346632 Dec 16 12:48:26.575000 audit: BPF prog-id=220 op=LOAD Dec 16 12:48:26.575000 audit[4328]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001866e8 a2=98 a3=0 items=0 ppid=4291 pid=4328 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:26.575000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6434386536373635343739303031646365333464323735396332346632 Dec 16 12:48:26.591923 containerd[1623]: time="2025-12-16T12:48:26.591887931Z" level=info msg="StartContainer for \"d48e6765479001dce34d2759c24f2e1ea8c98770d793ee4648b5f2a827143d99\" returns successfully" Dec 16 12:48:27.179033 containerd[1623]: time="2025-12-16T12:48:27.178962711Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-m8p7d,Uid:2053803c-11b9-48f6-949c-d0d885185bfe,Namespace:calico-system,Attempt:0,}" Dec 16 12:48:27.359256 systemd-networkd[1527]: cali75320bddb25: Link UP Dec 16 12:48:27.362143 systemd-networkd[1527]: cali75320bddb25: Gained carrier Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.244 [INFO][4360] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4547--0--0--6--b8319fed62-k8s-csi--node--driver--m8p7d-eth0 csi-node-driver- calico-system 2053803c-11b9-48f6-949c-d0d885185bfe 684 0 2025-12-16 12:48:06 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:857b56db8f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4547-0-0-6-b8319fed62 csi-node-driver-m8p7d eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali75320bddb25 [] [] }} ContainerID="b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" Namespace="calico-system" Pod="csi-node-driver-m8p7d" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-csi--node--driver--m8p7d-" Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.245 [INFO][4360] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" Namespace="calico-system" Pod="csi-node-driver-m8p7d" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-csi--node--driver--m8p7d-eth0" Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.294 [INFO][4374] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" HandleID="k8s-pod-network.b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" Workload="ci--4547--0--0--6--b8319fed62-k8s-csi--node--driver--m8p7d-eth0" Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.294 [INFO][4374] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" HandleID="k8s-pod-network.b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" Workload="ci--4547--0--0--6--b8319fed62-k8s-csi--node--driver--m8p7d-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00024f830), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4547-0-0-6-b8319fed62", "pod":"csi-node-driver-m8p7d", "timestamp":"2025-12-16 12:48:27.294103207 +0000 UTC"}, Hostname:"ci-4547-0-0-6-b8319fed62", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.294 [INFO][4374] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.294 [INFO][4374] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.294 [INFO][4374] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4547-0-0-6-b8319fed62' Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.306 [INFO][4374] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.314 [INFO][4374] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.323 [INFO][4374] ipam/ipam.go 511: Trying affinity for 192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.326 [INFO][4374] ipam/ipam.go 158: Attempting to load block cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.329 [INFO][4374] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.329 [INFO][4374] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.75.64/26 handle="k8s-pod-network.b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.331 [INFO][4374] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953 Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.337 [INFO][4374] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.75.64/26 handle="k8s-pod-network.b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.345 [INFO][4374] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.75.67/26] block=192.168.75.64/26 handle="k8s-pod-network.b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.345 [INFO][4374] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.75.67/26] handle="k8s-pod-network.b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.346 [INFO][4374] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 16 12:48:27.406795 containerd[1623]: 2025-12-16 12:48:27.347 [INFO][4374] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.75.67/26] IPv6=[] ContainerID="b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" HandleID="k8s-pod-network.b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" Workload="ci--4547--0--0--6--b8319fed62-k8s-csi--node--driver--m8p7d-eth0" Dec 16 12:48:27.408509 containerd[1623]: 2025-12-16 12:48:27.353 [INFO][4360] cni-plugin/k8s.go 418: Populated endpoint ContainerID="b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" Namespace="calico-system" Pod="csi-node-driver-m8p7d" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-csi--node--driver--m8p7d-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-csi--node--driver--m8p7d-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"2053803c-11b9-48f6-949c-d0d885185bfe", ResourceVersion:"684", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 48, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"", Pod:"csi-node-driver-m8p7d", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.75.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali75320bddb25", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:27.408509 containerd[1623]: 2025-12-16 12:48:27.353 [INFO][4360] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.75.67/32] ContainerID="b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" Namespace="calico-system" Pod="csi-node-driver-m8p7d" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-csi--node--driver--m8p7d-eth0" Dec 16 12:48:27.408509 containerd[1623]: 2025-12-16 12:48:27.353 [INFO][4360] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali75320bddb25 ContainerID="b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" Namespace="calico-system" Pod="csi-node-driver-m8p7d" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-csi--node--driver--m8p7d-eth0" Dec 16 12:48:27.408509 containerd[1623]: 2025-12-16 12:48:27.363 [INFO][4360] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" Namespace="calico-system" Pod="csi-node-driver-m8p7d" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-csi--node--driver--m8p7d-eth0" Dec 16 12:48:27.408509 containerd[1623]: 2025-12-16 12:48:27.364 [INFO][4360] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" Namespace="calico-system" Pod="csi-node-driver-m8p7d" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-csi--node--driver--m8p7d-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-csi--node--driver--m8p7d-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"2053803c-11b9-48f6-949c-d0d885185bfe", ResourceVersion:"684", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 48, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953", Pod:"csi-node-driver-m8p7d", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.75.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali75320bddb25", MAC:"f2:fd:eb:63:30:cd", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:27.408509 containerd[1623]: 2025-12-16 12:48:27.396 [INFO][4360] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" Namespace="calico-system" Pod="csi-node-driver-m8p7d" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-csi--node--driver--m8p7d-eth0" Dec 16 12:48:27.422059 kubelet[2796]: I1216 12:48:27.421963 2796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-48cft" podStartSLOduration=33.421942172 podStartE2EDuration="33.421942172s" podCreationTimestamp="2025-12-16 12:47:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-16 12:48:27.42041194 +0000 UTC m=+40.354139967" watchObservedRunningTime="2025-12-16 12:48:27.421942172 +0000 UTC m=+40.355670199" Dec 16 12:48:27.438792 systemd-networkd[1527]: calidf1eb0e5316: Gained IPv6LL Dec 16 12:48:27.474903 containerd[1623]: time="2025-12-16T12:48:27.473658002Z" level=info msg="connecting to shim b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953" address="unix:///run/containerd/s/4814d01914064261b5caf866464d5d3260ea158fc5e3aac89f6eaeebc4a612c1" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:48:27.486000 audit[4389]: NETFILTER_CFG table=filter:126 family=2 entries=20 op=nft_register_rule pid=4389 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:27.486000 audit[4389]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffc4f3b5c10 a2=0 a3=7ffc4f3b5bfc items=0 ppid=2898 pid=4389 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:27.486000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:27.490000 audit[4389]: NETFILTER_CFG table=nat:127 family=2 entries=14 op=nft_register_rule pid=4389 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:27.490000 audit[4389]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffc4f3b5c10 a2=0 a3=0 items=0 ppid=2898 pid=4389 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:27.490000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:27.520000 audit[4407]: NETFILTER_CFG table=filter:128 family=2 entries=40 op=nft_register_chain pid=4407 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Dec 16 12:48:27.520000 audit[4407]: SYSCALL arch=c000003e syscall=46 success=yes exit=20764 a0=3 a1=7ffc03e4e100 a2=0 a3=7ffc03e4e0ec items=0 ppid=4004 pid=4407 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:27.520000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Dec 16 12:48:27.524258 systemd[1]: Started cri-containerd-b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953.scope - libcontainer container b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953. Dec 16 12:48:27.541000 audit[4430]: NETFILTER_CFG table=filter:129 family=2 entries=17 op=nft_register_rule pid=4430 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:27.541000 audit[4430]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffc86481e10 a2=0 a3=7ffc86481dfc items=0 ppid=2898 pid=4430 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:27.541000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:27.543000 audit: BPF prog-id=221 op=LOAD Dec 16 12:48:27.543000 audit: BPF prog-id=222 op=LOAD Dec 16 12:48:27.543000 audit[4409]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=4397 pid=4409 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:27.543000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6238333264303134353739366632333335356666386335343734353464 Dec 16 12:48:27.543000 audit: BPF prog-id=222 op=UNLOAD Dec 16 12:48:27.543000 audit[4409]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4397 pid=4409 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:27.543000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6238333264303134353739366632333335356666386335343734353464 Dec 16 12:48:27.543000 audit: BPF prog-id=223 op=LOAD Dec 16 12:48:27.543000 audit[4409]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=4397 pid=4409 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:27.543000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6238333264303134353739366632333335356666386335343734353464 Dec 16 12:48:27.543000 audit: BPF prog-id=224 op=LOAD Dec 16 12:48:27.543000 audit[4409]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=4397 pid=4409 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:27.543000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6238333264303134353739366632333335356666386335343734353464 Dec 16 12:48:27.543000 audit: BPF prog-id=224 op=UNLOAD Dec 16 12:48:27.543000 audit[4409]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4397 pid=4409 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:27.543000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6238333264303134353739366632333335356666386335343734353464 Dec 16 12:48:27.543000 audit: BPF prog-id=223 op=UNLOAD Dec 16 12:48:27.543000 audit[4409]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4397 pid=4409 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:27.543000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6238333264303134353739366632333335356666386335343734353464 Dec 16 12:48:27.543000 audit: BPF prog-id=225 op=LOAD Dec 16 12:48:27.543000 audit[4409]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=4397 pid=4409 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:27.543000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6238333264303134353739366632333335356666386335343734353464 Dec 16 12:48:27.544000 audit[4430]: NETFILTER_CFG table=nat:130 family=2 entries=35 op=nft_register_chain pid=4430 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:27.544000 audit[4430]: SYSCALL arch=c000003e syscall=46 success=yes exit=14196 a0=3 a1=7ffc86481e10 a2=0 a3=7ffc86481dfc items=0 ppid=2898 pid=4430 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:27.544000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:27.568046 kubelet[2796]: E1216 12:48:27.567985 2796 cadvisor_stats_provider.go:522] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2053803c_11b9_48f6_949c_d0d885185bfe.slice/cri-containerd-b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953.scope\": RecentStats: unable to find data in memory cache]" Dec 16 12:48:27.571788 containerd[1623]: time="2025-12-16T12:48:27.571757429Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-m8p7d,Uid:2053803c-11b9-48f6-949c-d0d885185bfe,Namespace:calico-system,Attempt:0,} returns sandbox id \"b832d0145796f23355ff8c547454d4802f5604516b13490c18320f6b1a44a953\"" Dec 16 12:48:27.573605 containerd[1623]: time="2025-12-16T12:48:27.573585594Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Dec 16 12:48:28.000679 containerd[1623]: time="2025-12-16T12:48:28.000579357Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:28.002169 containerd[1623]: time="2025-12-16T12:48:28.002084754Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Dec 16 12:48:28.002262 containerd[1623]: time="2025-12-16T12:48:28.002187782Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:28.002409 kubelet[2796]: E1216 12:48:28.002367 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 16 12:48:28.002487 kubelet[2796]: E1216 12:48:28.002423 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 16 12:48:28.002683 kubelet[2796]: E1216 12:48:28.002569 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6k7vq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-m8p7d_calico-system(2053803c-11b9-48f6-949c-d0d885185bfe): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:28.005512 containerd[1623]: time="2025-12-16T12:48:28.005437762Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Dec 16 12:48:28.525272 systemd-networkd[1527]: cali75320bddb25: Gained IPv6LL Dec 16 12:48:28.645722 containerd[1623]: time="2025-12-16T12:48:28.645675410Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:28.647009 containerd[1623]: time="2025-12-16T12:48:28.646908440Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Dec 16 12:48:28.647168 containerd[1623]: time="2025-12-16T12:48:28.647084743Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:28.647365 kubelet[2796]: E1216 12:48:28.647298 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 16 12:48:28.647962 kubelet[2796]: E1216 12:48:28.647363 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 16 12:48:28.647962 kubelet[2796]: E1216 12:48:28.647506 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6k7vq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-m8p7d_calico-system(2053803c-11b9-48f6-949c-d0d885185bfe): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:28.649383 kubelet[2796]: E1216 12:48:28.649300 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:48:29.155301 containerd[1623]: time="2025-12-16T12:48:29.154995915Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-srwr6,Uid:d74a60d1-a1a7-4b5b-ad38-2c2ebbe76fd5,Namespace:kube-system,Attempt:0,}" Dec 16 12:48:29.155790 containerd[1623]: time="2025-12-16T12:48:29.155768625Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-5hslq,Uid:0e649fa6-ddd7-43a4-b265-cc0585bd2a9e,Namespace:calico-system,Attempt:0,}" Dec 16 12:48:29.156375 containerd[1623]: time="2025-12-16T12:48:29.156356334Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-98457786-pp6vg,Uid:6200e57e-82e7-4316-9456-f3714e3f66da,Namespace:calico-system,Attempt:0,}" Dec 16 12:48:29.314758 systemd-networkd[1527]: calic728c5bed9f: Link UP Dec 16 12:48:29.315482 systemd-networkd[1527]: calic728c5bed9f: Gained carrier Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.238 [INFO][4444] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--srwr6-eth0 coredns-668d6bf9bc- kube-system d74a60d1-a1a7-4b5b-ad38-2c2ebbe76fd5 782 0 2025-12-16 12:47:54 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4547-0-0-6-b8319fed62 coredns-668d6bf9bc-srwr6 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calic728c5bed9f [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" Namespace="kube-system" Pod="coredns-668d6bf9bc-srwr6" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--srwr6-" Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.239 [INFO][4444] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" Namespace="kube-system" Pod="coredns-668d6bf9bc-srwr6" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--srwr6-eth0" Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.270 [INFO][4491] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" HandleID="k8s-pod-network.e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" Workload="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--srwr6-eth0" Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.270 [INFO][4491] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" HandleID="k8s-pod-network.e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" Workload="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--srwr6-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5870), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4547-0-0-6-b8319fed62", "pod":"coredns-668d6bf9bc-srwr6", "timestamp":"2025-12-16 12:48:29.270299072 +0000 UTC"}, Hostname:"ci-4547-0-0-6-b8319fed62", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.270 [INFO][4491] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.270 [INFO][4491] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.270 [INFO][4491] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4547-0-0-6-b8319fed62' Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.283 [INFO][4491] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.289 [INFO][4491] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.293 [INFO][4491] ipam/ipam.go 511: Trying affinity for 192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.296 [INFO][4491] ipam/ipam.go 158: Attempting to load block cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.298 [INFO][4491] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.298 [INFO][4491] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.75.64/26 handle="k8s-pod-network.e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.299 [INFO][4491] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50 Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.303 [INFO][4491] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.75.64/26 handle="k8s-pod-network.e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.307 [INFO][4491] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.75.68/26] block=192.168.75.64/26 handle="k8s-pod-network.e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.309 [INFO][4491] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.75.68/26] handle="k8s-pod-network.e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.309 [INFO][4491] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 16 12:48:29.336578 containerd[1623]: 2025-12-16 12:48:29.309 [INFO][4491] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.75.68/26] IPv6=[] ContainerID="e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" HandleID="k8s-pod-network.e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" Workload="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--srwr6-eth0" Dec 16 12:48:29.337092 containerd[1623]: 2025-12-16 12:48:29.311 [INFO][4444] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" Namespace="kube-system" Pod="coredns-668d6bf9bc-srwr6" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--srwr6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--srwr6-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"d74a60d1-a1a7-4b5b-ad38-2c2ebbe76fd5", ResourceVersion:"782", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 47, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"", Pod:"coredns-668d6bf9bc-srwr6", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.75.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic728c5bed9f", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:29.337092 containerd[1623]: 2025-12-16 12:48:29.311 [INFO][4444] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.75.68/32] ContainerID="e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" Namespace="kube-system" Pod="coredns-668d6bf9bc-srwr6" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--srwr6-eth0" Dec 16 12:48:29.337092 containerd[1623]: 2025-12-16 12:48:29.311 [INFO][4444] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic728c5bed9f ContainerID="e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" Namespace="kube-system" Pod="coredns-668d6bf9bc-srwr6" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--srwr6-eth0" Dec 16 12:48:29.337092 containerd[1623]: 2025-12-16 12:48:29.315 [INFO][4444] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" Namespace="kube-system" Pod="coredns-668d6bf9bc-srwr6" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--srwr6-eth0" Dec 16 12:48:29.337092 containerd[1623]: 2025-12-16 12:48:29.315 [INFO][4444] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" Namespace="kube-system" Pod="coredns-668d6bf9bc-srwr6" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--srwr6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--srwr6-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"d74a60d1-a1a7-4b5b-ad38-2c2ebbe76fd5", ResourceVersion:"782", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 47, 54, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50", Pod:"coredns-668d6bf9bc-srwr6", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.75.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic728c5bed9f", MAC:"ba:e6:79:ab:0c:79", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:29.337092 containerd[1623]: 2025-12-16 12:48:29.332 [INFO][4444] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" Namespace="kube-system" Pod="coredns-668d6bf9bc-srwr6" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-coredns--668d6bf9bc--srwr6-eth0" Dec 16 12:48:29.345000 audit[4514]: NETFILTER_CFG table=filter:131 family=2 entries=40 op=nft_register_chain pid=4514 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Dec 16 12:48:29.345000 audit[4514]: SYSCALL arch=c000003e syscall=46 success=yes exit=20344 a0=3 a1=7fff87bef1a0 a2=0 a3=7fff87bef18c items=0 ppid=4004 pid=4514 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.345000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Dec 16 12:48:29.364388 containerd[1623]: time="2025-12-16T12:48:29.364220709Z" level=info msg="connecting to shim e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50" address="unix:///run/containerd/s/393215f8989768f2ac3bab1c8ff587a4dd1dfdf37f918bf1176faab62a20662e" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:48:29.390351 systemd[1]: Started cri-containerd-e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50.scope - libcontainer container e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50. Dec 16 12:48:29.405000 audit: BPF prog-id=226 op=LOAD Dec 16 12:48:29.407672 kubelet[2796]: E1216 12:48:29.407575 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:48:29.406000 audit: BPF prog-id=227 op=LOAD Dec 16 12:48:29.406000 audit[4534]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=4523 pid=4534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.406000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6532656630343634323461366362666238643035366633633761663039 Dec 16 12:48:29.408000 audit: BPF prog-id=227 op=UNLOAD Dec 16 12:48:29.408000 audit[4534]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4523 pid=4534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.408000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6532656630343634323461366362666238643035366633633761663039 Dec 16 12:48:29.408000 audit: BPF prog-id=228 op=LOAD Dec 16 12:48:29.408000 audit[4534]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=4523 pid=4534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.408000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6532656630343634323461366362666238643035366633633761663039 Dec 16 12:48:29.408000 audit: BPF prog-id=229 op=LOAD Dec 16 12:48:29.408000 audit[4534]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=4523 pid=4534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.408000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6532656630343634323461366362666238643035366633633761663039 Dec 16 12:48:29.408000 audit: BPF prog-id=229 op=UNLOAD Dec 16 12:48:29.408000 audit[4534]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=4523 pid=4534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.408000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6532656630343634323461366362666238643035366633633761663039 Dec 16 12:48:29.408000 audit: BPF prog-id=228 op=UNLOAD Dec 16 12:48:29.408000 audit[4534]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4523 pid=4534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.408000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6532656630343634323461366362666238643035366633633761663039 Dec 16 12:48:29.408000 audit: BPF prog-id=230 op=LOAD Dec 16 12:48:29.408000 audit[4534]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=4523 pid=4534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.408000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6532656630343634323461366362666238643035366633633761663039 Dec 16 12:48:29.446960 systemd-networkd[1527]: cali8796873e69e: Link UP Dec 16 12:48:29.456132 systemd-networkd[1527]: cali8796873e69e: Gained carrier Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.243 [INFO][4458] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4547--0--0--6--b8319fed62-k8s-calico--kube--controllers--98457786--pp6vg-eth0 calico-kube-controllers-98457786- calico-system 6200e57e-82e7-4316-9456-f3714e3f66da 790 0 2025-12-16 12:48:06 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:98457786 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4547-0-0-6-b8319fed62 calico-kube-controllers-98457786-pp6vg eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali8796873e69e [] [] }} ContainerID="713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" Namespace="calico-system" Pod="calico-kube-controllers-98457786-pp6vg" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--kube--controllers--98457786--pp6vg-" Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.243 [INFO][4458] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" Namespace="calico-system" Pod="calico-kube-controllers-98457786-pp6vg" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--kube--controllers--98457786--pp6vg-eth0" Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.288 [INFO][4493] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" HandleID="k8s-pod-network.713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" Workload="ci--4547--0--0--6--b8319fed62-k8s-calico--kube--controllers--98457786--pp6vg-eth0" Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.289 [INFO][4493] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" HandleID="k8s-pod-network.713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" Workload="ci--4547--0--0--6--b8319fed62-k8s-calico--kube--controllers--98457786--pp6vg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f320), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4547-0-0-6-b8319fed62", "pod":"calico-kube-controllers-98457786-pp6vg", "timestamp":"2025-12-16 12:48:29.288831742 +0000 UTC"}, Hostname:"ci-4547-0-0-6-b8319fed62", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.290 [INFO][4493] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.309 [INFO][4493] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.309 [INFO][4493] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4547-0-0-6-b8319fed62' Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.385 [INFO][4493] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.394 [INFO][4493] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.399 [INFO][4493] ipam/ipam.go 511: Trying affinity for 192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.400 [INFO][4493] ipam/ipam.go 158: Attempting to load block cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.402 [INFO][4493] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.402 [INFO][4493] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.75.64/26 handle="k8s-pod-network.713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.404 [INFO][4493] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1 Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.415 [INFO][4493] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.75.64/26 handle="k8s-pod-network.713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.421 [INFO][4493] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.75.69/26] block=192.168.75.64/26 handle="k8s-pod-network.713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.421 [INFO][4493] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.75.69/26] handle="k8s-pod-network.713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.423 [INFO][4493] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 16 12:48:29.473718 containerd[1623]: 2025-12-16 12:48:29.425 [INFO][4493] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.75.69/26] IPv6=[] ContainerID="713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" HandleID="k8s-pod-network.713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" Workload="ci--4547--0--0--6--b8319fed62-k8s-calico--kube--controllers--98457786--pp6vg-eth0" Dec 16 12:48:29.475600 containerd[1623]: 2025-12-16 12:48:29.438 [INFO][4458] cni-plugin/k8s.go 418: Populated endpoint ContainerID="713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" Namespace="calico-system" Pod="calico-kube-controllers-98457786-pp6vg" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--kube--controllers--98457786--pp6vg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-calico--kube--controllers--98457786--pp6vg-eth0", GenerateName:"calico-kube-controllers-98457786-", Namespace:"calico-system", SelfLink:"", UID:"6200e57e-82e7-4316-9456-f3714e3f66da", ResourceVersion:"790", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 48, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"98457786", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"", Pod:"calico-kube-controllers-98457786-pp6vg", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.75.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali8796873e69e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:29.475600 containerd[1623]: 2025-12-16 12:48:29.438 [INFO][4458] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.75.69/32] ContainerID="713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" Namespace="calico-system" Pod="calico-kube-controllers-98457786-pp6vg" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--kube--controllers--98457786--pp6vg-eth0" Dec 16 12:48:29.475600 containerd[1623]: 2025-12-16 12:48:29.438 [INFO][4458] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali8796873e69e ContainerID="713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" Namespace="calico-system" Pod="calico-kube-controllers-98457786-pp6vg" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--kube--controllers--98457786--pp6vg-eth0" Dec 16 12:48:29.475600 containerd[1623]: 2025-12-16 12:48:29.456 [INFO][4458] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" Namespace="calico-system" Pod="calico-kube-controllers-98457786-pp6vg" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--kube--controllers--98457786--pp6vg-eth0" Dec 16 12:48:29.475600 containerd[1623]: 2025-12-16 12:48:29.459 [INFO][4458] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" Namespace="calico-system" Pod="calico-kube-controllers-98457786-pp6vg" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--kube--controllers--98457786--pp6vg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-calico--kube--controllers--98457786--pp6vg-eth0", GenerateName:"calico-kube-controllers-98457786-", Namespace:"calico-system", SelfLink:"", UID:"6200e57e-82e7-4316-9456-f3714e3f66da", ResourceVersion:"790", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 48, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"98457786", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1", Pod:"calico-kube-controllers-98457786-pp6vg", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.75.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali8796873e69e", MAC:"76:39:a5:38:e1:b2", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:29.475600 containerd[1623]: 2025-12-16 12:48:29.471 [INFO][4458] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" Namespace="calico-system" Pod="calico-kube-controllers-98457786-pp6vg" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--kube--controllers--98457786--pp6vg-eth0" Dec 16 12:48:29.498297 containerd[1623]: time="2025-12-16T12:48:29.498261979Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-srwr6,Uid:d74a60d1-a1a7-4b5b-ad38-2c2ebbe76fd5,Namespace:kube-system,Attempt:0,} returns sandbox id \"e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50\"" Dec 16 12:48:29.504877 containerd[1623]: time="2025-12-16T12:48:29.504843940Z" level=info msg="CreateContainer within sandbox \"e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Dec 16 12:48:29.514426 containerd[1623]: time="2025-12-16T12:48:29.514181771Z" level=info msg="connecting to shim 713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1" address="unix:///run/containerd/s/1409653b02149d81d00c40158fb96e3c5ebb06da1528f98acbdff895396622cf" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:48:29.517182 containerd[1623]: time="2025-12-16T12:48:29.517153728Z" level=info msg="Container 93bf83674b25d8423a3741f3f87ab1e984f76f28017aeedf2a2ad81466c62f56: CDI devices from CRI Config.CDIDevices: []" Dec 16 12:48:29.515000 audit[4570]: NETFILTER_CFG table=filter:132 family=2 entries=48 op=nft_register_chain pid=4570 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Dec 16 12:48:29.515000 audit[4570]: SYSCALL arch=c000003e syscall=46 success=yes exit=23140 a0=3 a1=7fff02fb3420 a2=0 a3=7fff02fb340c items=0 ppid=4004 pid=4570 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.515000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Dec 16 12:48:29.538845 containerd[1623]: time="2025-12-16T12:48:29.538790163Z" level=info msg="CreateContainer within sandbox \"e2ef046424a6cbfb8d056f3c7af09e00e0a7e8b09778e05c028a880bf303ed50\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"93bf83674b25d8423a3741f3f87ab1e984f76f28017aeedf2a2ad81466c62f56\"" Dec 16 12:48:29.540453 containerd[1623]: time="2025-12-16T12:48:29.540402915Z" level=info msg="StartContainer for \"93bf83674b25d8423a3741f3f87ab1e984f76f28017aeedf2a2ad81466c62f56\"" Dec 16 12:48:29.541675 containerd[1623]: time="2025-12-16T12:48:29.541656739Z" level=info msg="connecting to shim 93bf83674b25d8423a3741f3f87ab1e984f76f28017aeedf2a2ad81466c62f56" address="unix:///run/containerd/s/393215f8989768f2ac3bab1c8ff587a4dd1dfdf37f918bf1176faab62a20662e" protocol=ttrpc version=3 Dec 16 12:48:29.563338 systemd[1]: Started cri-containerd-713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1.scope - libcontainer container 713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1. Dec 16 12:48:29.567934 systemd-networkd[1527]: calidb70bb30ae9: Link UP Dec 16 12:48:29.569246 systemd-networkd[1527]: calidb70bb30ae9: Gained carrier Dec 16 12:48:29.588292 systemd[1]: Started cri-containerd-93bf83674b25d8423a3741f3f87ab1e984f76f28017aeedf2a2ad81466c62f56.scope - libcontainer container 93bf83674b25d8423a3741f3f87ab1e984f76f28017aeedf2a2ad81466c62f56. Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.238 [INFO][4447] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4547--0--0--6--b8319fed62-k8s-goldmane--666569f655--5hslq-eth0 goldmane-666569f655- calico-system 0e649fa6-ddd7-43a4-b265-cc0585bd2a9e 792 0 2025-12-16 12:48:04 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:666569f655 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4547-0-0-6-b8319fed62 goldmane-666569f655-5hslq eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] calidb70bb30ae9 [] [] }} ContainerID="fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" Namespace="calico-system" Pod="goldmane-666569f655-5hslq" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-goldmane--666569f655--5hslq-" Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.239 [INFO][4447] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" Namespace="calico-system" Pod="goldmane-666569f655-5hslq" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-goldmane--666569f655--5hslq-eth0" Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.297 [INFO][4485] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" HandleID="k8s-pod-network.fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" Workload="ci--4547--0--0--6--b8319fed62-k8s-goldmane--666569f655--5hslq-eth0" Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.297 [INFO][4485] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" HandleID="k8s-pod-network.fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" Workload="ci--4547--0--0--6--b8319fed62-k8s-goldmane--666569f655--5hslq-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003db470), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4547-0-0-6-b8319fed62", "pod":"goldmane-666569f655-5hslq", "timestamp":"2025-12-16 12:48:29.297575571 +0000 UTC"}, Hostname:"ci-4547-0-0-6-b8319fed62", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.298 [INFO][4485] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.423 [INFO][4485] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.425 [INFO][4485] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4547-0-0-6-b8319fed62' Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.486 [INFO][4485] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.495 [INFO][4485] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.500 [INFO][4485] ipam/ipam.go 511: Trying affinity for 192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.505 [INFO][4485] ipam/ipam.go 158: Attempting to load block cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.511 [INFO][4485] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.512 [INFO][4485] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.75.64/26 handle="k8s-pod-network.fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.515 [INFO][4485] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954 Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.524 [INFO][4485] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.75.64/26 handle="k8s-pod-network.fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.558 [INFO][4485] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.75.70/26] block=192.168.75.64/26 handle="k8s-pod-network.fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.559 [INFO][4485] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.75.70/26] handle="k8s-pod-network.fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.559 [INFO][4485] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 16 12:48:29.591962 containerd[1623]: 2025-12-16 12:48:29.559 [INFO][4485] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.75.70/26] IPv6=[] ContainerID="fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" HandleID="k8s-pod-network.fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" Workload="ci--4547--0--0--6--b8319fed62-k8s-goldmane--666569f655--5hslq-eth0" Dec 16 12:48:29.592495 containerd[1623]: 2025-12-16 12:48:29.561 [INFO][4447] cni-plugin/k8s.go 418: Populated endpoint ContainerID="fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" Namespace="calico-system" Pod="goldmane-666569f655-5hslq" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-goldmane--666569f655--5hslq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-goldmane--666569f655--5hslq-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"0e649fa6-ddd7-43a4-b265-cc0585bd2a9e", ResourceVersion:"792", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 48, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"", Pod:"goldmane-666569f655-5hslq", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.75.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calidb70bb30ae9", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:29.592495 containerd[1623]: 2025-12-16 12:48:29.562 [INFO][4447] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.75.70/32] ContainerID="fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" Namespace="calico-system" Pod="goldmane-666569f655-5hslq" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-goldmane--666569f655--5hslq-eth0" Dec 16 12:48:29.592495 containerd[1623]: 2025-12-16 12:48:29.562 [INFO][4447] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calidb70bb30ae9 ContainerID="fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" Namespace="calico-system" Pod="goldmane-666569f655-5hslq" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-goldmane--666569f655--5hslq-eth0" Dec 16 12:48:29.592495 containerd[1623]: 2025-12-16 12:48:29.574 [INFO][4447] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" Namespace="calico-system" Pod="goldmane-666569f655-5hslq" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-goldmane--666569f655--5hslq-eth0" Dec 16 12:48:29.592495 containerd[1623]: 2025-12-16 12:48:29.574 [INFO][4447] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" Namespace="calico-system" Pod="goldmane-666569f655-5hslq" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-goldmane--666569f655--5hslq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-goldmane--666569f655--5hslq-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"0e649fa6-ddd7-43a4-b265-cc0585bd2a9e", ResourceVersion:"792", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 48, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954", Pod:"goldmane-666569f655-5hslq", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.75.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calidb70bb30ae9", MAC:"0e:f3:b8:10:61:80", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:29.592495 containerd[1623]: 2025-12-16 12:48:29.584 [INFO][4447] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" Namespace="calico-system" Pod="goldmane-666569f655-5hslq" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-goldmane--666569f655--5hslq-eth0" Dec 16 12:48:29.593000 audit: BPF prog-id=231 op=LOAD Dec 16 12:48:29.594000 audit: BPF prog-id=232 op=LOAD Dec 16 12:48:29.594000 audit[4591]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=4580 pid=4591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.594000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3731336362323566373834643730656334343063653137633763633238 Dec 16 12:48:29.595000 audit: BPF prog-id=232 op=UNLOAD Dec 16 12:48:29.595000 audit[4591]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4580 pid=4591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.595000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3731336362323566373834643730656334343063653137633763633238 Dec 16 12:48:29.595000 audit: BPF prog-id=233 op=LOAD Dec 16 12:48:29.595000 audit[4591]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=4580 pid=4591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.595000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3731336362323566373834643730656334343063653137633763633238 Dec 16 12:48:29.595000 audit: BPF prog-id=234 op=LOAD Dec 16 12:48:29.595000 audit[4591]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=4580 pid=4591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.595000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3731336362323566373834643730656334343063653137633763633238 Dec 16 12:48:29.595000 audit: BPF prog-id=234 op=UNLOAD Dec 16 12:48:29.595000 audit[4591]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4580 pid=4591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.595000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3731336362323566373834643730656334343063653137633763633238 Dec 16 12:48:29.595000 audit: BPF prog-id=233 op=UNLOAD Dec 16 12:48:29.595000 audit[4591]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4580 pid=4591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.595000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3731336362323566373834643730656334343063653137633763633238 Dec 16 12:48:29.595000 audit: BPF prog-id=235 op=LOAD Dec 16 12:48:29.595000 audit[4591]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=4580 pid=4591 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.595000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3731336362323566373834643730656334343063653137633763633238 Dec 16 12:48:29.605000 audit: BPF prog-id=236 op=LOAD Dec 16 12:48:29.606000 audit: BPF prog-id=237 op=LOAD Dec 16 12:48:29.606000 audit[4602]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000128238 a2=98 a3=0 items=0 ppid=4523 pid=4602 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.606000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3933626638333637346232356438343233613337343166336638376162 Dec 16 12:48:29.607000 audit: BPF prog-id=237 op=UNLOAD Dec 16 12:48:29.607000 audit[4602]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4523 pid=4602 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.607000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3933626638333637346232356438343233613337343166336638376162 Dec 16 12:48:29.607000 audit: BPF prog-id=238 op=LOAD Dec 16 12:48:29.607000 audit[4602]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000128488 a2=98 a3=0 items=0 ppid=4523 pid=4602 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.607000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3933626638333637346232356438343233613337343166336638376162 Dec 16 12:48:29.609000 audit: BPF prog-id=239 op=LOAD Dec 16 12:48:29.609000 audit[4602]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000128218 a2=98 a3=0 items=0 ppid=4523 pid=4602 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.609000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3933626638333637346232356438343233613337343166336638376162 Dec 16 12:48:29.609000 audit: BPF prog-id=239 op=UNLOAD Dec 16 12:48:29.609000 audit[4602]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4523 pid=4602 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.609000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3933626638333637346232356438343233613337343166336638376162 Dec 16 12:48:29.609000 audit: BPF prog-id=238 op=UNLOAD Dec 16 12:48:29.609000 audit[4602]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4523 pid=4602 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.609000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3933626638333637346232356438343233613337343166336638376162 Dec 16 12:48:29.609000 audit: BPF prog-id=240 op=LOAD Dec 16 12:48:29.609000 audit[4602]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001286e8 a2=98 a3=0 items=0 ppid=4523 pid=4602 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.609000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3933626638333637346232356438343233613337343166336638376162 Dec 16 12:48:29.617000 audit[4636]: NETFILTER_CFG table=filter:133 family=2 entries=66 op=nft_register_chain pid=4636 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Dec 16 12:48:29.617000 audit[4636]: SYSCALL arch=c000003e syscall=46 success=yes exit=32784 a0=3 a1=7ffd0061b9a0 a2=0 a3=7ffd0061b98c items=0 ppid=4004 pid=4636 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.617000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Dec 16 12:48:29.620493 containerd[1623]: time="2025-12-16T12:48:29.620466951Z" level=info msg="connecting to shim fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954" address="unix:///run/containerd/s/cfcf1655057ac475b4a0d4dd2b25811161f0171a64ab11a2a3dc662ba30aa637" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:48:29.631494 containerd[1623]: time="2025-12-16T12:48:29.631443838Z" level=info msg="StartContainer for \"93bf83674b25d8423a3741f3f87ab1e984f76f28017aeedf2a2ad81466c62f56\" returns successfully" Dec 16 12:48:29.650167 systemd[1]: Started cri-containerd-fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954.scope - libcontainer container fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954. Dec 16 12:48:29.692569 containerd[1623]: time="2025-12-16T12:48:29.692453362Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-98457786-pp6vg,Uid:6200e57e-82e7-4316-9456-f3714e3f66da,Namespace:calico-system,Attempt:0,} returns sandbox id \"713cb25f784d70ec440ce17c7cc28414e60f78213fd767885274dcc7d38ab6c1\"" Dec 16 12:48:29.698577 containerd[1623]: time="2025-12-16T12:48:29.698430577Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Dec 16 12:48:29.708000 audit: BPF prog-id=241 op=LOAD Dec 16 12:48:29.709000 audit: BPF prog-id=242 op=LOAD Dec 16 12:48:29.709000 audit[4665]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=4644 pid=4665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.709000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6666663266636666363061626439616531643230323961373265633938 Dec 16 12:48:29.709000 audit: BPF prog-id=242 op=UNLOAD Dec 16 12:48:29.709000 audit[4665]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4644 pid=4665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.709000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6666663266636666363061626439616531643230323961373265633938 Dec 16 12:48:29.709000 audit: BPF prog-id=243 op=LOAD Dec 16 12:48:29.709000 audit[4665]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=4644 pid=4665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.709000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6666663266636666363061626439616531643230323961373265633938 Dec 16 12:48:29.709000 audit: BPF prog-id=244 op=LOAD Dec 16 12:48:29.709000 audit[4665]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=4644 pid=4665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.709000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6666663266636666363061626439616531643230323961373265633938 Dec 16 12:48:29.709000 audit: BPF prog-id=244 op=UNLOAD Dec 16 12:48:29.709000 audit[4665]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4644 pid=4665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.709000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6666663266636666363061626439616531643230323961373265633938 Dec 16 12:48:29.709000 audit: BPF prog-id=243 op=UNLOAD Dec 16 12:48:29.709000 audit[4665]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4644 pid=4665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.709000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6666663266636666363061626439616531643230323961373265633938 Dec 16 12:48:29.709000 audit: BPF prog-id=245 op=LOAD Dec 16 12:48:29.709000 audit[4665]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=4644 pid=4665 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:29.709000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6666663266636666363061626439616531643230323961373265633938 Dec 16 12:48:29.743074 containerd[1623]: time="2025-12-16T12:48:29.742996045Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-5hslq,Uid:0e649fa6-ddd7-43a4-b265-cc0585bd2a9e,Namespace:calico-system,Attempt:0,} returns sandbox id \"fff2fcff60abd9ae1d2029a72ec98c0b674c4d5a9ccc2a6689460d9a5b051954\"" Dec 16 12:48:30.126819 containerd[1623]: time="2025-12-16T12:48:30.126768038Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:30.128079 containerd[1623]: time="2025-12-16T12:48:30.127998715Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Dec 16 12:48:30.128188 containerd[1623]: time="2025-12-16T12:48:30.128148750Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:30.128360 kubelet[2796]: E1216 12:48:30.128310 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 16 12:48:30.128663 kubelet[2796]: E1216 12:48:30.128363 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 16 12:48:30.128735 kubelet[2796]: E1216 12:48:30.128648 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gcpvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-98457786-pp6vg_calico-system(6200e57e-82e7-4316-9456-f3714e3f66da): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:30.129631 containerd[1623]: time="2025-12-16T12:48:30.129610130Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Dec 16 12:48:30.129891 kubelet[2796]: E1216 12:48:30.129817 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-98457786-pp6vg" podUID="6200e57e-82e7-4316-9456-f3714e3f66da" Dec 16 12:48:30.153903 containerd[1623]: time="2025-12-16T12:48:30.153867329Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5668559856-lv4tb,Uid:dd8550e8-a28d-4154-8fdc-fe7811fd3803,Namespace:calico-apiserver,Attempt:0,}" Dec 16 12:48:30.154172 containerd[1623]: time="2025-12-16T12:48:30.153910570Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5668559856-xl9w8,Uid:f4651243-c791-471b-a364-a0f9a5a677b8,Namespace:calico-apiserver,Attempt:0,}" Dec 16 12:48:30.292447 systemd-networkd[1527]: cali68692ed9b6d: Link UP Dec 16 12:48:30.293593 systemd-networkd[1527]: cali68692ed9b6d: Gained carrier Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.218 [INFO][4712] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--lv4tb-eth0 calico-apiserver-5668559856- calico-apiserver dd8550e8-a28d-4154-8fdc-fe7811fd3803 791 0 2025-12-16 12:48:01 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:5668559856 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4547-0-0-6-b8319fed62 calico-apiserver-5668559856-lv4tb eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali68692ed9b6d [] [] }} ContainerID="638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-lv4tb" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--lv4tb-" Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.220 [INFO][4712] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-lv4tb" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--lv4tb-eth0" Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.247 [INFO][4729] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" HandleID="k8s-pod-network.638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" Workload="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--lv4tb-eth0" Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.247 [INFO][4729] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" HandleID="k8s-pod-network.638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" Workload="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--lv4tb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d56e0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4547-0-0-6-b8319fed62", "pod":"calico-apiserver-5668559856-lv4tb", "timestamp":"2025-12-16 12:48:30.247285648 +0000 UTC"}, Hostname:"ci-4547-0-0-6-b8319fed62", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.247 [INFO][4729] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.247 [INFO][4729] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.247 [INFO][4729] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4547-0-0-6-b8319fed62' Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.254 [INFO][4729] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.259 [INFO][4729] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.262 [INFO][4729] ipam/ipam.go 511: Trying affinity for 192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.264 [INFO][4729] ipam/ipam.go 158: Attempting to load block cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.266 [INFO][4729] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.266 [INFO][4729] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.75.64/26 handle="k8s-pod-network.638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.267 [INFO][4729] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89 Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.272 [INFO][4729] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.75.64/26 handle="k8s-pod-network.638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.282 [INFO][4729] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.75.71/26] block=192.168.75.64/26 handle="k8s-pod-network.638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.282 [INFO][4729] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.75.71/26] handle="k8s-pod-network.638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.282 [INFO][4729] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 16 12:48:30.309712 containerd[1623]: 2025-12-16 12:48:30.282 [INFO][4729] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.75.71/26] IPv6=[] ContainerID="638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" HandleID="k8s-pod-network.638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" Workload="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--lv4tb-eth0" Dec 16 12:48:30.313893 containerd[1623]: 2025-12-16 12:48:30.284 [INFO][4712] cni-plugin/k8s.go 418: Populated endpoint ContainerID="638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-lv4tb" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--lv4tb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--lv4tb-eth0", GenerateName:"calico-apiserver-5668559856-", Namespace:"calico-apiserver", SelfLink:"", UID:"dd8550e8-a28d-4154-8fdc-fe7811fd3803", ResourceVersion:"791", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 48, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5668559856", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"", Pod:"calico-apiserver-5668559856-lv4tb", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.75.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali68692ed9b6d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:30.313893 containerd[1623]: 2025-12-16 12:48:30.284 [INFO][4712] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.75.71/32] ContainerID="638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-lv4tb" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--lv4tb-eth0" Dec 16 12:48:30.313893 containerd[1623]: 2025-12-16 12:48:30.284 [INFO][4712] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali68692ed9b6d ContainerID="638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-lv4tb" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--lv4tb-eth0" Dec 16 12:48:30.313893 containerd[1623]: 2025-12-16 12:48:30.293 [INFO][4712] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-lv4tb" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--lv4tb-eth0" Dec 16 12:48:30.313893 containerd[1623]: 2025-12-16 12:48:30.295 [INFO][4712] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-lv4tb" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--lv4tb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--lv4tb-eth0", GenerateName:"calico-apiserver-5668559856-", Namespace:"calico-apiserver", SelfLink:"", UID:"dd8550e8-a28d-4154-8fdc-fe7811fd3803", ResourceVersion:"791", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 48, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5668559856", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89", Pod:"calico-apiserver-5668559856-lv4tb", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.75.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali68692ed9b6d", MAC:"ca:5a:6d:0c:d6:5d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:30.313893 containerd[1623]: 2025-12-16 12:48:30.306 [INFO][4712] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-lv4tb" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--lv4tb-eth0" Dec 16 12:48:30.330000 audit[4750]: NETFILTER_CFG table=filter:134 family=2 entries=66 op=nft_register_chain pid=4750 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Dec 16 12:48:30.330000 audit[4750]: SYSCALL arch=c000003e syscall=46 success=yes exit=32944 a0=3 a1=7ffea8c7c450 a2=0 a3=7ffea8c7c43c items=0 ppid=4004 pid=4750 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.330000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Dec 16 12:48:30.342770 containerd[1623]: time="2025-12-16T12:48:30.342729764Z" level=info msg="connecting to shim 638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89" address="unix:///run/containerd/s/457286bfc5cfd037e2fdb00d1d9bd593ba6851153f4082fdb0be8909a835b0f4" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:48:30.377299 systemd[1]: Started cri-containerd-638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89.scope - libcontainer container 638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89. Dec 16 12:48:30.395000 audit: BPF prog-id=246 op=LOAD Dec 16 12:48:30.396000 audit: BPF prog-id=247 op=LOAD Dec 16 12:48:30.396000 audit[4772]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=4760 pid=4772 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.396000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3633386536663537663533373061396234633034393464333665653933 Dec 16 12:48:30.396000 audit: BPF prog-id=247 op=UNLOAD Dec 16 12:48:30.396000 audit[4772]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4760 pid=4772 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.396000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3633386536663537663533373061396234633034393464333665653933 Dec 16 12:48:30.397000 audit: BPF prog-id=248 op=LOAD Dec 16 12:48:30.397000 audit[4772]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=4760 pid=4772 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.397000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3633386536663537663533373061396234633034393464333665653933 Dec 16 12:48:30.397000 audit: BPF prog-id=249 op=LOAD Dec 16 12:48:30.397000 audit[4772]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=4760 pid=4772 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.397000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3633386536663537663533373061396234633034393464333665653933 Dec 16 12:48:30.397000 audit: BPF prog-id=249 op=UNLOAD Dec 16 12:48:30.397000 audit[4772]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4760 pid=4772 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.397000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3633386536663537663533373061396234633034393464333665653933 Dec 16 12:48:30.397000 audit: BPF prog-id=248 op=UNLOAD Dec 16 12:48:30.397000 audit[4772]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4760 pid=4772 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.397000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3633386536663537663533373061396234633034393464333665653933 Dec 16 12:48:30.398000 audit: BPF prog-id=250 op=LOAD Dec 16 12:48:30.398000 audit[4772]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=4760 pid=4772 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.398000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3633386536663537663533373061396234633034393464333665653933 Dec 16 12:48:30.420627 systemd-networkd[1527]: calic0766619aa4: Link UP Dec 16 12:48:30.421843 systemd-networkd[1527]: calic0766619aa4: Gained carrier Dec 16 12:48:30.422513 kubelet[2796]: E1216 12:48:30.422473 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-98457786-pp6vg" podUID="6200e57e-82e7-4316-9456-f3714e3f66da" Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.220 [INFO][4703] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--xl9w8-eth0 calico-apiserver-5668559856- calico-apiserver f4651243-c791-471b-a364-a0f9a5a677b8 795 0 2025-12-16 12:48:01 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:5668559856 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4547-0-0-6-b8319fed62 calico-apiserver-5668559856-xl9w8 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calic0766619aa4 [] [] }} ContainerID="bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-xl9w8" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--xl9w8-" Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.220 [INFO][4703] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-xl9w8" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--xl9w8-eth0" Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.255 [INFO][4730] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" HandleID="k8s-pod-network.bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" Workload="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--xl9w8-eth0" Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.255 [INFO][4730] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" HandleID="k8s-pod-network.bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" Workload="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--xl9w8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5a30), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4547-0-0-6-b8319fed62", "pod":"calico-apiserver-5668559856-xl9w8", "timestamp":"2025-12-16 12:48:30.255323808 +0000 UTC"}, Hostname:"ci-4547-0-0-6-b8319fed62", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.255 [INFO][4730] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.282 [INFO][4730] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.282 [INFO][4730] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4547-0-0-6-b8319fed62' Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.356 [INFO][4730] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.361 [INFO][4730] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.367 [INFO][4730] ipam/ipam.go 511: Trying affinity for 192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.370 [INFO][4730] ipam/ipam.go 158: Attempting to load block cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.374 [INFO][4730] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.75.64/26 host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.374 [INFO][4730] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.75.64/26 handle="k8s-pod-network.bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.376 [INFO][4730] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5 Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.380 [INFO][4730] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.75.64/26 handle="k8s-pod-network.bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.389 [INFO][4730] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.75.72/26] block=192.168.75.64/26 handle="k8s-pod-network.bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.389 [INFO][4730] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.75.72/26] handle="k8s-pod-network.bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" host="ci-4547-0-0-6-b8319fed62" Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.389 [INFO][4730] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Dec 16 12:48:30.444388 containerd[1623]: 2025-12-16 12:48:30.389 [INFO][4730] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.75.72/26] IPv6=[] ContainerID="bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" HandleID="k8s-pod-network.bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" Workload="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--xl9w8-eth0" Dec 16 12:48:30.444873 containerd[1623]: 2025-12-16 12:48:30.396 [INFO][4703] cni-plugin/k8s.go 418: Populated endpoint ContainerID="bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-xl9w8" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--xl9w8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--xl9w8-eth0", GenerateName:"calico-apiserver-5668559856-", Namespace:"calico-apiserver", SelfLink:"", UID:"f4651243-c791-471b-a364-a0f9a5a677b8", ResourceVersion:"795", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 48, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5668559856", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"", Pod:"calico-apiserver-5668559856-xl9w8", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.75.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calic0766619aa4", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:30.444873 containerd[1623]: 2025-12-16 12:48:30.396 [INFO][4703] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.75.72/32] ContainerID="bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-xl9w8" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--xl9w8-eth0" Dec 16 12:48:30.444873 containerd[1623]: 2025-12-16 12:48:30.396 [INFO][4703] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic0766619aa4 ContainerID="bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-xl9w8" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--xl9w8-eth0" Dec 16 12:48:30.444873 containerd[1623]: 2025-12-16 12:48:30.424 [INFO][4703] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-xl9w8" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--xl9w8-eth0" Dec 16 12:48:30.444873 containerd[1623]: 2025-12-16 12:48:30.426 [INFO][4703] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-xl9w8" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--xl9w8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--xl9w8-eth0", GenerateName:"calico-apiserver-5668559856-", Namespace:"calico-apiserver", SelfLink:"", UID:"f4651243-c791-471b-a364-a0f9a5a677b8", ResourceVersion:"795", Generation:0, CreationTimestamp:time.Date(2025, time.December, 16, 12, 48, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5668559856", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4547-0-0-6-b8319fed62", ContainerID:"bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5", Pod:"calico-apiserver-5668559856-xl9w8", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.75.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calic0766619aa4", MAC:"3a:9d:59:71:80:89", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Dec 16 12:48:30.444873 containerd[1623]: 2025-12-16 12:48:30.441 [INFO][4703] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" Namespace="calico-apiserver" Pod="calico-apiserver-5668559856-xl9w8" WorkloadEndpoint="ci--4547--0--0--6--b8319fed62-k8s-calico--apiserver--5668559856--xl9w8-eth0" Dec 16 12:48:30.487317 containerd[1623]: time="2025-12-16T12:48:30.487233242Z" level=info msg="connecting to shim bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5" address="unix:///run/containerd/s/5a8eb10b2d06126358361fdfd5c5da137b84e5c12d6df8805922489b1aa3ef97" namespace=k8s.io protocol=ttrpc version=3 Dec 16 12:48:30.494000 audit[4812]: NETFILTER_CFG table=filter:135 family=2 entries=63 op=nft_register_chain pid=4812 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Dec 16 12:48:30.494000 audit[4812]: SYSCALL arch=c000003e syscall=46 success=yes exit=30664 a0=3 a1=7ffc3e564c50 a2=0 a3=7ffc3e564c3c items=0 ppid=4004 pid=4812 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.494000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Dec 16 12:48:30.497349 containerd[1623]: time="2025-12-16T12:48:30.497296407Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5668559856-lv4tb,Uid:dd8550e8-a28d-4154-8fdc-fe7811fd3803,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"638e6f57f5370a9b4c0494d36ee938d5cd97dcdb02f68b8eb904f4543a7f5b89\"" Dec 16 12:48:30.512000 audit[4837]: NETFILTER_CFG table=filter:136 family=2 entries=14 op=nft_register_rule pid=4837 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:30.512000 audit[4837]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7fff3af7be50 a2=0 a3=7fff3af7be3c items=0 ppid=2898 pid=4837 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.512000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:30.524201 systemd[1]: Started cri-containerd-bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5.scope - libcontainer container bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5. Dec 16 12:48:30.529000 audit[4837]: NETFILTER_CFG table=nat:137 family=2 entries=56 op=nft_register_chain pid=4837 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:30.529000 audit[4837]: SYSCALL arch=c000003e syscall=46 success=yes exit=19860 a0=3 a1=7fff3af7be50 a2=0 a3=7fff3af7be3c items=0 ppid=2898 pid=4837 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.529000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:30.541000 audit: BPF prog-id=251 op=LOAD Dec 16 12:48:30.541000 audit: BPF prog-id=252 op=LOAD Dec 16 12:48:30.541000 audit[4828]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=4811 pid=4828 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.541000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6263633130386362656530633863663737613133623765636630366430 Dec 16 12:48:30.542000 audit: BPF prog-id=252 op=UNLOAD Dec 16 12:48:30.542000 audit[4828]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4811 pid=4828 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.542000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6263633130386362656530633863663737613133623765636630366430 Dec 16 12:48:30.542000 audit: BPF prog-id=253 op=LOAD Dec 16 12:48:30.542000 audit[4828]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=4811 pid=4828 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.542000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6263633130386362656530633863663737613133623765636630366430 Dec 16 12:48:30.542000 audit: BPF prog-id=254 op=LOAD Dec 16 12:48:30.542000 audit[4828]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=4811 pid=4828 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.542000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6263633130386362656530633863663737613133623765636630366430 Dec 16 12:48:30.542000 audit: BPF prog-id=254 op=UNLOAD Dec 16 12:48:30.542000 audit[4828]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4811 pid=4828 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.542000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6263633130386362656530633863663737613133623765636630366430 Dec 16 12:48:30.542000 audit: BPF prog-id=253 op=UNLOAD Dec 16 12:48:30.542000 audit[4828]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4811 pid=4828 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.542000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6263633130386362656530633863663737613133623765636630366430 Dec 16 12:48:30.543000 audit: BPF prog-id=255 op=LOAD Dec 16 12:48:30.543000 audit[4828]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=4811 pid=4828 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:30.543000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6263633130386362656530633863663737613133623765636630366430 Dec 16 12:48:30.569441 containerd[1623]: time="2025-12-16T12:48:30.569399809Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:30.572279 containerd[1623]: time="2025-12-16T12:48:30.571002697Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Dec 16 12:48:30.572387 containerd[1623]: time="2025-12-16T12:48:30.572092226Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:30.573185 kubelet[2796]: E1216 12:48:30.573135 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 16 12:48:30.573283 kubelet[2796]: E1216 12:48:30.573196 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 16 12:48:30.573486 kubelet[2796]: E1216 12:48:30.573429 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kw75k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-5hslq_calico-system(0e649fa6-ddd7-43a4-b265-cc0585bd2a9e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:30.573878 containerd[1623]: time="2025-12-16T12:48:30.573829657Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 16 12:48:30.575273 kubelet[2796]: E1216 12:48:30.575234 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-5hslq" podUID="0e649fa6-ddd7-43a4-b265-cc0585bd2a9e" Dec 16 12:48:30.587214 containerd[1623]: time="2025-12-16T12:48:30.587160848Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5668559856-xl9w8,Uid:f4651243-c791-471b-a364-a0f9a5a677b8,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"bcc108cbee0c8cf77a13b7ecf06d029cd3f5934447e55e495b0f1f9e267cc7b5\"" Dec 16 12:48:30.829226 systemd-networkd[1527]: cali8796873e69e: Gained IPv6LL Dec 16 12:48:31.014710 containerd[1623]: time="2025-12-16T12:48:31.014580950Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:31.016530 containerd[1623]: time="2025-12-16T12:48:31.016431255Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 16 12:48:31.016756 containerd[1623]: time="2025-12-16T12:48:31.016545173Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:31.016933 kubelet[2796]: E1216 12:48:31.016773 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:48:31.016933 kubelet[2796]: E1216 12:48:31.016836 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:48:31.018068 kubelet[2796]: E1216 12:48:31.017277 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sp8j9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5668559856-lv4tb_calico-apiserver(dd8550e8-a28d-4154-8fdc-fe7811fd3803): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:31.018733 containerd[1623]: time="2025-12-16T12:48:31.017496064Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 16 12:48:31.018858 kubelet[2796]: E1216 12:48:31.018438 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" podUID="dd8550e8-a28d-4154-8fdc-fe7811fd3803" Dec 16 12:48:31.085363 systemd-networkd[1527]: calic728c5bed9f: Gained IPv6LL Dec 16 12:48:31.341560 systemd-networkd[1527]: calidb70bb30ae9: Gained IPv6LL Dec 16 12:48:31.437223 kubelet[2796]: E1216 12:48:31.436053 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" podUID="dd8550e8-a28d-4154-8fdc-fe7811fd3803" Dec 16 12:48:31.439156 kubelet[2796]: E1216 12:48:31.439063 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-5hslq" podUID="0e649fa6-ddd7-43a4-b265-cc0585bd2a9e" Dec 16 12:48:31.439971 kubelet[2796]: E1216 12:48:31.439423 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-98457786-pp6vg" podUID="6200e57e-82e7-4316-9456-f3714e3f66da" Dec 16 12:48:31.450873 containerd[1623]: time="2025-12-16T12:48:31.450623760Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:31.452249 containerd[1623]: time="2025-12-16T12:48:31.452132529Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 16 12:48:31.452249 containerd[1623]: time="2025-12-16T12:48:31.452202516Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:31.452441 kubelet[2796]: E1216 12:48:31.452394 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:48:31.452529 kubelet[2796]: E1216 12:48:31.452515 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:48:31.452743 kubelet[2796]: E1216 12:48:31.452713 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9gsv7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5668559856-xl9w8_calico-apiserver(f4651243-c791-471b-a364-a0f9a5a677b8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:31.456333 kubelet[2796]: E1216 12:48:31.455558 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" podUID="f4651243-c791-471b-a364-a0f9a5a677b8" Dec 16 12:48:31.457051 kubelet[2796]: I1216 12:48:31.456965 2796 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-srwr6" podStartSLOduration=37.456807269 podStartE2EDuration="37.456807269s" podCreationTimestamp="2025-12-16 12:47:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-16 12:48:30.474476458 +0000 UTC m=+43.408204475" watchObservedRunningTime="2025-12-16 12:48:31.456807269 +0000 UTC m=+44.390535296" Dec 16 12:48:31.501074 kernel: kauditd_printk_skb: 227 callbacks suppressed Dec 16 12:48:31.501789 kernel: audit: type=1325 audit(1765889311.497:744): table=filter:138 family=2 entries=14 op=nft_register_rule pid=4856 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:31.497000 audit[4856]: NETFILTER_CFG table=filter:138 family=2 entries=14 op=nft_register_rule pid=4856 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:31.497000 audit[4856]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffca3a2bf10 a2=0 a3=7ffca3a2befc items=0 ppid=2898 pid=4856 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:31.504830 kernel: audit: type=1300 audit(1765889311.497:744): arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffca3a2bf10 a2=0 a3=7ffca3a2befc items=0 ppid=2898 pid=4856 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:31.497000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:31.513571 kernel: audit: type=1327 audit(1765889311.497:744): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:31.509000 audit[4856]: NETFILTER_CFG table=nat:139 family=2 entries=20 op=nft_register_rule pid=4856 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:31.514294 kernel: audit: type=1325 audit(1765889311.509:745): table=nat:139 family=2 entries=20 op=nft_register_rule pid=4856 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:31.509000 audit[4856]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffca3a2bf10 a2=0 a3=7ffca3a2befc items=0 ppid=2898 pid=4856 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:31.509000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:31.524417 kernel: audit: type=1300 audit(1765889311.509:745): arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffca3a2bf10 a2=0 a3=7ffca3a2befc items=0 ppid=2898 pid=4856 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:31.524468 kernel: audit: type=1327 audit(1765889311.509:745): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:31.524000 audit[4858]: NETFILTER_CFG table=filter:140 family=2 entries=14 op=nft_register_rule pid=4858 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:31.524000 audit[4858]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffd77bb0540 a2=0 a3=7ffd77bb052c items=0 ppid=2898 pid=4858 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:31.532816 kernel: audit: type=1325 audit(1765889311.524:746): table=filter:140 family=2 entries=14 op=nft_register_rule pid=4858 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:31.532866 kernel: audit: type=1300 audit(1765889311.524:746): arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffd77bb0540 a2=0 a3=7ffd77bb052c items=0 ppid=2898 pid=4858 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:31.539051 kernel: audit: type=1327 audit(1765889311.524:746): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:31.524000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:31.532000 audit[4858]: NETFILTER_CFG table=nat:141 family=2 entries=20 op=nft_register_rule pid=4858 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:31.543212 kernel: audit: type=1325 audit(1765889311.532:747): table=nat:141 family=2 entries=20 op=nft_register_rule pid=4858 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:31.532000 audit[4858]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffd77bb0540 a2=0 a3=7ffd77bb052c items=0 ppid=2898 pid=4858 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:31.532000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:31.853588 systemd-networkd[1527]: calic0766619aa4: Gained IPv6LL Dec 16 12:48:32.301166 systemd-networkd[1527]: cali68692ed9b6d: Gained IPv6LL Dec 16 12:48:32.437519 kubelet[2796]: E1216 12:48:32.437476 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" podUID="f4651243-c791-471b-a364-a0f9a5a677b8" Dec 16 12:48:32.438905 kubelet[2796]: E1216 12:48:32.437674 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" podUID="dd8550e8-a28d-4154-8fdc-fe7811fd3803" Dec 16 12:48:32.552000 audit[4860]: NETFILTER_CFG table=filter:142 family=2 entries=14 op=nft_register_rule pid=4860 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:32.552000 audit[4860]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffd2bdb7160 a2=0 a3=7ffd2bdb714c items=0 ppid=2898 pid=4860 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:32.552000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:32.558000 audit[4860]: NETFILTER_CFG table=nat:143 family=2 entries=20 op=nft_register_rule pid=4860 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:48:32.558000 audit[4860]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffd2bdb7160 a2=0 a3=7ffd2bdb714c items=0 ppid=2898 pid=4860 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:48:32.558000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:48:39.155779 containerd[1623]: time="2025-12-16T12:48:39.155725491Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Dec 16 12:48:39.572369 containerd[1623]: time="2025-12-16T12:48:39.572315359Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:39.573749 containerd[1623]: time="2025-12-16T12:48:39.573713286Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Dec 16 12:48:39.573823 containerd[1623]: time="2025-12-16T12:48:39.573789412Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:39.573981 kubelet[2796]: E1216 12:48:39.573916 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 16 12:48:39.573981 kubelet[2796]: E1216 12:48:39.573965 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 16 12:48:39.574416 kubelet[2796]: E1216 12:48:39.574084 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:f9261a3a5fe8412497181a324a67db5c,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-79n4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-f48cdf5c5-pnl6x_calico-system(ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:39.577204 containerd[1623]: time="2025-12-16T12:48:39.577177873Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Dec 16 12:48:39.995887 containerd[1623]: time="2025-12-16T12:48:39.995675565Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:39.997231 containerd[1623]: time="2025-12-16T12:48:39.997178277Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Dec 16 12:48:39.997480 containerd[1623]: time="2025-12-16T12:48:39.997276569Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:39.997621 kubelet[2796]: E1216 12:48:39.997563 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 16 12:48:39.997697 kubelet[2796]: E1216 12:48:39.997641 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 16 12:48:39.998084 kubelet[2796]: E1216 12:48:39.997802 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-79n4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-f48cdf5c5-pnl6x_calico-system(ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:39.999200 kubelet[2796]: E1216 12:48:39.999109 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-f48cdf5c5-pnl6x" podUID="ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1" Dec 16 12:48:43.156439 containerd[1623]: time="2025-12-16T12:48:43.156097474Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Dec 16 12:48:43.606796 containerd[1623]: time="2025-12-16T12:48:43.606742726Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:43.608433 containerd[1623]: time="2025-12-16T12:48:43.608367004Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Dec 16 12:48:43.608519 containerd[1623]: time="2025-12-16T12:48:43.608492088Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:43.608668 kubelet[2796]: E1216 12:48:43.608606 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 16 12:48:43.608668 kubelet[2796]: E1216 12:48:43.608657 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 16 12:48:43.610127 kubelet[2796]: E1216 12:48:43.608803 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6k7vq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-m8p7d_calico-system(2053803c-11b9-48f6-949c-d0d885185bfe): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:43.612736 containerd[1623]: time="2025-12-16T12:48:43.612707694Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Dec 16 12:48:44.033952 containerd[1623]: time="2025-12-16T12:48:44.033901853Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:44.035289 containerd[1623]: time="2025-12-16T12:48:44.035261468Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Dec 16 12:48:44.035358 containerd[1623]: time="2025-12-16T12:48:44.035333735Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:44.035545 kubelet[2796]: E1216 12:48:44.035500 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 16 12:48:44.035599 kubelet[2796]: E1216 12:48:44.035569 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 16 12:48:44.036207 kubelet[2796]: E1216 12:48:44.035704 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6k7vq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-m8p7d_calico-system(2053803c-11b9-48f6-949c-d0d885185bfe): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:44.037743 kubelet[2796]: E1216 12:48:44.037702 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:48:45.155934 containerd[1623]: time="2025-12-16T12:48:45.155661832Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 16 12:48:45.596388 containerd[1623]: time="2025-12-16T12:48:45.596338231Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:45.597651 containerd[1623]: time="2025-12-16T12:48:45.597613250Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 16 12:48:45.598061 containerd[1623]: time="2025-12-16T12:48:45.597686378Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:45.598104 kubelet[2796]: E1216 12:48:45.597813 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:48:45.598104 kubelet[2796]: E1216 12:48:45.597873 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:48:45.598419 kubelet[2796]: E1216 12:48:45.598095 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9gsv7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5668559856-xl9w8_calico-apiserver(f4651243-c791-471b-a364-a0f9a5a677b8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:45.599496 containerd[1623]: time="2025-12-16T12:48:45.599192498Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Dec 16 12:48:45.599547 kubelet[2796]: E1216 12:48:45.599445 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" podUID="f4651243-c791-471b-a364-a0f9a5a677b8" Dec 16 12:48:46.039011 containerd[1623]: time="2025-12-16T12:48:46.038935950Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:46.040203 containerd[1623]: time="2025-12-16T12:48:46.040143338Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Dec 16 12:48:46.040316 containerd[1623]: time="2025-12-16T12:48:46.040238090Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:46.040814 kubelet[2796]: E1216 12:48:46.040772 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 16 12:48:46.040955 kubelet[2796]: E1216 12:48:46.040928 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 16 12:48:46.041680 kubelet[2796]: E1216 12:48:46.041513 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gcpvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-98457786-pp6vg_calico-system(6200e57e-82e7-4316-9456-f3714e3f66da): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:46.043243 kubelet[2796]: E1216 12:48:46.043193 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-98457786-pp6vg" podUID="6200e57e-82e7-4316-9456-f3714e3f66da" Dec 16 12:48:47.172864 containerd[1623]: time="2025-12-16T12:48:47.172106950Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 16 12:48:47.623073 containerd[1623]: time="2025-12-16T12:48:47.623037597Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:47.624426 containerd[1623]: time="2025-12-16T12:48:47.624365587Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 16 12:48:47.624527 containerd[1623]: time="2025-12-16T12:48:47.624416900Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:47.624596 kubelet[2796]: E1216 12:48:47.624572 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:48:47.624961 kubelet[2796]: E1216 12:48:47.624610 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:48:47.624961 kubelet[2796]: E1216 12:48:47.624805 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sp8j9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5668559856-lv4tb_calico-apiserver(dd8550e8-a28d-4154-8fdc-fe7811fd3803): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:47.625217 containerd[1623]: time="2025-12-16T12:48:47.625193443Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Dec 16 12:48:47.626843 kubelet[2796]: E1216 12:48:47.626532 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" podUID="dd8550e8-a28d-4154-8fdc-fe7811fd3803" Dec 16 12:48:48.049229 containerd[1623]: time="2025-12-16T12:48:48.049174709Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:48:48.050514 containerd[1623]: time="2025-12-16T12:48:48.050460169Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Dec 16 12:48:48.050588 containerd[1623]: time="2025-12-16T12:48:48.050533298Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Dec 16 12:48:48.050747 kubelet[2796]: E1216 12:48:48.050678 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 16 12:48:48.050822 kubelet[2796]: E1216 12:48:48.050757 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 16 12:48:48.050933 kubelet[2796]: E1216 12:48:48.050880 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kw75k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-5hslq_calico-system(0e649fa6-ddd7-43a4-b265-cc0585bd2a9e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Dec 16 12:48:48.052466 kubelet[2796]: E1216 12:48:48.052420 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-5hslq" podUID="0e649fa6-ddd7-43a4-b265-cc0585bd2a9e" Dec 16 12:48:52.157290 kubelet[2796]: E1216 12:48:52.157174 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-f48cdf5c5-pnl6x" podUID="ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1" Dec 16 12:48:57.160866 kubelet[2796]: E1216 12:48:57.159991 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" podUID="f4651243-c791-471b-a364-a0f9a5a677b8" Dec 16 12:48:58.155958 kubelet[2796]: E1216 12:48:58.155264 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-5hslq" podUID="0e649fa6-ddd7-43a4-b265-cc0585bd2a9e" Dec 16 12:48:58.156169 kubelet[2796]: E1216 12:48:58.156130 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:48:59.158396 kubelet[2796]: E1216 12:48:59.158329 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" podUID="dd8550e8-a28d-4154-8fdc-fe7811fd3803" Dec 16 12:49:00.155242 kubelet[2796]: E1216 12:49:00.155139 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-98457786-pp6vg" podUID="6200e57e-82e7-4316-9456-f3714e3f66da" Dec 16 12:49:07.158134 containerd[1623]: time="2025-12-16T12:49:07.157767629Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Dec 16 12:49:07.626265 containerd[1623]: time="2025-12-16T12:49:07.625777892Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:07.628035 containerd[1623]: time="2025-12-16T12:49:07.627115142Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Dec 16 12:49:07.628308 containerd[1623]: time="2025-12-16T12:49:07.628163806Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:07.628946 kubelet[2796]: E1216 12:49:07.628446 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 16 12:49:07.628946 kubelet[2796]: E1216 12:49:07.628490 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 16 12:49:07.628946 kubelet[2796]: E1216 12:49:07.628576 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:f9261a3a5fe8412497181a324a67db5c,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-79n4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-f48cdf5c5-pnl6x_calico-system(ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:07.631198 containerd[1623]: time="2025-12-16T12:49:07.630758241Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Dec 16 12:49:08.068526 containerd[1623]: time="2025-12-16T12:49:08.068383616Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:08.069768 containerd[1623]: time="2025-12-16T12:49:08.069678636Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Dec 16 12:49:08.069843 containerd[1623]: time="2025-12-16T12:49:08.069760956Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:08.070211 kubelet[2796]: E1216 12:49:08.070132 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 16 12:49:08.070372 kubelet[2796]: E1216 12:49:08.070187 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 16 12:49:08.070640 kubelet[2796]: E1216 12:49:08.070577 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-79n4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-f48cdf5c5-pnl6x_calico-system(ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:08.072156 kubelet[2796]: E1216 12:49:08.071918 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-f48cdf5c5-pnl6x" podUID="ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1" Dec 16 12:49:09.157164 containerd[1623]: time="2025-12-16T12:49:09.157130420Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 16 12:49:09.585802 containerd[1623]: time="2025-12-16T12:49:09.585760675Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:09.587209 containerd[1623]: time="2025-12-16T12:49:09.587139606Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 16 12:49:09.587288 containerd[1623]: time="2025-12-16T12:49:09.587227356Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:09.587724 kubelet[2796]: E1216 12:49:09.587684 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:49:09.588002 kubelet[2796]: E1216 12:49:09.587753 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:49:09.588402 kubelet[2796]: E1216 12:49:09.587883 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9gsv7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5668559856-xl9w8_calico-apiserver(f4651243-c791-471b-a364-a0f9a5a677b8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:09.589608 kubelet[2796]: E1216 12:49:09.589525 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" podUID="f4651243-c791-471b-a364-a0f9a5a677b8" Dec 16 12:49:11.160133 containerd[1623]: time="2025-12-16T12:49:11.159985232Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Dec 16 12:49:11.581148 containerd[1623]: time="2025-12-16T12:49:11.581079710Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:11.582224 containerd[1623]: time="2025-12-16T12:49:11.582169317Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Dec 16 12:49:11.582338 containerd[1623]: time="2025-12-16T12:49:11.582257417Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:11.582592 kubelet[2796]: E1216 12:49:11.582523 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 16 12:49:11.583220 kubelet[2796]: E1216 12:49:11.582603 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 16 12:49:11.583220 kubelet[2796]: E1216 12:49:11.582865 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6k7vq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-m8p7d_calico-system(2053803c-11b9-48f6-949c-d0d885185bfe): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:11.584870 containerd[1623]: time="2025-12-16T12:49:11.584397856Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Dec 16 12:49:12.019839 containerd[1623]: time="2025-12-16T12:49:12.019754888Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:12.021402 containerd[1623]: time="2025-12-16T12:49:12.021237870Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Dec 16 12:49:12.021674 containerd[1623]: time="2025-12-16T12:49:12.021364699Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:12.023416 kubelet[2796]: E1216 12:49:12.023166 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 16 12:49:12.023416 kubelet[2796]: E1216 12:49:12.023224 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 16 12:49:12.023873 kubelet[2796]: E1216 12:49:12.023540 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kw75k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-5hslq_calico-system(0e649fa6-ddd7-43a4-b265-cc0585bd2a9e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:12.024004 containerd[1623]: time="2025-12-16T12:49:12.023673102Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 16 12:49:12.027183 kubelet[2796]: E1216 12:49:12.025355 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-5hslq" podUID="0e649fa6-ddd7-43a4-b265-cc0585bd2a9e" Dec 16 12:49:12.474567 containerd[1623]: time="2025-12-16T12:49:12.474421724Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:12.475591 containerd[1623]: time="2025-12-16T12:49:12.475542103Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 16 12:49:12.475673 containerd[1623]: time="2025-12-16T12:49:12.475645120Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:12.476152 kubelet[2796]: E1216 12:49:12.476094 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:49:12.476227 kubelet[2796]: E1216 12:49:12.476182 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:49:12.476516 kubelet[2796]: E1216 12:49:12.476450 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sp8j9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5668559856-lv4tb_calico-apiserver(dd8550e8-a28d-4154-8fdc-fe7811fd3803): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:12.477346 containerd[1623]: time="2025-12-16T12:49:12.477311980Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Dec 16 12:49:12.478834 kubelet[2796]: E1216 12:49:12.478329 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" podUID="dd8550e8-a28d-4154-8fdc-fe7811fd3803" Dec 16 12:49:12.898131 containerd[1623]: time="2025-12-16T12:49:12.898064415Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:12.899517 containerd[1623]: time="2025-12-16T12:49:12.899453256Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Dec 16 12:49:12.899657 containerd[1623]: time="2025-12-16T12:49:12.899565148Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:12.899892 kubelet[2796]: E1216 12:49:12.899840 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 16 12:49:12.900261 kubelet[2796]: E1216 12:49:12.899931 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 16 12:49:12.900466 kubelet[2796]: E1216 12:49:12.900289 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gcpvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-98457786-pp6vg_calico-system(6200e57e-82e7-4316-9456-f3714e3f66da): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:12.901344 containerd[1623]: time="2025-12-16T12:49:12.901303691Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Dec 16 12:49:12.902455 kubelet[2796]: E1216 12:49:12.902354 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-98457786-pp6vg" podUID="6200e57e-82e7-4316-9456-f3714e3f66da" Dec 16 12:49:13.507381 containerd[1623]: time="2025-12-16T12:49:13.507222390Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:13.508806 containerd[1623]: time="2025-12-16T12:49:13.508697360Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Dec 16 12:49:13.508806 containerd[1623]: time="2025-12-16T12:49:13.508769835Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:13.508976 kubelet[2796]: E1216 12:49:13.508917 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 16 12:49:13.509066 kubelet[2796]: E1216 12:49:13.508986 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 16 12:49:13.509491 kubelet[2796]: E1216 12:49:13.509137 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6k7vq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-m8p7d_calico-system(2053803c-11b9-48f6-949c-d0d885185bfe): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:13.510621 kubelet[2796]: E1216 12:49:13.510580 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:49:20.051000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-77.42.19.41:22-147.75.109.163:54636 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:20.056697 kernel: kauditd_printk_skb: 8 callbacks suppressed Dec 16 12:49:20.056784 kernel: audit: type=1130 audit(1765889360.051:750): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-77.42.19.41:22-147.75.109.163:54636 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:20.050843 systemd[1]: Started sshd@7-77.42.19.41:22-147.75.109.163:54636.service - OpenSSH per-connection server daemon (147.75.109.163:54636). Dec 16 12:49:21.035000 audit[4925]: USER_ACCT pid=4925 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:21.037890 sshd[4925]: Accepted publickey for core from 147.75.109.163 port 54636 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:49:21.044423 sshd-session[4925]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:49:21.046123 kernel: audit: type=1101 audit(1765889361.035:751): pid=4925 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:21.037000 audit[4925]: CRED_ACQ pid=4925 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:21.056282 kernel: audit: type=1103 audit(1765889361.037:752): pid=4925 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:21.056457 kernel: audit: type=1006 audit(1765889361.037:753): pid=4925 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=9 res=1 Dec 16 12:49:21.037000 audit[4925]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd3f91c730 a2=3 a3=0 items=0 ppid=1 pid=4925 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=9 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:21.065075 kernel: audit: type=1300 audit(1765889361.037:753): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd3f91c730 a2=3 a3=0 items=0 ppid=1 pid=4925 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=9 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:21.037000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:21.074749 kernel: audit: type=1327 audit(1765889361.037:753): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:21.075936 systemd-logind[1601]: New session 9 of user core. Dec 16 12:49:21.081326 systemd[1]: Started session-9.scope - Session 9 of User core. Dec 16 12:49:21.085000 audit[4925]: USER_START pid=4925 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:21.096058 kernel: audit: type=1105 audit(1765889361.085:754): pid=4925 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:21.096000 audit[4929]: CRED_ACQ pid=4929 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:21.104070 kernel: audit: type=1103 audit(1765889361.096:755): pid=4929 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:22.143539 sshd[4929]: Connection closed by 147.75.109.163 port 54636 Dec 16 12:49:22.144068 sshd-session[4925]: pam_unix(sshd:session): session closed for user core Dec 16 12:49:22.150000 audit[4925]: USER_END pid=4925 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:22.159046 kernel: audit: type=1106 audit(1765889362.150:756): pid=4925 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:22.160278 kubelet[2796]: E1216 12:49:22.160243 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" podUID="f4651243-c791-471b-a364-a0f9a5a677b8" Dec 16 12:49:22.160997 kubelet[2796]: E1216 12:49:22.160931 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-f48cdf5c5-pnl6x" podUID="ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1" Dec 16 12:49:22.162134 systemd[1]: sshd@7-77.42.19.41:22-147.75.109.163:54636.service: Deactivated successfully. Dec 16 12:49:22.163585 systemd[1]: session-9.scope: Deactivated successfully. Dec 16 12:49:22.150000 audit[4925]: CRED_DISP pid=4925 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:22.167277 systemd-logind[1601]: Session 9 logged out. Waiting for processes to exit. Dec 16 12:49:22.173071 kernel: audit: type=1104 audit(1765889362.150:757): pid=4925 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:22.161000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-77.42.19.41:22-147.75.109.163:54636 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:22.173458 systemd-logind[1601]: Removed session 9. Dec 16 12:49:26.156432 kubelet[2796]: E1216 12:49:26.155951 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-98457786-pp6vg" podUID="6200e57e-82e7-4316-9456-f3714e3f66da" Dec 16 12:49:27.157949 kubelet[2796]: E1216 12:49:27.157277 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" podUID="dd8550e8-a28d-4154-8fdc-fe7811fd3803" Dec 16 12:49:27.158563 kubelet[2796]: E1216 12:49:27.157351 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-5hslq" podUID="0e649fa6-ddd7-43a4-b265-cc0585bd2a9e" Dec 16 12:49:27.302668 kernel: kauditd_printk_skb: 1 callbacks suppressed Dec 16 12:49:27.302836 kernel: audit: type=1130 audit(1765889367.297:759): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-77.42.19.41:22-147.75.109.163:41130 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:27.297000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-77.42.19.41:22-147.75.109.163:41130 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:27.298642 systemd[1]: Started sshd@8-77.42.19.41:22-147.75.109.163:41130.service - OpenSSH per-connection server daemon (147.75.109.163:41130). Dec 16 12:49:28.156771 kubelet[2796]: E1216 12:49:28.156659 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:49:28.191000 audit[4967]: USER_ACCT pid=4967 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:28.206172 kernel: audit: type=1101 audit(1765889368.191:760): pid=4967 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:28.198208 sshd-session[4967]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:49:28.206593 sshd[4967]: Accepted publickey for core from 147.75.109.163 port 41130 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:49:28.216959 systemd-logind[1601]: New session 10 of user core. Dec 16 12:49:28.193000 audit[4967]: CRED_ACQ pid=4967 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:28.228995 systemd[1]: Started session-10.scope - Session 10 of User core. Dec 16 12:49:28.229138 kernel: audit: type=1103 audit(1765889368.193:761): pid=4967 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:28.193000 audit[4967]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc9c0b32e0 a2=3 a3=0 items=0 ppid=1 pid=4967 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:28.239383 kernel: audit: type=1006 audit(1765889368.193:762): pid=4967 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=10 res=1 Dec 16 12:49:28.239466 kernel: audit: type=1300 audit(1765889368.193:762): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc9c0b32e0 a2=3 a3=0 items=0 ppid=1 pid=4967 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:28.193000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:28.247030 kernel: audit: type=1327 audit(1765889368.193:762): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:28.243000 audit[4967]: USER_START pid=4967 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:28.250207 kernel: audit: type=1105 audit(1765889368.243:763): pid=4967 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:28.246000 audit[4971]: CRED_ACQ pid=4971 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:28.258057 kernel: audit: type=1103 audit(1765889368.246:764): pid=4971 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:28.842867 sshd[4971]: Connection closed by 147.75.109.163 port 41130 Dec 16 12:49:28.843987 sshd-session[4967]: pam_unix(sshd:session): session closed for user core Dec 16 12:49:28.845000 audit[4967]: USER_END pid=4967 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:28.850400 systemd[1]: sshd@8-77.42.19.41:22-147.75.109.163:41130.service: Deactivated successfully. Dec 16 12:49:28.855132 systemd[1]: session-10.scope: Deactivated successfully. Dec 16 12:49:28.856134 kernel: audit: type=1106 audit(1765889368.845:765): pid=4967 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:28.846000 audit[4967]: CRED_DISP pid=4967 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:28.860148 systemd-logind[1601]: Session 10 logged out. Waiting for processes to exit. Dec 16 12:49:28.861606 systemd-logind[1601]: Removed session 10. Dec 16 12:49:28.850000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-77.42.19.41:22-147.75.109.163:41130 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:28.865483 kernel: audit: type=1104 audit(1765889368.846:766): pid=4967 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:29.055795 systemd[1]: Started sshd@9-77.42.19.41:22-147.75.109.163:41134.service - OpenSSH per-connection server daemon (147.75.109.163:41134). Dec 16 12:49:29.054000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-77.42.19.41:22-147.75.109.163:41134 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:30.015000 audit[4984]: USER_ACCT pid=4984 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:30.017181 sshd[4984]: Accepted publickey for core from 147.75.109.163 port 41134 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:49:30.018000 audit[4984]: CRED_ACQ pid=4984 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:30.018000 audit[4984]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fffc1db2060 a2=3 a3=0 items=0 ppid=1 pid=4984 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=11 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:30.018000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:30.020240 sshd-session[4984]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:49:30.025408 systemd-logind[1601]: New session 11 of user core. Dec 16 12:49:30.031180 systemd[1]: Started session-11.scope - Session 11 of User core. Dec 16 12:49:30.033000 audit[4984]: USER_START pid=4984 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:30.035000 audit[4988]: CRED_ACQ pid=4988 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:30.664685 sshd[4988]: Connection closed by 147.75.109.163 port 41134 Dec 16 12:49:30.665674 sshd-session[4984]: pam_unix(sshd:session): session closed for user core Dec 16 12:49:30.668000 audit[4984]: USER_END pid=4984 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:30.668000 audit[4984]: CRED_DISP pid=4984 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:30.674888 systemd[1]: sshd@9-77.42.19.41:22-147.75.109.163:41134.service: Deactivated successfully. Dec 16 12:49:30.674000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-77.42.19.41:22-147.75.109.163:41134 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:30.681463 systemd[1]: session-11.scope: Deactivated successfully. Dec 16 12:49:30.684694 systemd-logind[1601]: Session 11 logged out. Waiting for processes to exit. Dec 16 12:49:30.691105 systemd-logind[1601]: Removed session 11. Dec 16 12:49:30.817577 systemd[1]: Started sshd@10-77.42.19.41:22-147.75.109.163:41142.service - OpenSSH per-connection server daemon (147.75.109.163:41142). Dec 16 12:49:30.816000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-77.42.19.41:22-147.75.109.163:41142 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:31.694000 audit[4999]: USER_ACCT pid=4999 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:31.696060 sshd[4999]: Accepted publickey for core from 147.75.109.163 port 41142 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:49:31.695000 audit[4999]: CRED_ACQ pid=4999 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:31.695000 audit[4999]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe2c576b00 a2=3 a3=0 items=0 ppid=1 pid=4999 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=12 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:31.695000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:31.698269 sshd-session[4999]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:49:31.703170 systemd-logind[1601]: New session 12 of user core. Dec 16 12:49:31.709174 systemd[1]: Started session-12.scope - Session 12 of User core. Dec 16 12:49:31.715000 audit[4999]: USER_START pid=4999 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:31.717000 audit[5003]: CRED_ACQ pid=5003 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:32.253678 sshd[5003]: Connection closed by 147.75.109.163 port 41142 Dec 16 12:49:32.255335 sshd-session[4999]: pam_unix(sshd:session): session closed for user core Dec 16 12:49:32.256000 audit[4999]: USER_END pid=4999 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:32.256000 audit[4999]: CRED_DISP pid=4999 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:32.261855 systemd-logind[1601]: Session 12 logged out. Waiting for processes to exit. Dec 16 12:49:32.262320 systemd[1]: sshd@10-77.42.19.41:22-147.75.109.163:41142.service: Deactivated successfully. Dec 16 12:49:32.262000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-77.42.19.41:22-147.75.109.163:41142 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:32.265662 systemd[1]: session-12.scope: Deactivated successfully. Dec 16 12:49:32.269645 systemd-logind[1601]: Removed session 12. Dec 16 12:49:33.156421 kubelet[2796]: E1216 12:49:33.156122 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" podUID="f4651243-c791-471b-a364-a0f9a5a677b8" Dec 16 12:49:35.158862 kubelet[2796]: E1216 12:49:35.157963 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-f48cdf5c5-pnl6x" podUID="ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1" Dec 16 12:49:37.460463 systemd[1]: Started sshd@11-77.42.19.41:22-147.75.109.163:35764.service - OpenSSH per-connection server daemon (147.75.109.163:35764). Dec 16 12:49:37.459000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-77.42.19.41:22-147.75.109.163:35764 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:37.462232 kernel: kauditd_printk_skb: 23 callbacks suppressed Dec 16 12:49:37.462332 kernel: audit: type=1130 audit(1765889377.459:786): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-77.42.19.41:22-147.75.109.163:35764 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:38.154934 kubelet[2796]: E1216 12:49:38.154897 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-5hslq" podUID="0e649fa6-ddd7-43a4-b265-cc0585bd2a9e" Dec 16 12:49:38.416000 audit[5020]: USER_ACCT pid=5020 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:38.418753 sshd[5020]: Accepted publickey for core from 147.75.109.163 port 35764 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:49:38.420954 sshd-session[5020]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:49:38.427140 kernel: audit: type=1101 audit(1765889378.416:787): pid=5020 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:38.417000 audit[5020]: CRED_ACQ pid=5020 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:38.439599 kernel: audit: type=1103 audit(1765889378.417:788): pid=5020 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:38.439663 kernel: audit: type=1006 audit(1765889378.417:789): pid=5020 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=13 res=1 Dec 16 12:49:38.447231 kernel: audit: type=1300 audit(1765889378.417:789): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd6cf04530 a2=3 a3=0 items=0 ppid=1 pid=5020 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=13 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:38.417000 audit[5020]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd6cf04530 a2=3 a3=0 items=0 ppid=1 pid=5020 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=13 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:38.442000 systemd-logind[1601]: New session 13 of user core. Dec 16 12:49:38.450643 kernel: audit: type=1327 audit(1765889378.417:789): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:38.417000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:38.452204 systemd[1]: Started session-13.scope - Session 13 of User core. Dec 16 12:49:38.456000 audit[5020]: USER_START pid=5020 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:38.467035 kernel: audit: type=1105 audit(1765889378.456:790): pid=5020 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:38.465000 audit[5024]: CRED_ACQ pid=5024 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:38.473200 kernel: audit: type=1103 audit(1765889378.465:791): pid=5024 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:39.053203 sshd[5024]: Connection closed by 147.75.109.163 port 35764 Dec 16 12:49:39.054179 sshd-session[5020]: pam_unix(sshd:session): session closed for user core Dec 16 12:49:39.056000 audit[5020]: USER_END pid=5020 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:39.061216 systemd-logind[1601]: Session 13 logged out. Waiting for processes to exit. Dec 16 12:49:39.063081 systemd[1]: sshd@11-77.42.19.41:22-147.75.109.163:35764.service: Deactivated successfully. Dec 16 12:49:39.075538 kernel: audit: type=1106 audit(1765889379.056:792): pid=5020 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:39.068795 systemd[1]: session-13.scope: Deactivated successfully. Dec 16 12:49:39.075352 systemd-logind[1601]: Removed session 13. Dec 16 12:49:39.056000 audit[5020]: CRED_DISP pid=5020 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:39.063000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-77.42.19.41:22-147.75.109.163:35764 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:39.089159 kernel: audit: type=1104 audit(1765889379.056:793): pid=5020 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:40.156909 kubelet[2796]: E1216 12:49:40.156567 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-98457786-pp6vg" podUID="6200e57e-82e7-4316-9456-f3714e3f66da" Dec 16 12:49:41.156517 kubelet[2796]: E1216 12:49:41.156326 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:49:42.156357 kubelet[2796]: E1216 12:49:42.156240 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" podUID="dd8550e8-a28d-4154-8fdc-fe7811fd3803" Dec 16 12:49:44.157363 kubelet[2796]: E1216 12:49:44.156575 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" podUID="f4651243-c791-471b-a364-a0f9a5a677b8" Dec 16 12:49:44.212440 systemd[1]: Started sshd@12-77.42.19.41:22-147.75.109.163:33274.service - OpenSSH per-connection server daemon (147.75.109.163:33274). Dec 16 12:49:44.227154 kernel: kauditd_printk_skb: 1 callbacks suppressed Dec 16 12:49:44.227236 kernel: audit: type=1130 audit(1765889384.212:795): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-77.42.19.41:22-147.75.109.163:33274 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:44.212000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-77.42.19.41:22-147.75.109.163:33274 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:45.073000 audit[5042]: USER_ACCT pid=5042 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:45.076207 sshd[5042]: Accepted publickey for core from 147.75.109.163 port 33274 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:49:45.079785 sshd-session[5042]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:49:45.077000 audit[5042]: CRED_ACQ pid=5042 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:45.092308 kernel: audit: type=1101 audit(1765889385.073:796): pid=5042 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:45.092410 kernel: audit: type=1103 audit(1765889385.077:797): pid=5042 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:45.097801 systemd-logind[1601]: New session 14 of user core. Dec 16 12:49:45.104882 kernel: audit: type=1006 audit(1765889385.077:798): pid=5042 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=14 res=1 Dec 16 12:49:45.105335 systemd[1]: Started session-14.scope - Session 14 of User core. Dec 16 12:49:45.077000 audit[5042]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff3e666a50 a2=3 a3=0 items=0 ppid=1 pid=5042 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:45.114888 kernel: audit: type=1300 audit(1765889385.077:798): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff3e666a50 a2=3 a3=0 items=0 ppid=1 pid=5042 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:45.077000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:45.126706 kernel: audit: type=1327 audit(1765889385.077:798): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:45.114000 audit[5042]: USER_START pid=5042 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:45.133622 kernel: audit: type=1105 audit(1765889385.114:799): pid=5042 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:45.118000 audit[5046]: CRED_ACQ pid=5046 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:45.145355 kernel: audit: type=1103 audit(1765889385.118:800): pid=5046 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:45.743610 sshd[5046]: Connection closed by 147.75.109.163 port 33274 Dec 16 12:49:45.746482 sshd-session[5042]: pam_unix(sshd:session): session closed for user core Dec 16 12:49:45.750000 audit[5042]: USER_END pid=5042 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:45.758000 audit[5042]: CRED_DISP pid=5042 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:45.761741 kernel: audit: type=1106 audit(1765889385.750:801): pid=5042 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:45.761791 kernel: audit: type=1104 audit(1765889385.758:802): pid=5042 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:45.762033 systemd-logind[1601]: Session 14 logged out. Waiting for processes to exit. Dec 16 12:49:45.765152 systemd[1]: sshd@12-77.42.19.41:22-147.75.109.163:33274.service: Deactivated successfully. Dec 16 12:49:45.760000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-77.42.19.41:22-147.75.109.163:33274 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:45.767637 systemd[1]: session-14.scope: Deactivated successfully. Dec 16 12:49:45.769308 systemd-logind[1601]: Removed session 14. Dec 16 12:49:48.155204 containerd[1623]: time="2025-12-16T12:49:48.154861138Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Dec 16 12:49:48.652074 containerd[1623]: time="2025-12-16T12:49:48.651458398Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:48.654449 containerd[1623]: time="2025-12-16T12:49:48.654145139Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Dec 16 12:49:48.654449 containerd[1623]: time="2025-12-16T12:49:48.654257911Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:48.655188 kubelet[2796]: E1216 12:49:48.654498 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 16 12:49:48.655188 kubelet[2796]: E1216 12:49:48.654571 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Dec 16 12:49:48.655188 kubelet[2796]: E1216 12:49:48.654780 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:f9261a3a5fe8412497181a324a67db5c,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-79n4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-f48cdf5c5-pnl6x_calico-system(ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:48.660276 containerd[1623]: time="2025-12-16T12:49:48.660131320Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Dec 16 12:49:49.102155 containerd[1623]: time="2025-12-16T12:49:49.101998818Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:49.104005 containerd[1623]: time="2025-12-16T12:49:49.103969710Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Dec 16 12:49:49.106065 containerd[1623]: time="2025-12-16T12:49:49.104041375Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:49.106207 kubelet[2796]: E1216 12:49:49.104370 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 16 12:49:49.106207 kubelet[2796]: E1216 12:49:49.104422 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Dec 16 12:49:49.106207 kubelet[2796]: E1216 12:49:49.104553 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-79n4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-f48cdf5c5-pnl6x_calico-system(ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:49.106207 kubelet[2796]: E1216 12:49:49.106115 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-f48cdf5c5-pnl6x" podUID="ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1" Dec 16 12:49:50.946135 systemd[1]: Started sshd@13-77.42.19.41:22-147.75.109.163:33284.service - OpenSSH per-connection server daemon (147.75.109.163:33284). Dec 16 12:49:50.945000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-77.42.19.41:22-147.75.109.163:33284 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:50.947938 kernel: kauditd_printk_skb: 1 callbacks suppressed Dec 16 12:49:50.947979 kernel: audit: type=1130 audit(1765889390.945:804): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-77.42.19.41:22-147.75.109.163:33284 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:51.155142 kubelet[2796]: E1216 12:49:51.155098 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-5hslq" podUID="0e649fa6-ddd7-43a4-b265-cc0585bd2a9e" Dec 16 12:49:51.927000 audit[5062]: USER_ACCT pid=5062 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:51.937039 kernel: audit: type=1101 audit(1765889391.927:805): pid=5062 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:51.940163 kernel: audit: type=1103 audit(1765889391.936:806): pid=5062 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:51.936000 audit[5062]: CRED_ACQ pid=5062 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:51.940289 sshd[5062]: Accepted publickey for core from 147.75.109.163 port 33284 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:49:51.946162 sshd-session[5062]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:49:51.947836 kernel: audit: type=1006 audit(1765889391.936:807): pid=5062 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=15 res=1 Dec 16 12:49:51.936000 audit[5062]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff86e2fd30 a2=3 a3=0 items=0 ppid=1 pid=5062 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:51.954336 kernel: audit: type=1300 audit(1765889391.936:807): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff86e2fd30 a2=3 a3=0 items=0 ppid=1 pid=5062 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:51.959227 systemd-logind[1601]: New session 15 of user core. Dec 16 12:49:51.936000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:51.961614 kernel: audit: type=1327 audit(1765889391.936:807): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:51.966318 systemd[1]: Started session-15.scope - Session 15 of User core. Dec 16 12:49:51.968000 audit[5062]: USER_START pid=5062 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:51.971000 audit[5066]: CRED_ACQ pid=5066 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:51.981160 kernel: audit: type=1105 audit(1765889391.968:808): pid=5062 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:51.981234 kernel: audit: type=1103 audit(1765889391.971:809): pid=5066 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:52.574801 sshd[5066]: Connection closed by 147.75.109.163 port 33284 Dec 16 12:49:52.577151 sshd-session[5062]: pam_unix(sshd:session): session closed for user core Dec 16 12:49:52.577000 audit[5062]: USER_END pid=5062 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:52.581465 systemd[1]: sshd@13-77.42.19.41:22-147.75.109.163:33284.service: Deactivated successfully. Dec 16 12:49:52.582828 systemd[1]: session-15.scope: Deactivated successfully. Dec 16 12:49:52.583695 systemd-logind[1601]: Session 15 logged out. Waiting for processes to exit. Dec 16 12:49:52.587605 kernel: audit: type=1106 audit(1765889392.577:810): pid=5062 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:52.588329 systemd-logind[1601]: Removed session 15. Dec 16 12:49:52.577000 audit[5062]: CRED_DISP pid=5062 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:52.596040 kernel: audit: type=1104 audit(1765889392.577:811): pid=5062 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:52.580000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-77.42.19.41:22-147.75.109.163:33284 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:52.724000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-77.42.19.41:22-147.75.109.163:46848 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:52.725229 systemd[1]: Started sshd@14-77.42.19.41:22-147.75.109.163:46848.service - OpenSSH per-connection server daemon (147.75.109.163:46848). Dec 16 12:49:53.560000 audit[5102]: USER_ACCT pid=5102 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:53.561000 audit[5102]: CRED_ACQ pid=5102 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:53.563158 sshd[5102]: Accepted publickey for core from 147.75.109.163 port 46848 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:49:53.562000 audit[5102]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd472d0020 a2=3 a3=0 items=0 ppid=1 pid=5102 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:53.562000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:53.564739 sshd-session[5102]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:49:53.575607 systemd-logind[1601]: New session 16 of user core. Dec 16 12:49:53.579180 systemd[1]: Started session-16.scope - Session 16 of User core. Dec 16 12:49:53.582000 audit[5102]: USER_START pid=5102 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:53.584000 audit[5106]: CRED_ACQ pid=5106 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:54.154856 containerd[1623]: time="2025-12-16T12:49:54.154811556Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 16 12:49:54.427593 sshd[5106]: Connection closed by 147.75.109.163 port 46848 Dec 16 12:49:54.428270 sshd-session[5102]: pam_unix(sshd:session): session closed for user core Dec 16 12:49:54.429000 audit[5102]: USER_END pid=5102 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:54.429000 audit[5102]: CRED_DISP pid=5102 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:54.442303 systemd[1]: sshd@14-77.42.19.41:22-147.75.109.163:46848.service: Deactivated successfully. Dec 16 12:49:54.441000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-77.42.19.41:22-147.75.109.163:46848 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:54.444800 systemd[1]: session-16.scope: Deactivated successfully. Dec 16 12:49:54.446380 systemd-logind[1601]: Session 16 logged out. Waiting for processes to exit. Dec 16 12:49:54.449141 systemd-logind[1601]: Removed session 16. Dec 16 12:49:54.595747 systemd[1]: Started sshd@15-77.42.19.41:22-147.75.109.163:46856.service - OpenSSH per-connection server daemon (147.75.109.163:46856). Dec 16 12:49:54.594000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-77.42.19.41:22-147.75.109.163:46856 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:54.607037 containerd[1623]: time="2025-12-16T12:49:54.606290939Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:54.608180 containerd[1623]: time="2025-12-16T12:49:54.608107511Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 16 12:49:54.608270 containerd[1623]: time="2025-12-16T12:49:54.608245451Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:54.608447 kubelet[2796]: E1216 12:49:54.608374 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:49:54.608732 kubelet[2796]: E1216 12:49:54.608455 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:49:54.608732 kubelet[2796]: E1216 12:49:54.608664 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sp8j9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5668559856-lv4tb_calico-apiserver(dd8550e8-a28d-4154-8fdc-fe7811fd3803): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:54.609295 containerd[1623]: time="2025-12-16T12:49:54.609266499Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Dec 16 12:49:54.610353 kubelet[2796]: E1216 12:49:54.610310 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" podUID="dd8550e8-a28d-4154-8fdc-fe7811fd3803" Dec 16 12:49:55.030721 containerd[1623]: time="2025-12-16T12:49:55.030657592Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:55.032511 containerd[1623]: time="2025-12-16T12:49:55.032452254Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Dec 16 12:49:55.032639 containerd[1623]: time="2025-12-16T12:49:55.032608420Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:55.032826 kubelet[2796]: E1216 12:49:55.032771 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 16 12:49:55.032894 kubelet[2796]: E1216 12:49:55.032845 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Dec 16 12:49:55.033544 kubelet[2796]: E1216 12:49:55.032995 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6k7vq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-m8p7d_calico-system(2053803c-11b9-48f6-949c-d0d885185bfe): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:55.036829 containerd[1623]: time="2025-12-16T12:49:55.036789284Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Dec 16 12:49:55.455000 audit[5116]: USER_ACCT pid=5116 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:55.458039 sshd[5116]: Accepted publickey for core from 147.75.109.163 port 46856 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:49:55.457000 audit[5116]: CRED_ACQ pid=5116 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:55.457000 audit[5116]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffeae9771e0 a2=3 a3=0 items=0 ppid=1 pid=5116 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=17 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:55.457000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:55.459739 sshd-session[5116]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:49:55.466893 systemd-logind[1601]: New session 17 of user core. Dec 16 12:49:55.470269 containerd[1623]: time="2025-12-16T12:49:55.470207375Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:55.472158 containerd[1623]: time="2025-12-16T12:49:55.472119480Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Dec 16 12:49:55.472212 containerd[1623]: time="2025-12-16T12:49:55.472180396Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:55.472590 kubelet[2796]: E1216 12:49:55.472395 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 16 12:49:55.472590 kubelet[2796]: E1216 12:49:55.472440 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Dec 16 12:49:55.472677 kubelet[2796]: E1216 12:49:55.472644 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6k7vq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-m8p7d_calico-system(2053803c-11b9-48f6-949c-d0d885185bfe): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:55.472983 containerd[1623]: time="2025-12-16T12:49:55.472959107Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Dec 16 12:49:55.473234 systemd[1]: Started session-17.scope - Session 17 of User core. Dec 16 12:49:55.476604 kubelet[2796]: E1216 12:49:55.476275 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:49:55.476000 audit[5116]: USER_START pid=5116 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:55.478000 audit[5122]: CRED_ACQ pid=5122 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:56.102813 containerd[1623]: time="2025-12-16T12:49:56.102763208Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:56.104232 containerd[1623]: time="2025-12-16T12:49:56.104169990Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Dec 16 12:49:56.105132 containerd[1623]: time="2025-12-16T12:49:56.104207681Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:56.105370 kubelet[2796]: E1216 12:49:56.105324 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 16 12:49:56.105924 kubelet[2796]: E1216 12:49:56.105689 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Dec 16 12:49:56.105924 kubelet[2796]: E1216 12:49:56.105854 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gcpvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-98457786-pp6vg_calico-system(6200e57e-82e7-4316-9456-f3714e3f66da): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:56.107326 kubelet[2796]: E1216 12:49:56.107294 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-98457786-pp6vg" podUID="6200e57e-82e7-4316-9456-f3714e3f66da" Dec 16 12:49:56.486063 kernel: kauditd_printk_skb: 20 callbacks suppressed Dec 16 12:49:56.486395 kernel: audit: type=1325 audit(1765889396.481:828): table=filter:144 family=2 entries=26 op=nft_register_rule pid=5132 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:49:56.481000 audit[5132]: NETFILTER_CFG table=filter:144 family=2 entries=26 op=nft_register_rule pid=5132 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:49:56.481000 audit[5132]: SYSCALL arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7fff373425b0 a2=0 a3=7fff3734259c items=0 ppid=2898 pid=5132 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:56.494363 kernel: audit: type=1300 audit(1765889396.481:828): arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7fff373425b0 a2=0 a3=7fff3734259c items=0 ppid=2898 pid=5132 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:56.504383 kernel: audit: type=1327 audit(1765889396.481:828): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:49:56.481000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:49:56.510506 kernel: audit: type=1325 audit(1765889396.493:829): table=nat:145 family=2 entries=20 op=nft_register_rule pid=5132 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:49:56.493000 audit[5132]: NETFILTER_CFG table=nat:145 family=2 entries=20 op=nft_register_rule pid=5132 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:49:56.519956 kernel: audit: type=1300 audit(1765889396.493:829): arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7fff373425b0 a2=0 a3=0 items=0 ppid=2898 pid=5132 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:56.525200 kernel: audit: type=1327 audit(1765889396.493:829): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:49:56.493000 audit[5132]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7fff373425b0 a2=0 a3=0 items=0 ppid=2898 pid=5132 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:56.493000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:49:56.539000 audit[5134]: NETFILTER_CFG table=filter:146 family=2 entries=38 op=nft_register_rule pid=5134 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:49:56.546476 kernel: audit: type=1325 audit(1765889396.539:830): table=filter:146 family=2 entries=38 op=nft_register_rule pid=5134 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:49:56.546557 kernel: audit: type=1300 audit(1765889396.539:830): arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7ffd01f3c7c0 a2=0 a3=7ffd01f3c7ac items=0 ppid=2898 pid=5134 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:56.539000 audit[5134]: SYSCALL arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7ffd01f3c7c0 a2=0 a3=7ffd01f3c7ac items=0 ppid=2898 pid=5134 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:56.539000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:49:56.555991 kernel: audit: type=1327 audit(1765889396.539:830): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:49:56.547000 audit[5134]: NETFILTER_CFG table=nat:147 family=2 entries=20 op=nft_register_rule pid=5134 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:49:56.560817 kernel: audit: type=1325 audit(1765889396.547:831): table=nat:147 family=2 entries=20 op=nft_register_rule pid=5134 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:49:56.547000 audit[5134]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffd01f3c7c0 a2=0 a3=0 items=0 ppid=2898 pid=5134 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:56.547000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:49:56.633985 sshd[5122]: Connection closed by 147.75.109.163 port 46856 Dec 16 12:49:56.638312 sshd-session[5116]: pam_unix(sshd:session): session closed for user core Dec 16 12:49:56.643000 audit[5116]: USER_END pid=5116 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:56.644000 audit[5116]: CRED_DISP pid=5116 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:56.647476 systemd[1]: sshd@15-77.42.19.41:22-147.75.109.163:46856.service: Deactivated successfully. Dec 16 12:49:56.647000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-77.42.19.41:22-147.75.109.163:46856 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:56.651040 systemd[1]: session-17.scope: Deactivated successfully. Dec 16 12:49:56.652853 systemd-logind[1601]: Session 17 logged out. Waiting for processes to exit. Dec 16 12:49:56.655136 systemd-logind[1601]: Removed session 17. Dec 16 12:49:56.845866 systemd[1]: Started sshd@16-77.42.19.41:22-147.75.109.163:46862.service - OpenSSH per-connection server daemon (147.75.109.163:46862). Dec 16 12:49:56.844000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-77.42.19.41:22-147.75.109.163:46862 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:57.804000 audit[5139]: USER_ACCT pid=5139 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:57.806130 sshd[5139]: Accepted publickey for core from 147.75.109.163 port 46862 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:49:57.805000 audit[5139]: CRED_ACQ pid=5139 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:57.805000 audit[5139]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff3f7b8560 a2=3 a3=0 items=0 ppid=1 pid=5139 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=18 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:57.805000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:57.807966 sshd-session[5139]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:49:57.813048 systemd-logind[1601]: New session 18 of user core. Dec 16 12:49:57.818138 systemd[1]: Started session-18.scope - Session 18 of User core. Dec 16 12:49:57.819000 audit[5139]: USER_START pid=5139 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:57.821000 audit[5157]: CRED_ACQ pid=5157 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:58.674054 sshd[5157]: Connection closed by 147.75.109.163 port 46862 Dec 16 12:49:58.674660 sshd-session[5139]: pam_unix(sshd:session): session closed for user core Dec 16 12:49:58.679000 audit[5139]: USER_END pid=5139 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:58.679000 audit[5139]: CRED_DISP pid=5139 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:58.683424 systemd-logind[1601]: Session 18 logged out. Waiting for processes to exit. Dec 16 12:49:58.686512 systemd[1]: sshd@16-77.42.19.41:22-147.75.109.163:46862.service: Deactivated successfully. Dec 16 12:49:58.686000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-77.42.19.41:22-147.75.109.163:46862 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:58.694045 systemd[1]: session-18.scope: Deactivated successfully. Dec 16 12:49:58.699052 systemd-logind[1601]: Removed session 18. Dec 16 12:49:58.823000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-77.42.19.41:22-147.75.109.163:46868 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:49:58.824334 systemd[1]: Started sshd@17-77.42.19.41:22-147.75.109.163:46868.service - OpenSSH per-connection server daemon (147.75.109.163:46868). Dec 16 12:49:59.157372 containerd[1623]: time="2025-12-16T12:49:59.157297376Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Dec 16 12:49:59.586597 containerd[1623]: time="2025-12-16T12:49:59.586484046Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:49:59.588565 containerd[1623]: time="2025-12-16T12:49:59.588451770Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Dec 16 12:49:59.588952 containerd[1623]: time="2025-12-16T12:49:59.588680455Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Dec 16 12:49:59.589466 kubelet[2796]: E1216 12:49:59.589245 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:49:59.589466 kubelet[2796]: E1216 12:49:59.589374 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Dec 16 12:49:59.590767 kubelet[2796]: E1216 12:49:59.589533 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9gsv7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5668559856-xl9w8_calico-apiserver(f4651243-c791-471b-a364-a0f9a5a677b8): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Dec 16 12:49:59.590767 kubelet[2796]: E1216 12:49:59.590744 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" podUID="f4651243-c791-471b-a364-a0f9a5a677b8" Dec 16 12:49:59.692000 audit[5167]: USER_ACCT pid=5167 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:59.694472 sshd[5167]: Accepted publickey for core from 147.75.109.163 port 46868 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:49:59.698000 audit[5167]: CRED_ACQ pid=5167 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:59.699000 audit[5167]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe1d8ccbb0 a2=3 a3=0 items=0 ppid=1 pid=5167 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=19 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:49:59.699000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:49:59.702460 sshd-session[5167]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:49:59.712375 systemd-logind[1601]: New session 19 of user core. Dec 16 12:49:59.723284 systemd[1]: Started session-19.scope - Session 19 of User core. Dec 16 12:49:59.729000 audit[5167]: USER_START pid=5167 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:49:59.732000 audit[5172]: CRED_ACQ pid=5172 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:00.257553 sshd[5172]: Connection closed by 147.75.109.163 port 46868 Dec 16 12:50:00.258244 sshd-session[5167]: pam_unix(sshd:session): session closed for user core Dec 16 12:50:00.259000 audit[5167]: USER_END pid=5167 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:00.259000 audit[5167]: CRED_DISP pid=5167 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:00.262862 systemd[1]: sshd@17-77.42.19.41:22-147.75.109.163:46868.service: Deactivated successfully. Dec 16 12:50:00.262000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-77.42.19.41:22-147.75.109.163:46868 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:50:00.263179 systemd-logind[1601]: Session 19 logged out. Waiting for processes to exit. Dec 16 12:50:00.265051 systemd[1]: session-19.scope: Deactivated successfully. Dec 16 12:50:00.266902 systemd-logind[1601]: Removed session 19. Dec 16 12:50:00.527000 audit[5184]: NETFILTER_CFG table=filter:148 family=2 entries=26 op=nft_register_rule pid=5184 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:50:00.527000 audit[5184]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffed09c3c70 a2=0 a3=7ffed09c3c5c items=0 ppid=2898 pid=5184 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:00.527000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:50:00.531000 audit[5184]: NETFILTER_CFG table=nat:149 family=2 entries=104 op=nft_register_chain pid=5184 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Dec 16 12:50:00.531000 audit[5184]: SYSCALL arch=c000003e syscall=46 success=yes exit=48684 a0=3 a1=7ffed09c3c70 a2=0 a3=7ffed09c3c5c items=0 ppid=2898 pid=5184 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:00.531000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Dec 16 12:50:02.156646 kubelet[2796]: E1216 12:50:02.156541 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-f48cdf5c5-pnl6x" podUID="ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1" Dec 16 12:50:04.156304 containerd[1623]: time="2025-12-16T12:50:04.156197704Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Dec 16 12:50:04.622640 containerd[1623]: time="2025-12-16T12:50:04.622428973Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Dec 16 12:50:04.623620 containerd[1623]: time="2025-12-16T12:50:04.623555078Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Dec 16 12:50:04.623807 containerd[1623]: time="2025-12-16T12:50:04.623652824Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Dec 16 12:50:04.624221 kubelet[2796]: E1216 12:50:04.624081 2796 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 16 12:50:04.624221 kubelet[2796]: E1216 12:50:04.624159 2796 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Dec 16 12:50:04.625450 kubelet[2796]: E1216 12:50:04.625218 2796 kuberuntime_manager.go:1341] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kw75k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-5hslq_calico-system(0e649fa6-ddd7-43a4-b265-cc0585bd2a9e): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Dec 16 12:50:04.626731 kubelet[2796]: E1216 12:50:04.626675 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-5hslq" podUID="0e649fa6-ddd7-43a4-b265-cc0585bd2a9e" Dec 16 12:50:05.423996 systemd[1]: Started sshd@18-77.42.19.41:22-147.75.109.163:46476.service - OpenSSH per-connection server daemon (147.75.109.163:46476). Dec 16 12:50:05.425000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-77.42.19.41:22-147.75.109.163:46476 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:50:05.427290 kernel: kauditd_printk_skb: 33 callbacks suppressed Dec 16 12:50:05.427666 kernel: audit: type=1130 audit(1765889405.425:855): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-77.42.19.41:22-147.75.109.163:46476 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:50:06.268000 audit[5193]: USER_ACCT pid=5193 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:06.271406 sshd-session[5193]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:50:06.272424 sshd[5193]: Accepted publickey for core from 147.75.109.163 port 46476 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:50:06.269000 audit[5193]: CRED_ACQ pid=5193 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:06.280380 kernel: audit: type=1101 audit(1765889406.268:856): pid=5193 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:06.280434 kernel: audit: type=1103 audit(1765889406.269:857): pid=5193 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:06.281538 systemd-logind[1601]: New session 20 of user core. Dec 16 12:50:06.287885 kernel: audit: type=1006 audit(1765889406.269:858): pid=5193 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=20 res=1 Dec 16 12:50:06.269000 audit[5193]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc26eb01c0 a2=3 a3=0 items=0 ppid=1 pid=5193 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=20 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:06.292279 systemd[1]: Started session-20.scope - Session 20 of User core. Dec 16 12:50:06.294205 kernel: audit: type=1300 audit(1765889406.269:858): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc26eb01c0 a2=3 a3=0 items=0 ppid=1 pid=5193 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=20 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:06.269000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:50:06.302463 kernel: audit: type=1327 audit(1765889406.269:858): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:50:06.297000 audit[5193]: USER_START pid=5193 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:06.307585 kernel: audit: type=1105 audit(1765889406.297:859): pid=5193 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:06.301000 audit[5197]: CRED_ACQ pid=5197 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:06.316666 kernel: audit: type=1103 audit(1765889406.301:860): pid=5197 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:06.810316 sshd[5197]: Connection closed by 147.75.109.163 port 46476 Dec 16 12:50:06.812153 sshd-session[5193]: pam_unix(sshd:session): session closed for user core Dec 16 12:50:06.812000 audit[5193]: USER_END pid=5193 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:06.826580 kernel: audit: type=1106 audit(1765889406.812:861): pid=5193 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:06.825122 systemd[1]: sshd@18-77.42.19.41:22-147.75.109.163:46476.service: Deactivated successfully. Dec 16 12:50:06.826821 systemd-logind[1601]: Session 20 logged out. Waiting for processes to exit. Dec 16 12:50:06.829324 systemd[1]: session-20.scope: Deactivated successfully. Dec 16 12:50:06.833339 systemd-logind[1601]: Removed session 20. Dec 16 12:50:06.812000 audit[5193]: CRED_DISP pid=5193 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:06.841357 kernel: audit: type=1104 audit(1765889406.812:862): pid=5193 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:06.824000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-77.42.19.41:22-147.75.109.163:46476 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:50:07.160440 kubelet[2796]: E1216 12:50:07.159639 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:50:08.157187 kubelet[2796]: E1216 12:50:08.156369 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-98457786-pp6vg" podUID="6200e57e-82e7-4316-9456-f3714e3f66da" Dec 16 12:50:09.160012 kubelet[2796]: E1216 12:50:09.159590 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" podUID="dd8550e8-a28d-4154-8fdc-fe7811fd3803" Dec 16 12:50:11.987127 kernel: kauditd_printk_skb: 1 callbacks suppressed Dec 16 12:50:11.987233 kernel: audit: type=1130 audit(1765889411.978:864): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-77.42.19.41:22-147.75.109.163:46488 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:50:11.978000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-77.42.19.41:22-147.75.109.163:46488 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:50:11.979110 systemd[1]: Started sshd@19-77.42.19.41:22-147.75.109.163:46488.service - OpenSSH per-connection server daemon (147.75.109.163:46488). Dec 16 12:50:12.820000 audit[5209]: USER_ACCT pid=5209 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:12.830323 sshd[5209]: Accepted publickey for core from 147.75.109.163 port 46488 ssh2: RSA SHA256:EtC75xtaTSJ+84wEKbjmapFbgFAJxSn7WaNHLp2aTq4 Dec 16 12:50:12.832549 sshd-session[5209]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Dec 16 12:50:12.838049 kernel: audit: type=1101 audit(1765889412.820:865): pid=5209 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:12.838148 kernel: audit: type=1103 audit(1765889412.830:866): pid=5209 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:12.830000 audit[5209]: CRED_ACQ pid=5209 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:12.853511 systemd-logind[1601]: New session 21 of user core. Dec 16 12:50:12.859080 kernel: audit: type=1006 audit(1765889412.830:867): pid=5209 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=21 res=1 Dec 16 12:50:12.830000 audit[5209]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd9cc45790 a2=3 a3=0 items=0 ppid=1 pid=5209 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=21 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:12.830000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:50:12.875768 kernel: audit: type=1300 audit(1765889412.830:867): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd9cc45790 a2=3 a3=0 items=0 ppid=1 pid=5209 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=21 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:12.875986 kernel: audit: type=1327 audit(1765889412.830:867): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Dec 16 12:50:12.876293 systemd[1]: Started session-21.scope - Session 21 of User core. Dec 16 12:50:12.883000 audit[5209]: USER_START pid=5209 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:12.897254 kernel: audit: type=1105 audit(1765889412.883:868): pid=5209 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:12.898000 audit[5213]: CRED_ACQ pid=5213 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:12.908085 kernel: audit: type=1103 audit(1765889412.898:869): pid=5213 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:13.158479 kubelet[2796]: E1216 12:50:13.158383 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-f48cdf5c5-pnl6x" podUID="ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1" Dec 16 12:50:13.372508 sshd[5213]: Connection closed by 147.75.109.163 port 46488 Dec 16 12:50:13.372954 sshd-session[5209]: pam_unix(sshd:session): session closed for user core Dec 16 12:50:13.373000 audit[5209]: USER_END pid=5209 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:13.377787 systemd-logind[1601]: Session 21 logged out. Waiting for processes to exit. Dec 16 12:50:13.378978 systemd[1]: sshd@19-77.42.19.41:22-147.75.109.163:46488.service: Deactivated successfully. Dec 16 12:50:13.383036 kernel: audit: type=1106 audit(1765889413.373:870): pid=5209 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:13.383768 systemd[1]: session-21.scope: Deactivated successfully. Dec 16 12:50:13.385861 systemd-logind[1601]: Removed session 21. Dec 16 12:50:13.373000 audit[5209]: CRED_DISP pid=5209 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:13.376000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-77.42.19.41:22-147.75.109.163:46488 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 16 12:50:13.393047 kernel: audit: type=1104 audit(1765889413.373:871): pid=5209 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=147.75.109.163 addr=147.75.109.163 terminal=ssh res=success' Dec 16 12:50:14.156977 kubelet[2796]: E1216 12:50:14.156854 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" podUID="f4651243-c791-471b-a364-a0f9a5a677b8" Dec 16 12:50:19.156113 kubelet[2796]: E1216 12:50:19.155584 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-5hslq" podUID="0e649fa6-ddd7-43a4-b265-cc0585bd2a9e" Dec 16 12:50:20.156379 kubelet[2796]: E1216 12:50:20.156230 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:50:21.155720 kubelet[2796]: E1216 12:50:21.155594 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" podUID="dd8550e8-a28d-4154-8fdc-fe7811fd3803" Dec 16 12:50:21.156146 kubelet[2796]: E1216 12:50:21.155721 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-98457786-pp6vg" podUID="6200e57e-82e7-4316-9456-f3714e3f66da" Dec 16 12:50:27.154287 kubelet[2796]: E1216 12:50:27.154235 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-f48cdf5c5-pnl6x" podUID="ed1cdd8a-3d38-40e0-80a5-f1884fc6e6a1" Dec 16 12:50:28.154924 kubelet[2796]: E1216 12:50:28.154834 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-xl9w8" podUID="f4651243-c791-471b-a364-a0f9a5a677b8" Dec 16 12:50:29.242664 systemd[1]: cri-containerd-a6add43b584d0e8ff6fb00f86c9e01071c4a920f2c24cc94c9d8965295505031.scope: Deactivated successfully. Dec 16 12:50:29.245558 systemd[1]: cri-containerd-a6add43b584d0e8ff6fb00f86c9e01071c4a920f2c24cc94c9d8965295505031.scope: Consumed 3.294s CPU time, 85.7M memory peak, 62.9M read from disk. Dec 16 12:50:29.252471 kernel: kauditd_printk_skb: 1 callbacks suppressed Dec 16 12:50:29.252564 kernel: audit: type=1334 audit(1765889429.246:873): prog-id=256 op=LOAD Dec 16 12:50:29.246000 audit: BPF prog-id=256 op=LOAD Dec 16 12:50:29.246000 audit: BPF prog-id=86 op=UNLOAD Dec 16 12:50:29.252000 audit: BPF prog-id=103 op=UNLOAD Dec 16 12:50:29.260089 kernel: audit: type=1334 audit(1765889429.246:874): prog-id=86 op=UNLOAD Dec 16 12:50:29.260126 kernel: audit: type=1334 audit(1765889429.252:875): prog-id=103 op=UNLOAD Dec 16 12:50:29.252000 audit: BPF prog-id=107 op=UNLOAD Dec 16 12:50:29.263370 kernel: audit: type=1334 audit(1765889429.252:876): prog-id=107 op=UNLOAD Dec 16 12:50:29.358489 containerd[1623]: time="2025-12-16T12:50:29.358415929Z" level=info msg="received container exit event container_id:\"a6add43b584d0e8ff6fb00f86c9e01071c4a920f2c24cc94c9d8965295505031\" id:\"a6add43b584d0e8ff6fb00f86c9e01071c4a920f2c24cc94c9d8965295505031\" pid:2632 exit_status:1 exited_at:{seconds:1765889429 nanos:268279058}" Dec 16 12:50:29.444106 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a6add43b584d0e8ff6fb00f86c9e01071c4a920f2c24cc94c9d8965295505031-rootfs.mount: Deactivated successfully. Dec 16 12:50:29.497000 audit: BPF prog-id=257 op=LOAD Dec 16 12:50:29.502733 containerd[1623]: time="2025-12-16T12:50:29.502036965Z" level=info msg="received container exit event container_id:\"3b63931a1b4c85b68b640f0164e2951f0c6e73d5afd99f62522a8b46d33182ed\" id:\"3b63931a1b4c85b68b640f0164e2951f0c6e73d5afd99f62522a8b46d33182ed\" pid:2653 exit_status:1 exited_at:{seconds:1765889429 nanos:501646451}" Dec 16 12:50:29.498092 systemd[1]: cri-containerd-3b63931a1b4c85b68b640f0164e2951f0c6e73d5afd99f62522a8b46d33182ed.scope: Deactivated successfully. Dec 16 12:50:29.498357 systemd[1]: cri-containerd-3b63931a1b4c85b68b640f0164e2951f0c6e73d5afd99f62522a8b46d33182ed.scope: Consumed 2.306s CPU time, 37.9M memory peak, 35.6M read from disk. Dec 16 12:50:29.506570 kernel: audit: type=1334 audit(1765889429.497:877): prog-id=257 op=LOAD Dec 16 12:50:29.507585 kernel: audit: type=1334 audit(1765889429.497:878): prog-id=93 op=UNLOAD Dec 16 12:50:29.497000 audit: BPF prog-id=93 op=UNLOAD Dec 16 12:50:29.506000 audit: BPF prog-id=108 op=UNLOAD Dec 16 12:50:29.513244 kernel: audit: type=1334 audit(1765889429.506:879): prog-id=108 op=UNLOAD Dec 16 12:50:29.513304 kernel: audit: type=1334 audit(1765889429.506:880): prog-id=112 op=UNLOAD Dec 16 12:50:29.506000 audit: BPF prog-id=112 op=UNLOAD Dec 16 12:50:29.541256 systemd[1]: cri-containerd-3cf9cfb2aba45a3e247a26c87c419524f478a8503621e1fc8baea6bcda29d055.scope: Deactivated successfully. Dec 16 12:50:29.541563 systemd[1]: cri-containerd-3cf9cfb2aba45a3e247a26c87c419524f478a8503621e1fc8baea6bcda29d055.scope: Consumed 17.268s CPU time, 117.5M memory peak, 41.8M read from disk. Dec 16 12:50:29.543959 containerd[1623]: time="2025-12-16T12:50:29.543891066Z" level=info msg="received container exit event container_id:\"3cf9cfb2aba45a3e247a26c87c419524f478a8503621e1fc8baea6bcda29d055\" id:\"3cf9cfb2aba45a3e247a26c87c419524f478a8503621e1fc8baea6bcda29d055\" pid:3110 exit_status:1 exited_at:{seconds:1765889429 nanos:541129489}" Dec 16 12:50:29.544000 audit: BPF prog-id=146 op=UNLOAD Dec 16 12:50:29.551324 kernel: audit: type=1334 audit(1765889429.544:881): prog-id=146 op=UNLOAD Dec 16 12:50:29.551388 kernel: audit: type=1334 audit(1765889429.544:882): prog-id=150 op=UNLOAD Dec 16 12:50:29.544000 audit: BPF prog-id=150 op=UNLOAD Dec 16 12:50:29.557972 kubelet[2796]: E1216 12:50:29.557942 2796 controller.go:195] "Failed to update lease" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:38158->10.0.0.2:2379: read: connection timed out" Dec 16 12:50:29.576260 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3b63931a1b4c85b68b640f0164e2951f0c6e73d5afd99f62522a8b46d33182ed-rootfs.mount: Deactivated successfully. Dec 16 12:50:29.589231 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3cf9cfb2aba45a3e247a26c87c419524f478a8503621e1fc8baea6bcda29d055-rootfs.mount: Deactivated successfully. Dec 16 12:50:29.618460 kubelet[2796]: I1216 12:50:29.618382 2796 status_manager.go:890] "Failed to get status for pod" podUID="0e649fa6-ddd7-43a4-b265-cc0585bd2a9e" pod="calico-system/goldmane-666569f655-5hslq" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:38074->10.0.0.2:2379: read: connection timed out" Dec 16 12:50:29.633279 kubelet[2796]: E1216 12:50:29.618855 2796 event.go:359] "Server rejected event (will not retry!)" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:37954->10.0.0.2:2379: read: connection timed out" event="&Event{ObjectMeta:{goldmane-666569f655-5hslq.1881b2f900fa4609 calico-system 1402 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:calico-system,Name:goldmane-666569f655-5hslq,UID:0e649fa6-ddd7-43a4-b265-cc0585bd2a9e,APIVersion:v1,ResourceVersion:778,FieldPath:spec.containers{goldmane},},Reason:BackOff,Message:Back-off pulling image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\",Source:EventSource{Component:kubelet,Host:ci-4547-0-0-6-b8319fed62,},FirstTimestamp:2025-12-16 12:48:31 +0000 UTC,LastTimestamp:2025-12-16 12:50:19.1555304 +0000 UTC m=+152.089258437,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4547-0-0-6-b8319fed62,}" Dec 16 12:50:29.765726 kubelet[2796]: I1216 12:50:29.765675 2796 scope.go:117] "RemoveContainer" containerID="3b63931a1b4c85b68b640f0164e2951f0c6e73d5afd99f62522a8b46d33182ed" Dec 16 12:50:29.768537 kubelet[2796]: I1216 12:50:29.768503 2796 scope.go:117] "RemoveContainer" containerID="3cf9cfb2aba45a3e247a26c87c419524f478a8503621e1fc8baea6bcda29d055" Dec 16 12:50:29.777959 containerd[1623]: time="2025-12-16T12:50:29.777667239Z" level=info msg="CreateContainer within sandbox \"a4b11037aaa136258d84d308b5125184ee3d601f2045debbce68a868e343d15a\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Dec 16 12:50:29.778086 kubelet[2796]: I1216 12:50:29.777924 2796 scope.go:117] "RemoveContainer" containerID="a6add43b584d0e8ff6fb00f86c9e01071c4a920f2c24cc94c9d8965295505031" Dec 16 12:50:29.778844 containerd[1623]: time="2025-12-16T12:50:29.778791327Z" level=info msg="CreateContainer within sandbox \"02ac242cef836574fb5d0643d85fa43d58200df961b34aef937073f2bcd94794\" for container &ContainerMetadata{Name:tigera-operator,Attempt:1,}" Dec 16 12:50:29.839343 containerd[1623]: time="2025-12-16T12:50:29.839296052Z" level=info msg="CreateContainer within sandbox \"572b02fe65e614ab942b1985a6f3b351518d0a7bc06d4764c65afa080e172397\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Dec 16 12:50:29.858327 containerd[1623]: time="2025-12-16T12:50:29.858271824Z" level=info msg="Container af693ca8963359e37d4ec9e1481bc6021209ace6434638e75ff14ead3126ea0b: CDI devices from CRI Config.CDIDevices: []" Dec 16 12:50:29.858891 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2191187222.mount: Deactivated successfully. Dec 16 12:50:29.861994 containerd[1623]: time="2025-12-16T12:50:29.861899791Z" level=info msg="Container d94409a32f7c0af4b1b182e6008d2363fa35c6c880bc75cb617a86949c3e1eb9: CDI devices from CRI Config.CDIDevices: []" Dec 16 12:50:29.891495 containerd[1623]: time="2025-12-16T12:50:29.891073103Z" level=info msg="Container bd1d08f6a8a21d21cbb9dc364137df5d89be2937b0d88980fbb69834bfd5a345: CDI devices from CRI Config.CDIDevices: []" Dec 16 12:50:29.918127 containerd[1623]: time="2025-12-16T12:50:29.918071409Z" level=info msg="CreateContainer within sandbox \"572b02fe65e614ab942b1985a6f3b351518d0a7bc06d4764c65afa080e172397\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"bd1d08f6a8a21d21cbb9dc364137df5d89be2937b0d88980fbb69834bfd5a345\"" Dec 16 12:50:29.921653 containerd[1623]: time="2025-12-16T12:50:29.921631616Z" level=info msg="StartContainer for \"bd1d08f6a8a21d21cbb9dc364137df5d89be2937b0d88980fbb69834bfd5a345\"" Dec 16 12:50:29.926256 containerd[1623]: time="2025-12-16T12:50:29.926031753Z" level=info msg="CreateContainer within sandbox \"a4b11037aaa136258d84d308b5125184ee3d601f2045debbce68a868e343d15a\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"af693ca8963359e37d4ec9e1481bc6021209ace6434638e75ff14ead3126ea0b\"" Dec 16 12:50:29.926256 containerd[1623]: time="2025-12-16T12:50:29.926129723Z" level=info msg="connecting to shim bd1d08f6a8a21d21cbb9dc364137df5d89be2937b0d88980fbb69834bfd5a345" address="unix:///run/containerd/s/3f5b2b7fb4c529f7a8b552685c156d93aca8f3164ee4e0635d9ffb58f5d23524" protocol=ttrpc version=3 Dec 16 12:50:29.928334 containerd[1623]: time="2025-12-16T12:50:29.927065217Z" level=info msg="CreateContainer within sandbox \"02ac242cef836574fb5d0643d85fa43d58200df961b34aef937073f2bcd94794\" for &ContainerMetadata{Name:tigera-operator,Attempt:1,} returns container id \"d94409a32f7c0af4b1b182e6008d2363fa35c6c880bc75cb617a86949c3e1eb9\"" Dec 16 12:50:29.928334 containerd[1623]: time="2025-12-16T12:50:29.927365416Z" level=info msg="StartContainer for \"af693ca8963359e37d4ec9e1481bc6021209ace6434638e75ff14ead3126ea0b\"" Dec 16 12:50:29.928539 containerd[1623]: time="2025-12-16T12:50:29.928523710Z" level=info msg="StartContainer for \"d94409a32f7c0af4b1b182e6008d2363fa35c6c880bc75cb617a86949c3e1eb9\"" Dec 16 12:50:29.928710 containerd[1623]: time="2025-12-16T12:50:29.928556573Z" level=info msg="connecting to shim af693ca8963359e37d4ec9e1481bc6021209ace6434638e75ff14ead3126ea0b" address="unix:///run/containerd/s/69557a881e3a77eba4a2735934dcb6011ee3355c49c3fae5455d922b75574085" protocol=ttrpc version=3 Dec 16 12:50:29.929268 containerd[1623]: time="2025-12-16T12:50:29.929249720Z" level=info msg="connecting to shim d94409a32f7c0af4b1b182e6008d2363fa35c6c880bc75cb617a86949c3e1eb9" address="unix:///run/containerd/s/d13c141a964ed7933bf8b9828d04ff61c824a1d935f623eb0bedd7e1afc4c513" protocol=ttrpc version=3 Dec 16 12:50:29.954212 systemd[1]: Started cri-containerd-af693ca8963359e37d4ec9e1481bc6021209ace6434638e75ff14ead3126ea0b.scope - libcontainer container af693ca8963359e37d4ec9e1481bc6021209ace6434638e75ff14ead3126ea0b. Dec 16 12:50:29.961182 systemd[1]: Started cri-containerd-d94409a32f7c0af4b1b182e6008d2363fa35c6c880bc75cb617a86949c3e1eb9.scope - libcontainer container d94409a32f7c0af4b1b182e6008d2363fa35c6c880bc75cb617a86949c3e1eb9. Dec 16 12:50:29.964245 systemd[1]: Started cri-containerd-bd1d08f6a8a21d21cbb9dc364137df5d89be2937b0d88980fbb69834bfd5a345.scope - libcontainer container bd1d08f6a8a21d21cbb9dc364137df5d89be2937b0d88980fbb69834bfd5a345. Dec 16 12:50:29.992000 audit: BPF prog-id=258 op=LOAD Dec 16 12:50:29.992000 audit: BPF prog-id=259 op=LOAD Dec 16 12:50:29.992000 audit[5296]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=2924 pid=5296 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:29.992000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6439343430396133326637633061663462316231383265363030386432 Dec 16 12:50:29.992000 audit: BPF prog-id=259 op=UNLOAD Dec 16 12:50:29.992000 audit[5296]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2924 pid=5296 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:29.992000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6439343430396133326637633061663462316231383265363030386432 Dec 16 12:50:29.992000 audit: BPF prog-id=260 op=LOAD Dec 16 12:50:29.992000 audit[5296]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=2924 pid=5296 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:29.992000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6439343430396133326637633061663462316231383265363030386432 Dec 16 12:50:29.992000 audit: BPF prog-id=261 op=LOAD Dec 16 12:50:29.992000 audit[5296]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=2924 pid=5296 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:29.992000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6439343430396133326637633061663462316231383265363030386432 Dec 16 12:50:29.992000 audit: BPF prog-id=261 op=UNLOAD Dec 16 12:50:29.992000 audit[5296]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2924 pid=5296 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:29.992000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6439343430396133326637633061663462316231383265363030386432 Dec 16 12:50:29.992000 audit: BPF prog-id=260 op=UNLOAD Dec 16 12:50:29.992000 audit[5296]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2924 pid=5296 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:29.992000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6439343430396133326637633061663462316231383265363030386432 Dec 16 12:50:29.992000 audit: BPF prog-id=262 op=LOAD Dec 16 12:50:29.992000 audit[5296]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=2924 pid=5296 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:29.992000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6439343430396133326637633061663462316231383265363030386432 Dec 16 12:50:30.002000 audit: BPF prog-id=263 op=LOAD Dec 16 12:50:30.004000 audit: BPF prog-id=264 op=LOAD Dec 16 12:50:30.004000 audit[5294]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b0238 a2=98 a3=0 items=0 ppid=2477 pid=5294 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:30.004000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6264316430386636613861323164323163626239646333363431333764 Dec 16 12:50:30.004000 audit: BPF prog-id=264 op=UNLOAD Dec 16 12:50:30.004000 audit[5294]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2477 pid=5294 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:30.004000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6264316430386636613861323164323163626239646333363431333764 Dec 16 12:50:30.004000 audit: BPF prog-id=265 op=LOAD Dec 16 12:50:30.004000 audit[5294]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b0488 a2=98 a3=0 items=0 ppid=2477 pid=5294 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:30.004000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6264316430386636613861323164323163626239646333363431333764 Dec 16 12:50:30.004000 audit: BPF prog-id=266 op=LOAD Dec 16 12:50:30.004000 audit[5294]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001b0218 a2=98 a3=0 items=0 ppid=2477 pid=5294 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:30.004000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6264316430386636613861323164323163626239646333363431333764 Dec 16 12:50:30.004000 audit: BPF prog-id=266 op=UNLOAD Dec 16 12:50:30.004000 audit[5294]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2477 pid=5294 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:30.004000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6264316430386636613861323164323163626239646333363431333764 Dec 16 12:50:30.004000 audit: BPF prog-id=265 op=UNLOAD Dec 16 12:50:30.004000 audit[5294]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2477 pid=5294 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:30.004000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6264316430386636613861323164323163626239646333363431333764 Dec 16 12:50:30.004000 audit: BPF prog-id=267 op=LOAD Dec 16 12:50:30.004000 audit[5294]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b06e8 a2=98 a3=0 items=0 ppid=2477 pid=5294 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:30.004000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6264316430386636613861323164323163626239646333363431333764 Dec 16 12:50:30.005000 audit: BPF prog-id=268 op=LOAD Dec 16 12:50:30.008000 audit: BPF prog-id=269 op=LOAD Dec 16 12:50:30.008000 audit[5295]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b0238 a2=98 a3=0 items=0 ppid=2492 pid=5295 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:30.008000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166363933636138393633333539653337643465633965313438316263 Dec 16 12:50:30.008000 audit: BPF prog-id=269 op=UNLOAD Dec 16 12:50:30.008000 audit[5295]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2492 pid=5295 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:30.008000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166363933636138393633333539653337643465633965313438316263 Dec 16 12:50:30.008000 audit: BPF prog-id=270 op=LOAD Dec 16 12:50:30.008000 audit[5295]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b0488 a2=98 a3=0 items=0 ppid=2492 pid=5295 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:30.008000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166363933636138393633333539653337643465633965313438316263 Dec 16 12:50:30.009000 audit: BPF prog-id=271 op=LOAD Dec 16 12:50:30.009000 audit[5295]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001b0218 a2=98 a3=0 items=0 ppid=2492 pid=5295 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:30.009000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166363933636138393633333539653337643465633965313438316263 Dec 16 12:50:30.011000 audit: BPF prog-id=271 op=UNLOAD Dec 16 12:50:30.011000 audit[5295]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2492 pid=5295 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:30.011000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166363933636138393633333539653337643465633965313438316263 Dec 16 12:50:30.011000 audit: BPF prog-id=270 op=UNLOAD Dec 16 12:50:30.011000 audit[5295]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2492 pid=5295 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:30.011000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166363933636138393633333539653337643465633965313438316263 Dec 16 12:50:30.011000 audit: BPF prog-id=272 op=LOAD Dec 16 12:50:30.011000 audit[5295]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001b06e8 a2=98 a3=0 items=0 ppid=2492 pid=5295 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 16 12:50:30.011000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6166363933636138393633333539653337643465633965313438316263 Dec 16 12:50:30.023090 containerd[1623]: time="2025-12-16T12:50:30.022271019Z" level=info msg="StartContainer for \"d94409a32f7c0af4b1b182e6008d2363fa35c6c880bc75cb617a86949c3e1eb9\" returns successfully" Dec 16 12:50:30.090418 containerd[1623]: time="2025-12-16T12:50:30.090281794Z" level=info msg="StartContainer for \"af693ca8963359e37d4ec9e1481bc6021209ace6434638e75ff14ead3126ea0b\" returns successfully" Dec 16 12:50:30.096564 containerd[1623]: time="2025-12-16T12:50:30.096521003Z" level=info msg="StartContainer for \"bd1d08f6a8a21d21cbb9dc364137df5d89be2937b0d88980fbb69834bfd5a345\" returns successfully" Dec 16 12:50:30.448187 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2864807149.mount: Deactivated successfully. Dec 16 12:50:31.157729 kubelet[2796]: E1216 12:50:31.157589 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-m8p7d" podUID="2053803c-11b9-48f6-949c-d0d885185bfe" Dec 16 12:50:34.155363 kubelet[2796]: E1216 12:50:34.155296 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5668559856-lv4tb" podUID="dd8550e8-a28d-4154-8fdc-fe7811fd3803" Dec 16 12:50:34.155730 kubelet[2796]: E1216 12:50:34.155437 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-5hslq" podUID="0e649fa6-ddd7-43a4-b265-cc0585bd2a9e" Dec 16 12:50:34.155730 kubelet[2796]: E1216 12:50:34.155515 2796 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-98457786-pp6vg" podUID="6200e57e-82e7-4316-9456-f3714e3f66da"