Jan 14 00:46:44.174930 kernel: Linux version 6.12.65-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.1_p20250801 p4) 14.3.1 20250801, GNU ld (Gentoo 2.45 p3) 2.45.0) #1 SMP PREEMPT_DYNAMIC Tue Jan 13 22:15:29 -00 2026 Jan 14 00:46:44.174959 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=akamai verity.usrhash=6d34ab71a3dc5a0ab37eb2c851228af18a1e24f648223df9a1099dbd7db2cfcf Jan 14 00:46:44.174969 kernel: BIOS-provided physical RAM map: Jan 14 00:46:44.174976 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009f7ff] usable Jan 14 00:46:44.174982 kernel: BIOS-e820: [mem 0x000000000009f800-0x000000000009ffff] reserved Jan 14 00:46:44.174989 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Jan 14 00:46:44.175055 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdcfff] usable Jan 14 00:46:44.175062 kernel: BIOS-e820: [mem 0x000000007ffdd000-0x000000007fffffff] reserved Jan 14 00:46:44.175069 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Jan 14 00:46:44.175075 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Jan 14 00:46:44.175082 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Jan 14 00:46:44.175089 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Jan 14 00:46:44.175096 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000017fffffff] usable Jan 14 00:46:44.175102 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Jan 14 00:46:44.175128 kernel: NX (Execute Disable) protection: active Jan 14 00:46:44.175135 kernel: APIC: Static calls initialized Jan 14 00:46:44.175142 kernel: SMBIOS 2.8 present. Jan 14 00:46:44.175150 kernel: DMI: Linode Compute Instance/Standard PC (Q35 + ICH9, 2009), BIOS Not Specified Jan 14 00:46:44.175157 kernel: DMI: Memory slots populated: 1/1 Jan 14 00:46:44.175166 kernel: Hypervisor detected: KVM Jan 14 00:46:44.175173 kernel: last_pfn = 0x7ffdd max_arch_pfn = 0x400000000 Jan 14 00:46:44.175180 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Jan 14 00:46:44.175187 kernel: kvm-clock: using sched offset of 6043388566 cycles Jan 14 00:46:44.175199 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Jan 14 00:46:44.175211 kernel: tsc: Detected 2000.002 MHz processor Jan 14 00:46:44.175219 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jan 14 00:46:44.175227 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jan 14 00:46:44.175234 kernel: last_pfn = 0x180000 max_arch_pfn = 0x400000000 Jan 14 00:46:44.175244 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Jan 14 00:46:44.175252 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jan 14 00:46:44.175259 kernel: last_pfn = 0x7ffdd max_arch_pfn = 0x400000000 Jan 14 00:46:44.175267 kernel: Using GB pages for direct mapping Jan 14 00:46:44.175274 kernel: ACPI: Early table checksum verification disabled Jan 14 00:46:44.175282 kernel: ACPI: RSDP 0x00000000000F5160 000014 (v00 BOCHS ) Jan 14 00:46:44.175290 kernel: ACPI: RSDT 0x000000007FFE2307 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 14 00:46:44.180088 kernel: ACPI: FACP 0x000000007FFE20F7 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Jan 14 00:46:44.180102 kernel: ACPI: DSDT 0x000000007FFE0040 0020B7 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 14 00:46:44.180126 kernel: ACPI: FACS 0x000000007FFE0000 000040 Jan 14 00:46:44.180134 kernel: ACPI: APIC 0x000000007FFE21EB 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 14 00:46:44.180142 kernel: ACPI: HPET 0x000000007FFE226B 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 14 00:46:44.180155 kernel: ACPI: MCFG 0x000000007FFE22A3 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 14 00:46:44.180165 kernel: ACPI: WAET 0x000000007FFE22DF 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 14 00:46:44.180176 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe20f7-0x7ffe21ea] Jan 14 00:46:44.180189 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe20f6] Jan 14 00:46:44.180202 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Jan 14 00:46:44.180214 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe21eb-0x7ffe226a] Jan 14 00:46:44.180227 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe226b-0x7ffe22a2] Jan 14 00:46:44.180235 kernel: ACPI: Reserving MCFG table memory at [mem 0x7ffe22a3-0x7ffe22de] Jan 14 00:46:44.180243 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe22df-0x7ffe2306] Jan 14 00:46:44.180250 kernel: No NUMA configuration found Jan 14 00:46:44.180258 kernel: Faking a node at [mem 0x0000000000000000-0x000000017fffffff] Jan 14 00:46:44.180266 kernel: NODE_DATA(0) allocated [mem 0x17fff6dc0-0x17fffdfff] Jan 14 00:46:44.180274 kernel: Zone ranges: Jan 14 00:46:44.180282 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jan 14 00:46:44.180291 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Jan 14 00:46:44.180299 kernel: Normal [mem 0x0000000100000000-0x000000017fffffff] Jan 14 00:46:44.180307 kernel: Device empty Jan 14 00:46:44.180315 kernel: Movable zone start for each node Jan 14 00:46:44.180322 kernel: Early memory node ranges Jan 14 00:46:44.180330 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Jan 14 00:46:44.180338 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdcfff] Jan 14 00:46:44.180348 kernel: node 0: [mem 0x0000000100000000-0x000000017fffffff] Jan 14 00:46:44.180356 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000017fffffff] Jan 14 00:46:44.180363 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jan 14 00:46:44.180371 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Jan 14 00:46:44.180379 kernel: On node 0, zone Normal: 35 pages in unavailable ranges Jan 14 00:46:44.180387 kernel: ACPI: PM-Timer IO Port: 0x608 Jan 14 00:46:44.180395 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Jan 14 00:46:44.180403 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Jan 14 00:46:44.180413 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Jan 14 00:46:44.180421 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Jan 14 00:46:44.180428 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Jan 14 00:46:44.180436 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Jan 14 00:46:44.180444 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Jan 14 00:46:44.180452 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jan 14 00:46:44.180459 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Jan 14 00:46:44.180469 kernel: TSC deadline timer available Jan 14 00:46:44.180477 kernel: CPU topo: Max. logical packages: 1 Jan 14 00:46:44.180485 kernel: CPU topo: Max. logical dies: 1 Jan 14 00:46:44.180492 kernel: CPU topo: Max. dies per package: 1 Jan 14 00:46:44.180500 kernel: CPU topo: Max. threads per core: 1 Jan 14 00:46:44.180507 kernel: CPU topo: Num. cores per package: 2 Jan 14 00:46:44.180515 kernel: CPU topo: Num. threads per package: 2 Jan 14 00:46:44.181012 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs Jan 14 00:46:44.181023 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Jan 14 00:46:44.181031 kernel: kvm-guest: KVM setup pv remote TLB flush Jan 14 00:46:44.181039 kernel: kvm-guest: setup PV sched yield Jan 14 00:46:44.181047 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Jan 14 00:46:44.181055 kernel: Booting paravirtualized kernel on KVM Jan 14 00:46:44.181063 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jan 14 00:46:44.181071 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Jan 14 00:46:44.181081 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u1048576 Jan 14 00:46:44.181088 kernel: pcpu-alloc: s207832 r8192 d29736 u1048576 alloc=1*2097152 Jan 14 00:46:44.181096 kernel: pcpu-alloc: [0] 0 1 Jan 14 00:46:44.181104 kernel: kvm-guest: PV spinlocks enabled Jan 14 00:46:44.181131 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Jan 14 00:46:44.181141 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=akamai verity.usrhash=6d34ab71a3dc5a0ab37eb2c851228af18a1e24f648223df9a1099dbd7db2cfcf Jan 14 00:46:44.181152 kernel: random: crng init done Jan 14 00:46:44.181160 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jan 14 00:46:44.181168 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 14 00:46:44.181175 kernel: Fallback order for Node 0: 0 Jan 14 00:46:44.181183 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1048443 Jan 14 00:46:44.181191 kernel: Policy zone: Normal Jan 14 00:46:44.181199 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 14 00:46:44.181209 kernel: software IO TLB: area num 2. Jan 14 00:46:44.181216 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Jan 14 00:46:44.181224 kernel: ftrace: allocating 40097 entries in 157 pages Jan 14 00:46:44.181232 kernel: ftrace: allocated 157 pages with 5 groups Jan 14 00:46:44.181240 kernel: Dynamic Preempt: voluntary Jan 14 00:46:44.181248 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 14 00:46:44.181256 kernel: rcu: RCU event tracing is enabled. Jan 14 00:46:44.181265 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Jan 14 00:46:44.181275 kernel: Trampoline variant of Tasks RCU enabled. Jan 14 00:46:44.181283 kernel: Rude variant of Tasks RCU enabled. Jan 14 00:46:44.181290 kernel: Tracing variant of Tasks RCU enabled. Jan 14 00:46:44.181298 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 14 00:46:44.181306 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Jan 14 00:46:44.181314 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 14 00:46:44.181331 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 14 00:46:44.181339 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 14 00:46:44.181348 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Jan 14 00:46:44.181358 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jan 14 00:46:44.181366 kernel: Console: colour VGA+ 80x25 Jan 14 00:46:44.181374 kernel: printk: legacy console [tty0] enabled Jan 14 00:46:44.181382 kernel: printk: legacy console [ttyS0] enabled Jan 14 00:46:44.181390 kernel: ACPI: Core revision 20240827 Jan 14 00:46:44.181401 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Jan 14 00:46:44.181409 kernel: APIC: Switch to symmetric I/O mode setup Jan 14 00:46:44.181417 kernel: x2apic enabled Jan 14 00:46:44.181425 kernel: APIC: Switched APIC routing to: physical x2apic Jan 14 00:46:44.181433 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Jan 14 00:46:44.181442 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Jan 14 00:46:44.181452 kernel: kvm-guest: setup PV IPIs Jan 14 00:46:44.181460 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jan 14 00:46:44.181468 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x1cd42fed8cc, max_idle_ns: 440795202126 ns Jan 14 00:46:44.181476 kernel: Calibrating delay loop (skipped) preset value.. 4000.00 BogoMIPS (lpj=2000002) Jan 14 00:46:44.181485 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Jan 14 00:46:44.181493 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Jan 14 00:46:44.181501 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Jan 14 00:46:44.181511 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jan 14 00:46:44.181520 kernel: Spectre V2 : Mitigation: Retpolines Jan 14 00:46:44.181528 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Jan 14 00:46:44.181536 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Jan 14 00:46:44.181544 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Jan 14 00:46:44.181553 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Jan 14 00:46:44.181561 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Jan 14 00:46:44.181572 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Jan 14 00:46:44.181580 kernel: active return thunk: srso_alias_return_thunk Jan 14 00:46:44.181589 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Jan 14 00:46:44.181597 kernel: Transient Scheduler Attacks: Forcing mitigation on in a VM Jan 14 00:46:44.181605 kernel: Transient Scheduler Attacks: Vulnerable: Clear CPU buffers attempted, no microcode Jan 14 00:46:44.181613 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jan 14 00:46:44.181621 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jan 14 00:46:44.181644 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jan 14 00:46:44.181652 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Jan 14 00:46:44.181661 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jan 14 00:46:44.181669 kernel: x86/fpu: xstate_offset[9]: 832, xstate_sizes[9]: 8 Jan 14 00:46:44.181677 kernel: x86/fpu: Enabled xstate features 0x207, context size is 840 bytes, using 'compacted' format. Jan 14 00:46:44.181685 kernel: Freeing SMP alternatives memory: 32K Jan 14 00:46:44.181692 kernel: pid_max: default: 32768 minimum: 301 Jan 14 00:46:44.181702 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Jan 14 00:46:44.181710 kernel: landlock: Up and running. Jan 14 00:46:44.181718 kernel: SELinux: Initializing. Jan 14 00:46:44.181725 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 14 00:46:44.181733 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 14 00:46:44.181741 kernel: smpboot: CPU0: AMD EPYC 7713 64-Core Processor (family: 0x19, model: 0x1, stepping: 0x1) Jan 14 00:46:44.181749 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Jan 14 00:46:44.181759 kernel: ... version: 0 Jan 14 00:46:44.181766 kernel: ... bit width: 48 Jan 14 00:46:44.181774 kernel: ... generic registers: 6 Jan 14 00:46:44.181782 kernel: ... value mask: 0000ffffffffffff Jan 14 00:46:44.181955 kernel: ... max period: 00007fffffffffff Jan 14 00:46:44.181963 kernel: ... fixed-purpose events: 0 Jan 14 00:46:44.181971 kernel: ... event mask: 000000000000003f Jan 14 00:46:44.181980 kernel: signal: max sigframe size: 3376 Jan 14 00:46:44.181988 kernel: rcu: Hierarchical SRCU implementation. Jan 14 00:46:44.181996 kernel: rcu: Max phase no-delay instances is 400. Jan 14 00:46:44.182004 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Jan 14 00:46:44.182012 kernel: smp: Bringing up secondary CPUs ... Jan 14 00:46:44.182019 kernel: smpboot: x86: Booting SMP configuration: Jan 14 00:46:44.182027 kernel: .... node #0, CPUs: #1 Jan 14 00:46:44.182037 kernel: smp: Brought up 1 node, 2 CPUs Jan 14 00:46:44.182044 kernel: smpboot: Total of 2 processors activated (8000.00 BogoMIPS) Jan 14 00:46:44.182052 kernel: Memory: 3977432K/4193772K available (14336K kernel code, 2445K rwdata, 31636K rodata, 15536K init, 2504K bss, 210912K reserved, 0K cma-reserved) Jan 14 00:46:44.182060 kernel: devtmpfs: initialized Jan 14 00:46:44.182068 kernel: x86/mm: Memory block size: 128MB Jan 14 00:46:44.182076 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 14 00:46:44.182084 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Jan 14 00:46:44.182094 kernel: pinctrl core: initialized pinctrl subsystem Jan 14 00:46:44.182101 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 14 00:46:44.191141 kernel: audit: initializing netlink subsys (disabled) Jan 14 00:46:44.191159 kernel: audit: type=2000 audit(1768351600.613:1): state=initialized audit_enabled=0 res=1 Jan 14 00:46:44.191170 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 14 00:46:44.191179 kernel: thermal_sys: Registered thermal governor 'user_space' Jan 14 00:46:44.191188 kernel: cpuidle: using governor menu Jan 14 00:46:44.191197 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 14 00:46:44.191215 kernel: dca service started, version 1.12.1 Jan 14 00:46:44.191225 kernel: PCI: ECAM [mem 0xb0000000-0xbfffffff] (base 0xb0000000) for domain 0000 [bus 00-ff] Jan 14 00:46:44.191234 kernel: PCI: ECAM [mem 0xb0000000-0xbfffffff] reserved as E820 entry Jan 14 00:46:44.191243 kernel: PCI: Using configuration type 1 for base access Jan 14 00:46:44.191252 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jan 14 00:46:44.191260 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jan 14 00:46:44.191269 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jan 14 00:46:44.191280 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 14 00:46:44.191289 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jan 14 00:46:44.191297 kernel: ACPI: Added _OSI(Module Device) Jan 14 00:46:44.191306 kernel: ACPI: Added _OSI(Processor Device) Jan 14 00:46:44.191314 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 14 00:46:44.191323 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 14 00:46:44.191332 kernel: ACPI: Interpreter enabled Jan 14 00:46:44.191343 kernel: ACPI: PM: (supports S0 S3 S5) Jan 14 00:46:44.191351 kernel: ACPI: Using IOAPIC for interrupt routing Jan 14 00:46:44.191360 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jan 14 00:46:44.191369 kernel: PCI: Using E820 reservations for host bridge windows Jan 14 00:46:44.191378 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Jan 14 00:46:44.191386 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jan 14 00:46:44.191686 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jan 14 00:46:44.191886 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Jan 14 00:46:44.192281 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Jan 14 00:46:44.192295 kernel: PCI host bridge to bus 0000:00 Jan 14 00:46:44.192477 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jan 14 00:46:44.192643 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Jan 14 00:46:44.192813 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jan 14 00:46:44.192975 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xafffffff window] Jan 14 00:46:44.193563 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Jan 14 00:46:44.193745 kernel: pci_bus 0000:00: root bus resource [mem 0x180000000-0x97fffffff window] Jan 14 00:46:44.193913 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jan 14 00:46:44.194138 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint Jan 14 00:46:44.194343 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint Jan 14 00:46:44.194527 kernel: pci 0000:00:01.0: BAR 0 [mem 0xfd000000-0xfdffffff pref] Jan 14 00:46:44.194705 kernel: pci 0000:00:01.0: BAR 2 [mem 0xfebd0000-0xfebd0fff] Jan 14 00:46:44.194885 kernel: pci 0000:00:01.0: ROM [mem 0xfebc0000-0xfebcffff pref] Jan 14 00:46:44.195061 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jan 14 00:46:44.195309 kernel: pci 0000:00:02.0: [1af4:1004] type 00 class 0x010000 conventional PCI endpoint Jan 14 00:46:44.195492 kernel: pci 0000:00:02.0: BAR 0 [io 0xc000-0xc03f] Jan 14 00:46:44.195669 kernel: pci 0000:00:02.0: BAR 1 [mem 0xfebd1000-0xfebd1fff] Jan 14 00:46:44.195845 kernel: pci 0000:00:02.0: BAR 4 [mem 0xfe000000-0xfe003fff 64bit pref] Jan 14 00:46:44.196030 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint Jan 14 00:46:44.196231 kernel: pci 0000:00:03.0: BAR 0 [io 0xc040-0xc07f] Jan 14 00:46:44.196416 kernel: pci 0000:00:03.0: BAR 1 [mem 0xfebd2000-0xfebd2fff] Jan 14 00:46:44.196593 kernel: pci 0000:00:03.0: BAR 4 [mem 0xfe004000-0xfe007fff 64bit pref] Jan 14 00:46:44.196769 kernel: pci 0000:00:03.0: ROM [mem 0xfeb80000-0xfebbffff pref] Jan 14 00:46:44.196954 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint Jan 14 00:46:44.197460 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Jan 14 00:46:44.199282 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint Jan 14 00:46:44.199471 kernel: pci 0000:00:1f.2: BAR 4 [io 0xc0c0-0xc0df] Jan 14 00:46:44.199650 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xfebd3000-0xfebd3fff] Jan 14 00:46:44.199836 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint Jan 14 00:46:44.200013 kernel: pci 0000:00:1f.3: BAR 4 [io 0x0700-0x073f] Jan 14 00:46:44.200024 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Jan 14 00:46:44.200038 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Jan 14 00:46:44.200046 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jan 14 00:46:44.200055 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Jan 14 00:46:44.200063 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Jan 14 00:46:44.200072 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Jan 14 00:46:44.200081 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Jan 14 00:46:44.200089 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Jan 14 00:46:44.200100 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Jan 14 00:46:44.200125 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Jan 14 00:46:44.200133 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Jan 14 00:46:44.200142 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Jan 14 00:46:44.200150 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Jan 14 00:46:44.200159 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Jan 14 00:46:44.200167 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Jan 14 00:46:44.200179 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Jan 14 00:46:44.200187 kernel: iommu: Default domain type: Translated Jan 14 00:46:44.200195 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jan 14 00:46:44.200203 kernel: PCI: Using ACPI for IRQ routing Jan 14 00:46:44.200212 kernel: PCI: pci_cache_line_size set to 64 bytes Jan 14 00:46:44.200221 kernel: e820: reserve RAM buffer [mem 0x0009f800-0x0009ffff] Jan 14 00:46:44.200230 kernel: e820: reserve RAM buffer [mem 0x7ffdd000-0x7fffffff] Jan 14 00:46:44.200414 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Jan 14 00:46:44.200589 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Jan 14 00:46:44.200763 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jan 14 00:46:44.200774 kernel: vgaarb: loaded Jan 14 00:46:44.200782 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Jan 14 00:46:44.200791 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Jan 14 00:46:44.200799 kernel: clocksource: Switched to clocksource kvm-clock Jan 14 00:46:44.200811 kernel: VFS: Disk quotas dquot_6.6.0 Jan 14 00:46:44.200820 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 14 00:46:44.200828 kernel: pnp: PnP ACPI init Jan 14 00:46:44.201017 kernel: system 00:04: [mem 0xb0000000-0xbfffffff window] has been reserved Jan 14 00:46:44.201030 kernel: pnp: PnP ACPI: found 5 devices Jan 14 00:46:44.201038 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jan 14 00:46:44.201785 kernel: NET: Registered PF_INET protocol family Jan 14 00:46:44.201799 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jan 14 00:46:44.201809 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jan 14 00:46:44.201818 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 14 00:46:44.201826 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 14 00:46:44.201834 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jan 14 00:46:44.201843 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jan 14 00:46:44.201855 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 14 00:46:44.201864 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 14 00:46:44.201872 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 14 00:46:44.201881 kernel: NET: Registered PF_XDP protocol family Jan 14 00:46:44.202064 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Jan 14 00:46:44.202255 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Jan 14 00:46:44.202422 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Jan 14 00:46:44.202590 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xafffffff window] Jan 14 00:46:44.202752 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Jan 14 00:46:44.202915 kernel: pci_bus 0000:00: resource 9 [mem 0x180000000-0x97fffffff window] Jan 14 00:46:44.202926 kernel: PCI: CLS 0 bytes, default 64 Jan 14 00:46:44.202934 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Jan 14 00:46:44.202943 kernel: software IO TLB: mapped [mem 0x000000007bfdd000-0x000000007ffdd000] (64MB) Jan 14 00:46:44.202952 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x1cd42fed8cc, max_idle_ns: 440795202126 ns Jan 14 00:46:44.202963 kernel: Initialise system trusted keyrings Jan 14 00:46:44.202972 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jan 14 00:46:44.202980 kernel: Key type asymmetric registered Jan 14 00:46:44.202989 kernel: Asymmetric key parser 'x509' registered Jan 14 00:46:44.202997 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Jan 14 00:46:44.203005 kernel: io scheduler mq-deadline registered Jan 14 00:46:44.203013 kernel: io scheduler kyber registered Jan 14 00:46:44.203024 kernel: io scheduler bfq registered Jan 14 00:46:44.203032 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jan 14 00:46:44.203042 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Jan 14 00:46:44.203050 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Jan 14 00:46:44.203059 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 14 00:46:44.203068 kernel: 00:02: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jan 14 00:46:44.203077 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Jan 14 00:46:44.203087 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jan 14 00:46:44.203095 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jan 14 00:46:44.203151 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jan 14 00:46:44.205411 kernel: rtc_cmos 00:03: RTC can wake from S4 Jan 14 00:46:44.205592 kernel: rtc_cmos 00:03: registered as rtc0 Jan 14 00:46:44.205763 kernel: rtc_cmos 00:03: setting system clock to 2026-01-14T00:46:42 UTC (1768351602) Jan 14 00:46:44.205940 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Jan 14 00:46:44.205951 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Jan 14 00:46:44.205961 kernel: NET: Registered PF_INET6 protocol family Jan 14 00:46:44.205969 kernel: Segment Routing with IPv6 Jan 14 00:46:44.205978 kernel: In-situ OAM (IOAM) with IPv6 Jan 14 00:46:44.205986 kernel: NET: Registered PF_PACKET protocol family Jan 14 00:46:44.205995 kernel: Key type dns_resolver registered Jan 14 00:46:44.206003 kernel: IPI shorthand broadcast: enabled Jan 14 00:46:44.206015 kernel: sched_clock: Marking stable (1818007103, 327417829)->(2224257836, -78832904) Jan 14 00:46:44.206023 kernel: registered taskstats version 1 Jan 14 00:46:44.206031 kernel: Loading compiled-in X.509 certificates Jan 14 00:46:44.206040 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.65-flatcar: 58a78462583b088d099087e6f2d97e37d80e06bb' Jan 14 00:46:44.206048 kernel: Demotion targets for Node 0: null Jan 14 00:46:44.206056 kernel: Key type .fscrypt registered Jan 14 00:46:44.206064 kernel: Key type fscrypt-provisioning registered Jan 14 00:46:44.206075 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 14 00:46:44.206083 kernel: ima: Allocated hash algorithm: sha1 Jan 14 00:46:44.206092 kernel: ima: No architecture policies found Jan 14 00:46:44.206100 kernel: clk: Disabling unused clocks Jan 14 00:46:44.207296 kernel: Freeing unused kernel image (initmem) memory: 15536K Jan 14 00:46:44.207329 kernel: Write protecting the kernel read-only data: 47104k Jan 14 00:46:44.207339 kernel: Freeing unused kernel image (rodata/data gap) memory: 1132K Jan 14 00:46:44.208148 kernel: Run /init as init process Jan 14 00:46:44.208162 kernel: with arguments: Jan 14 00:46:44.208171 kernel: /init Jan 14 00:46:44.208180 kernel: with environment: Jan 14 00:46:44.208188 kernel: HOME=/ Jan 14 00:46:44.208213 kernel: TERM=linux Jan 14 00:46:44.208224 kernel: SCSI subsystem initialized Jan 14 00:46:44.208235 kernel: libata version 3.00 loaded. Jan 14 00:46:44.208432 kernel: ahci 0000:00:1f.2: version 3.0 Jan 14 00:46:44.208446 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Jan 14 00:46:44.208626 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode Jan 14 00:46:44.208806 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) Jan 14 00:46:44.208986 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Jan 14 00:46:44.211965 kernel: scsi host0: ahci Jan 14 00:46:44.212206 kernel: scsi host1: ahci Jan 14 00:46:44.212404 kernel: scsi host2: ahci Jan 14 00:46:44.212596 kernel: scsi host3: ahci Jan 14 00:46:44.212786 kernel: scsi host4: ahci Jan 14 00:46:44.212983 kernel: scsi host5: ahci Jan 14 00:46:44.212997 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3100 irq 24 lpm-pol 1 Jan 14 00:46:44.213006 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3180 irq 24 lpm-pol 1 Jan 14 00:46:44.213016 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3200 irq 24 lpm-pol 1 Jan 14 00:46:44.213025 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3280 irq 24 lpm-pol 1 Jan 14 00:46:44.213035 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3300 irq 24 lpm-pol 1 Jan 14 00:46:44.213044 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3380 irq 24 lpm-pol 1 Jan 14 00:46:44.213056 kernel: ata4: SATA link down (SStatus 0 SControl 300) Jan 14 00:46:44.213065 kernel: ata3: SATA link down (SStatus 0 SControl 300) Jan 14 00:46:44.213074 kernel: ata2: SATA link down (SStatus 0 SControl 300) Jan 14 00:46:44.213083 kernel: ata1: SATA link down (SStatus 0 SControl 300) Jan 14 00:46:44.213092 kernel: ata6: SATA link down (SStatus 0 SControl 300) Jan 14 00:46:44.213101 kernel: ata5: SATA link down (SStatus 0 SControl 300) Jan 14 00:46:44.213322 kernel: virtio_scsi virtio0: 2/0/0 default/read/poll queues Jan 14 00:46:44.213520 kernel: scsi host6: Virtio SCSI HBA Jan 14 00:46:44.213732 kernel: scsi 6:0:0:0: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Jan 14 00:46:44.213935 kernel: sd 6:0:0:0: Power-on or device reset occurred Jan 14 00:46:44.214188 kernel: sd 6:0:0:0: [sda] 167739392 512-byte logical blocks: (85.9 GB/80.0 GiB) Jan 14 00:46:44.214392 kernel: sd 6:0:0:0: [sda] Write Protect is off Jan 14 00:46:44.214593 kernel: sd 6:0:0:0: [sda] Mode Sense: 63 00 00 08 Jan 14 00:46:44.214787 kernel: sd 6:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Jan 14 00:46:44.214799 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jan 14 00:46:44.214810 kernel: GPT:25804799 != 167739391 Jan 14 00:46:44.214819 kernel: GPT:Alternate GPT header not at the end of the disk. Jan 14 00:46:44.214827 kernel: GPT:25804799 != 167739391 Jan 14 00:46:44.214836 kernel: GPT: Use GNU Parted to correct GPT errors. Jan 14 00:46:44.214846 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 14 00:46:44.215038 kernel: sd 6:0:0:0: [sda] Attached SCSI disk Jan 14 00:46:44.215050 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 14 00:46:44.215058 kernel: device-mapper: uevent: version 1.0.3 Jan 14 00:46:44.215067 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Jan 14 00:46:44.215076 kernel: device-mapper: verity: sha256 using shash "sha256-generic" Jan 14 00:46:44.215087 kernel: raid6: avx2x4 gen() 31258 MB/s Jan 14 00:46:44.215098 kernel: raid6: avx2x2 gen() 29797 MB/s Jan 14 00:46:44.215212 kernel: raid6: avx2x1 gen() 22050 MB/s Jan 14 00:46:44.215224 kernel: raid6: using algorithm avx2x4 gen() 31258 MB/s Jan 14 00:46:44.215232 kernel: raid6: .... xor() 3235 MB/s, rmw enabled Jan 14 00:46:44.215244 kernel: raid6: using avx2x2 recovery algorithm Jan 14 00:46:44.215253 kernel: xor: automatically using best checksumming function avx Jan 14 00:46:44.215262 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 14 00:46:44.215270 kernel: BTRFS: device fsid 315c4ba2-2b68-4ff5-9a58-ddeab520c9ac devid 1 transid 33 /dev/mapper/usr (254:0) scanned by mount (167) Jan 14 00:46:44.215279 kernel: BTRFS info (device dm-0): first mount of filesystem 315c4ba2-2b68-4ff5-9a58-ddeab520c9ac Jan 14 00:46:44.215288 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jan 14 00:46:44.215297 kernel: BTRFS info (device dm-0): enabling ssd optimizations Jan 14 00:46:44.215309 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 14 00:46:44.215318 kernel: BTRFS info (device dm-0): enabling free space tree Jan 14 00:46:44.215327 kernel: loop: module loaded Jan 14 00:46:44.215335 kernel: loop0: detected capacity change from 0 to 100552 Jan 14 00:46:44.215344 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jan 14 00:46:44.215354 systemd[1]: Successfully made /usr/ read-only. Jan 14 00:46:44.215370 systemd[1]: systemd 257.9 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jan 14 00:46:44.215380 systemd[1]: Detected virtualization kvm. Jan 14 00:46:44.215389 systemd[1]: Detected architecture x86-64. Jan 14 00:46:44.215398 systemd[1]: Running in initrd. Jan 14 00:46:44.215406 systemd[1]: No hostname configured, using default hostname. Jan 14 00:46:44.215416 systemd[1]: Hostname set to . Jan 14 00:46:44.215426 systemd[1]: Initializing machine ID from random generator. Jan 14 00:46:44.215436 systemd[1]: Queued start job for default target initrd.target. Jan 14 00:46:44.215445 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Jan 14 00:46:44.215454 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 14 00:46:44.215463 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 14 00:46:44.215473 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 14 00:46:44.215482 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 14 00:46:44.215494 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 14 00:46:44.215503 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 14 00:46:44.215513 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 14 00:46:44.215523 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 14 00:46:44.215532 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Jan 14 00:46:44.215541 systemd[1]: Reached target paths.target - Path Units. Jan 14 00:46:44.215552 systemd[1]: Reached target slices.target - Slice Units. Jan 14 00:46:44.215561 systemd[1]: Reached target swap.target - Swaps. Jan 14 00:46:44.215570 systemd[1]: Reached target timers.target - Timer Units. Jan 14 00:46:44.215579 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 14 00:46:44.215588 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 14 00:46:44.215596 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Jan 14 00:46:44.215605 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 14 00:46:44.215617 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Jan 14 00:46:44.215627 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 14 00:46:44.215636 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 14 00:46:44.215645 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 14 00:46:44.215654 systemd[1]: Reached target sockets.target - Socket Units. Jan 14 00:46:44.215663 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jan 14 00:46:44.215674 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 14 00:46:44.215683 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 14 00:46:44.215692 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 14 00:46:44.215702 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Jan 14 00:46:44.215711 systemd[1]: Starting systemd-fsck-usr.service... Jan 14 00:46:44.215721 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 14 00:46:44.215730 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 14 00:46:44.215741 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 14 00:46:44.215750 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 14 00:46:44.215760 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 14 00:46:44.215771 systemd[1]: Finished systemd-fsck-usr.service. Jan 14 00:46:44.215780 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 14 00:46:44.215820 systemd-journald[303]: Collecting audit messages is enabled. Jan 14 00:46:44.215846 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 14 00:46:44.215857 kernel: Bridge firewalling registered Jan 14 00:46:44.215866 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 14 00:46:44.215876 systemd-journald[303]: Journal started Jan 14 00:46:44.215895 systemd-journald[303]: Runtime Journal (/run/log/journal/2a1c9cd034314574a7c921e50d60a51b) is 8M, max 78.1M, 70.1M free. Jan 14 00:46:44.210381 systemd-modules-load[305]: Inserted module 'br_netfilter' Jan 14 00:46:44.286161 systemd[1]: Started systemd-journald.service - Journal Service. Jan 14 00:46:44.286238 kernel: audit: type=1130 audit(1768351604.282:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.282000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.293000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.300565 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 14 00:46:44.310310 kernel: audit: type=1130 audit(1768351604.293:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.310335 kernel: audit: type=1130 audit(1768351604.301:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.301000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.302415 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 14 00:46:44.324270 kernel: audit: type=1130 audit(1768351604.310:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.310000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.316217 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 14 00:46:44.327241 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 14 00:46:44.333616 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 14 00:46:44.342042 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 14 00:46:44.352668 systemd-tmpfiles[324]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Jan 14 00:46:44.365064 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 14 00:46:44.367000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.367474 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 14 00:46:44.383191 kernel: audit: type=1130 audit(1768351604.367:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.383218 kernel: audit: type=1130 audit(1768351604.375:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.375000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.383578 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 14 00:46:44.392595 kernel: audit: type=1130 audit(1768351604.384:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.384000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.385080 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 14 00:46:44.401315 kernel: audit: type=1130 audit(1768351604.393:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.393000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.395894 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 14 00:46:44.420000 audit: BPF prog-id=6 op=LOAD Jan 14 00:46:44.422344 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 14 00:46:44.425324 kernel: audit: type=1334 audit(1768351604.420:10): prog-id=6 op=LOAD Jan 14 00:46:44.439305 dracut-cmdline[340]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=akamai verity.usrhash=6d34ab71a3dc5a0ab37eb2c851228af18a1e24f648223df9a1099dbd7db2cfcf Jan 14 00:46:44.490211 systemd-resolved[342]: Positive Trust Anchors: Jan 14 00:46:44.490225 systemd-resolved[342]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 14 00:46:44.490230 systemd-resolved[342]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Jan 14 00:46:44.490259 systemd-resolved[342]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 14 00:46:44.516025 systemd-resolved[342]: Defaulting to hostname 'linux'. Jan 14 00:46:44.517955 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 14 00:46:44.519000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.519663 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 14 00:46:44.572158 kernel: Loading iSCSI transport class v2.0-870. Jan 14 00:46:44.588168 kernel: iscsi: registered transport (tcp) Jan 14 00:46:44.613535 kernel: iscsi: registered transport (qla4xxx) Jan 14 00:46:44.613612 kernel: QLogic iSCSI HBA Driver Jan 14 00:46:44.647349 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 14 00:46:44.687189 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 14 00:46:44.687000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.689845 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 14 00:46:44.758897 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 14 00:46:44.759000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.762187 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 14 00:46:44.764919 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 14 00:46:44.805590 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 14 00:46:44.807000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.808000 audit: BPF prog-id=7 op=LOAD Jan 14 00:46:44.808000 audit: BPF prog-id=8 op=LOAD Jan 14 00:46:44.810299 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 14 00:46:44.842700 systemd-udevd[578]: Using default interface naming scheme 'v257'. Jan 14 00:46:44.859630 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 14 00:46:44.861000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.867496 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 14 00:46:44.901545 dracut-pre-trigger[652]: rd.md=0: removing MD RAID activation Jan 14 00:46:44.904059 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 14 00:46:44.905000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.907000 audit: BPF prog-id=9 op=LOAD Jan 14 00:46:44.909251 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 14 00:46:44.943306 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 14 00:46:44.945000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.947202 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 14 00:46:44.970000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:44.969034 systemd-networkd[692]: lo: Link UP Jan 14 00:46:44.969041 systemd-networkd[692]: lo: Gained carrier Jan 14 00:46:44.969823 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 14 00:46:44.972239 systemd[1]: Reached target network.target - Network. Jan 14 00:46:45.061596 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 14 00:46:45.064000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:45.067974 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 14 00:46:45.206830 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Jan 14 00:46:45.224141 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 Jan 14 00:46:45.229749 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Jan 14 00:46:45.244680 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Jan 14 00:46:45.386239 kernel: cryptd: max_cpu_qlen set to 1000 Jan 14 00:46:45.401389 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Jan 14 00:46:45.438129 kernel: AES CTR mode by8 optimization enabled Jan 14 00:46:45.466381 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 14 00:46:45.473099 systemd-networkd[692]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 14 00:46:45.473125 systemd-networkd[692]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 14 00:46:45.474751 systemd-networkd[692]: eth0: Link UP Jan 14 00:46:45.474969 systemd-networkd[692]: eth0: Gained carrier Jan 14 00:46:45.474980 systemd-networkd[692]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 14 00:46:45.496669 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 14 00:46:45.497010 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 14 00:46:45.500000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:45.500626 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 14 00:46:45.504311 disk-uuid[819]: Primary Header is updated. Jan 14 00:46:45.504311 disk-uuid[819]: Secondary Entries is updated. Jan 14 00:46:45.504311 disk-uuid[819]: Secondary Header is updated. Jan 14 00:46:45.509835 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 14 00:46:45.520337 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 14 00:46:45.523000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:45.528616 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 14 00:46:45.531562 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 14 00:46:45.534581 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 14 00:46:45.541266 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 14 00:46:45.697009 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 14 00:46:45.699000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:45.715305 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 14 00:46:45.715000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:46.333205 systemd-networkd[692]: eth0: DHCPv4 address 172.236.110.177/24, gateway 172.236.110.1 acquired from 23.40.197.106 Jan 14 00:46:46.581058 disk-uuid[821]: Warning: The kernel is still using the old partition table. Jan 14 00:46:46.581058 disk-uuid[821]: The new table will be used at the next reboot or after you Jan 14 00:46:46.581058 disk-uuid[821]: run partprobe(8) or kpartx(8) Jan 14 00:46:46.581058 disk-uuid[821]: The operation has completed successfully. Jan 14 00:46:46.590041 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 14 00:46:46.590244 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 14 00:46:46.609966 kernel: kauditd_printk_skb: 16 callbacks suppressed Jan 14 00:46:46.609992 kernel: audit: type=1130 audit(1768351606.591:27): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:46.610011 kernel: audit: type=1131 audit(1768351606.591:28): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:46.591000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:46.591000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:46.594595 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 14 00:46:46.664566 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (850) Jan 14 00:46:46.665460 kernel: BTRFS info (device sda6): first mount of filesystem 87cf3d96-2540-4b91-98c0-7ae2e759a282 Jan 14 00:46:46.669143 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jan 14 00:46:46.675306 kernel: BTRFS info (device sda6): enabling ssd optimizations Jan 14 00:46:46.675334 kernel: BTRFS info (device sda6): turning on async discard Jan 14 00:46:46.680009 kernel: BTRFS info (device sda6): enabling free space tree Jan 14 00:46:46.691146 kernel: BTRFS info (device sda6): last unmount of filesystem 87cf3d96-2540-4b91-98c0-7ae2e759a282 Jan 14 00:46:46.692815 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 14 00:46:46.702074 kernel: audit: type=1130 audit(1768351606.693:29): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:46.693000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:46.695001 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 14 00:46:46.833786 ignition[869]: Ignition 2.24.0 Jan 14 00:46:46.833812 ignition[869]: Stage: fetch-offline Jan 14 00:46:46.833872 ignition[869]: no configs at "/usr/lib/ignition/base.d" Jan 14 00:46:46.833887 ignition[869]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jan 14 00:46:46.833985 ignition[869]: parsed url from cmdline: "" Jan 14 00:46:46.837600 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 14 00:46:46.847899 kernel: audit: type=1130 audit(1768351606.838:30): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:46.838000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:46.833990 ignition[869]: no config URL provided Jan 14 00:46:46.841265 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Jan 14 00:46:46.833996 ignition[869]: reading system config file "/usr/lib/ignition/user.ign" Jan 14 00:46:46.834010 ignition[869]: no config at "/usr/lib/ignition/user.ign" Jan 14 00:46:46.834016 ignition[869]: failed to fetch config: resource requires networking Jan 14 00:46:46.834707 ignition[869]: Ignition finished successfully Jan 14 00:46:46.868133 ignition[878]: Ignition 2.24.0 Jan 14 00:46:46.868141 ignition[878]: Stage: fetch Jan 14 00:46:46.868290 ignition[878]: no configs at "/usr/lib/ignition/base.d" Jan 14 00:46:46.868301 ignition[878]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jan 14 00:46:46.868390 ignition[878]: parsed url from cmdline: "" Jan 14 00:46:46.868395 ignition[878]: no config URL provided Jan 14 00:46:46.868404 ignition[878]: reading system config file "/usr/lib/ignition/user.ign" Jan 14 00:46:46.868412 ignition[878]: no config at "/usr/lib/ignition/user.ign" Jan 14 00:46:46.868450 ignition[878]: PUT http://169.254.169.254/v1/token: attempt #1 Jan 14 00:46:46.955852 ignition[878]: PUT result: OK Jan 14 00:46:46.955981 ignition[878]: GET http://169.254.169.254/v1/user-data: attempt #1 Jan 14 00:46:47.063699 ignition[878]: GET result: OK Jan 14 00:46:47.063816 ignition[878]: parsing config with SHA512: 36e04842fe7476c76c58b905ade73eb2ff58a8c6f05c452b61c5c992f24a8966cd0065b04e090048080106176ea576b6699c47e8ee6f6bf4663b85890afff0c2 Jan 14 00:46:47.071576 unknown[878]: fetched base config from "system" Jan 14 00:46:47.071589 unknown[878]: fetched base config from "system" Jan 14 00:46:47.071994 ignition[878]: fetch: fetch complete Jan 14 00:46:47.071595 unknown[878]: fetched user config from "akamai" Jan 14 00:46:47.077000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:47.072001 ignition[878]: fetch: fetch passed Jan 14 00:46:47.085261 kernel: audit: type=1130 audit(1768351607.077:31): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:47.076212 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Jan 14 00:46:47.072050 ignition[878]: Ignition finished successfully Jan 14 00:46:47.079356 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 14 00:46:47.113458 ignition[884]: Ignition 2.24.0 Jan 14 00:46:47.113473 ignition[884]: Stage: kargs Jan 14 00:46:47.113641 ignition[884]: no configs at "/usr/lib/ignition/base.d" Jan 14 00:46:47.113656 ignition[884]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jan 14 00:46:47.114685 ignition[884]: kargs: kargs passed Jan 14 00:46:47.114728 ignition[884]: Ignition finished successfully Jan 14 00:46:47.117439 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 14 00:46:47.118000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:47.122273 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 14 00:46:47.130595 kernel: audit: type=1130 audit(1768351607.118:32): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:47.160135 ignition[891]: Ignition 2.24.0 Jan 14 00:46:47.160150 ignition[891]: Stage: disks Jan 14 00:46:47.160293 ignition[891]: no configs at "/usr/lib/ignition/base.d" Jan 14 00:46:47.160304 ignition[891]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jan 14 00:46:47.163692 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 14 00:46:47.173632 kernel: audit: type=1130 audit(1768351607.165:33): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:47.165000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:47.161232 ignition[891]: disks: disks passed Jan 14 00:46:47.165344 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 14 00:46:47.161275 ignition[891]: Ignition finished successfully Jan 14 00:46:47.174586 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 14 00:46:47.176075 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 14 00:46:47.178001 systemd[1]: Reached target sysinit.target - System Initialization. Jan 14 00:46:47.179688 systemd[1]: Reached target basic.target - Basic System. Jan 14 00:46:47.182623 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 14 00:46:47.223384 systemd-fsck[899]: ROOT: clean, 15/1631200 files, 112378/1617920 blocks Jan 14 00:46:47.226079 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 14 00:46:47.227000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:47.231209 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 14 00:46:47.237447 kernel: audit: type=1130 audit(1768351607.227:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:47.371145 kernel: EXT4-fs (sda9): mounted filesystem 6efdc615-0e3c-4caf-8d0b-1f38e5c59ef0 r/w with ordered data mode. Quota mode: none. Jan 14 00:46:47.372349 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 14 00:46:47.373855 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 14 00:46:47.377178 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 14 00:46:47.381199 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 14 00:46:47.383140 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jan 14 00:46:47.383200 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 14 00:46:47.383240 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 14 00:46:47.395997 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 14 00:46:47.400253 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 14 00:46:47.409136 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (907) Jan 14 00:46:47.415645 kernel: BTRFS info (device sda6): first mount of filesystem 87cf3d96-2540-4b91-98c0-7ae2e759a282 Jan 14 00:46:47.415682 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jan 14 00:46:47.423244 kernel: BTRFS info (device sda6): enabling ssd optimizations Jan 14 00:46:47.423270 kernel: BTRFS info (device sda6): turning on async discard Jan 14 00:46:47.425419 kernel: BTRFS info (device sda6): enabling free space tree Jan 14 00:46:47.429829 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 14 00:46:47.509300 systemd-networkd[692]: eth0: Gained IPv6LL Jan 14 00:46:47.626714 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 14 00:46:47.627000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:47.631207 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 14 00:46:47.638465 kernel: audit: type=1130 audit(1768351607.627:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:47.646264 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 14 00:46:47.661914 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 14 00:46:47.667455 kernel: BTRFS info (device sda6): last unmount of filesystem 87cf3d96-2540-4b91-98c0-7ae2e759a282 Jan 14 00:46:47.694441 ignition[1005]: INFO : Ignition 2.24.0 Jan 14 00:46:47.704911 kernel: audit: type=1130 audit(1768351607.695:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:47.695000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:47.704997 ignition[1005]: INFO : Stage: mount Jan 14 00:46:47.704997 ignition[1005]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 14 00:46:47.704997 ignition[1005]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jan 14 00:46:47.704997 ignition[1005]: INFO : mount: mount passed Jan 14 00:46:47.704997 ignition[1005]: INFO : Ignition finished successfully Jan 14 00:46:47.705000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:47.694648 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 14 00:46:47.700450 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 14 00:46:47.707611 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 14 00:46:47.728807 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 14 00:46:47.755404 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 (8:6) scanned by mount (1017) Jan 14 00:46:47.755484 kernel: BTRFS info (device sda6): first mount of filesystem 87cf3d96-2540-4b91-98c0-7ae2e759a282 Jan 14 00:46:47.759459 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jan 14 00:46:47.769644 kernel: BTRFS info (device sda6): enabling ssd optimizations Jan 14 00:46:47.769684 kernel: BTRFS info (device sda6): turning on async discard Jan 14 00:46:47.769697 kernel: BTRFS info (device sda6): enabling free space tree Jan 14 00:46:47.774339 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 14 00:46:47.806643 ignition[1034]: INFO : Ignition 2.24.0 Jan 14 00:46:47.806643 ignition[1034]: INFO : Stage: files Jan 14 00:46:47.808497 ignition[1034]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 14 00:46:47.808497 ignition[1034]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jan 14 00:46:47.808497 ignition[1034]: DEBUG : files: compiled without relabeling support, skipping Jan 14 00:46:47.811739 ignition[1034]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 14 00:46:47.811739 ignition[1034]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 14 00:46:47.815287 ignition[1034]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 14 00:46:47.816419 ignition[1034]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 14 00:46:47.818328 unknown[1034]: wrote ssh authorized keys file for user: core Jan 14 00:46:47.819683 ignition[1034]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 14 00:46:47.843466 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Jan 14 00:46:47.843466 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Jan 14 00:46:48.007152 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jan 14 00:46:48.049450 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Jan 14 00:46:48.051541 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jan 14 00:46:48.051541 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jan 14 00:46:48.051541 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jan 14 00:46:48.051541 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jan 14 00:46:48.051541 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 14 00:46:48.051541 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 14 00:46:48.051541 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 14 00:46:48.051541 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 14 00:46:48.060550 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 14 00:46:48.060550 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 14 00:46:48.060550 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jan 14 00:46:48.060550 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jan 14 00:46:48.060550 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jan 14 00:46:48.060550 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Jan 14 00:46:48.561218 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jan 14 00:46:49.124498 ignition[1034]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jan 14 00:46:49.124498 ignition[1034]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Jan 14 00:46:49.127615 ignition[1034]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 14 00:46:49.130588 ignition[1034]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 14 00:46:49.130588 ignition[1034]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Jan 14 00:46:49.130588 ignition[1034]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Jan 14 00:46:49.135017 ignition[1034]: INFO : files: op(d): op(e): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Jan 14 00:46:49.135017 ignition[1034]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Jan 14 00:46:49.135017 ignition[1034]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Jan 14 00:46:49.135017 ignition[1034]: INFO : files: op(f): [started] setting preset to enabled for "prepare-helm.service" Jan 14 00:46:49.135017 ignition[1034]: INFO : files: op(f): [finished] setting preset to enabled for "prepare-helm.service" Jan 14 00:46:49.135017 ignition[1034]: INFO : files: createResultFile: createFiles: op(10): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 14 00:46:49.135017 ignition[1034]: INFO : files: createResultFile: createFiles: op(10): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 14 00:46:49.135017 ignition[1034]: INFO : files: files passed Jan 14 00:46:49.135017 ignition[1034]: INFO : Ignition finished successfully Jan 14 00:46:49.137000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.134501 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 14 00:46:49.140293 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 14 00:46:49.145396 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 14 00:46:49.160608 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 14 00:46:49.162000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.162000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.160732 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 14 00:46:49.175485 initrd-setup-root-after-ignition[1066]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 14 00:46:49.175485 initrd-setup-root-after-ignition[1066]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 14 00:46:49.178633 initrd-setup-root-after-ignition[1070]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 14 00:46:49.182505 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 14 00:46:49.183000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.183601 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 14 00:46:49.186345 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 14 00:46:49.256488 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 14 00:46:49.257000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.257000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.256629 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 14 00:46:49.258249 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 14 00:46:49.259364 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 14 00:46:49.262891 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 14 00:46:49.263828 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 14 00:46:49.293702 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 14 00:46:49.295000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.297289 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 14 00:46:49.318614 systemd[1]: Unnecessary job was removed for dev-mapper-usr.device - /dev/mapper/usr. Jan 14 00:46:49.319058 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 14 00:46:49.320267 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 14 00:46:49.322230 systemd[1]: Stopped target timers.target - Timer Units. Jan 14 00:46:49.325000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.324196 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 14 00:46:49.324353 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 14 00:46:49.326362 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 14 00:46:49.327682 systemd[1]: Stopped target basic.target - Basic System. Jan 14 00:46:49.329262 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 14 00:46:49.330987 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 14 00:46:49.332741 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 14 00:46:49.334616 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Jan 14 00:46:49.336569 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 14 00:46:49.338557 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 14 00:46:49.340720 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 14 00:46:49.342600 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 14 00:46:49.348000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.344751 systemd[1]: Stopped target swap.target - Swaps. Jan 14 00:46:49.346512 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 14 00:46:49.346667 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 14 00:46:49.348686 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 14 00:46:49.355000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.349941 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 14 00:46:49.357000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.351660 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 14 00:46:49.382000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.352183 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 14 00:46:49.353724 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 14 00:46:49.353830 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 14 00:46:49.356552 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 14 00:46:49.389000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.356710 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 14 00:46:49.357774 systemd[1]: ignition-files.service: Deactivated successfully. Jan 14 00:46:49.357917 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 14 00:46:49.385214 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 14 00:46:49.386471 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 14 00:46:49.386597 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 14 00:46:49.392292 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 14 00:46:49.397299 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 14 00:46:49.400000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.397426 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 14 00:46:49.402000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.400454 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 14 00:46:49.409000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.400594 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 14 00:46:49.402276 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 14 00:46:49.402438 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 14 00:46:49.417488 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 14 00:46:49.418000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.418000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.417703 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 14 00:46:49.427785 ignition[1090]: INFO : Ignition 2.24.0 Jan 14 00:46:49.429756 ignition[1090]: INFO : Stage: umount Jan 14 00:46:49.429756 ignition[1090]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 14 00:46:49.429756 ignition[1090]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jan 14 00:46:49.429756 ignition[1090]: INFO : umount: umount passed Jan 14 00:46:49.429756 ignition[1090]: INFO : Ignition finished successfully Jan 14 00:46:49.436000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.437000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.438000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.438000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.435444 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 14 00:46:49.435598 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 14 00:46:49.436596 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 14 00:46:49.436655 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 14 00:46:49.445000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.437493 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 14 00:46:49.437544 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 14 00:46:49.438302 systemd[1]: ignition-fetch.service: Deactivated successfully. Jan 14 00:46:49.438357 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Jan 14 00:46:49.439072 systemd[1]: Stopped target network.target - Network. Jan 14 00:46:49.441712 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 14 00:46:49.441773 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 14 00:46:49.445709 systemd[1]: Stopped target paths.target - Path Units. Jan 14 00:46:49.447485 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 14 00:46:49.453673 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 14 00:46:49.454619 systemd[1]: Stopped target slices.target - Slice Units. Jan 14 00:46:49.461917 systemd[1]: Stopped target sockets.target - Socket Units. Jan 14 00:46:49.464468 systemd[1]: iscsid.socket: Deactivated successfully. Jan 14 00:46:49.464519 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 14 00:46:49.465929 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 14 00:46:49.465989 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 14 00:46:49.473000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.469587 systemd[1]: systemd-journald-audit.socket: Deactivated successfully. Jan 14 00:46:49.475000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup-pre comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.469624 systemd[1]: Closed systemd-journald-audit.socket - Journal Audit Socket. Jan 14 00:46:49.471402 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 14 00:46:49.471464 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 14 00:46:49.473419 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 14 00:46:49.473471 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 14 00:46:49.475386 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 14 00:46:49.477485 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 14 00:46:49.484965 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 14 00:46:49.486045 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 14 00:46:49.487000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.486609 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 14 00:46:49.489000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.488202 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 14 00:46:49.488333 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 14 00:46:49.491450 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 14 00:46:49.493000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.491626 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 14 00:46:49.496000 audit: BPF prog-id=9 op=UNLOAD Jan 14 00:46:49.497000 audit: BPF prog-id=6 op=UNLOAD Jan 14 00:46:49.497681 systemd[1]: Stopped target network-pre.target - Preparation for Network. Jan 14 00:46:49.498727 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 14 00:46:49.498773 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 14 00:46:49.502000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.500764 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 14 00:46:49.500825 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 14 00:46:49.507000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.504217 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 14 00:46:49.506263 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 14 00:46:49.506388 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 14 00:46:49.510000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.509244 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 14 00:46:49.512000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.509303 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 14 00:46:49.510661 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 14 00:46:49.510715 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 14 00:46:49.512596 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 14 00:46:49.535538 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 14 00:46:49.536000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.535715 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 14 00:46:49.537548 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 14 00:46:49.537640 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 14 00:46:49.545000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.540922 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 14 00:46:49.540967 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 14 00:46:49.543247 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 14 00:46:49.548000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.543304 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 14 00:46:49.550000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.546784 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 14 00:46:49.546845 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 14 00:46:49.549099 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 14 00:46:49.556000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.549176 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 14 00:46:49.559000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.551777 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 14 00:46:49.561000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.553691 systemd[1]: systemd-network-generator.service: Deactivated successfully. Jan 14 00:46:49.564000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.553750 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Jan 14 00:46:49.556632 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 14 00:46:49.556685 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 14 00:46:49.591000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.591000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:49.559582 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 14 00:46:49.559637 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 14 00:46:49.563248 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 14 00:46:49.563378 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 14 00:46:49.589755 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 14 00:46:49.590082 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 14 00:46:49.591528 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 14 00:46:49.595301 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 14 00:46:49.627821 systemd[1]: Switching root. Jan 14 00:46:49.666634 systemd-journald[303]: Journal stopped Jan 14 00:46:51.064818 systemd-journald[303]: Received SIGTERM from PID 1 (systemd). Jan 14 00:46:51.064847 kernel: SELinux: policy capability network_peer_controls=1 Jan 14 00:46:51.064859 kernel: SELinux: policy capability open_perms=1 Jan 14 00:46:51.064870 kernel: SELinux: policy capability extended_socket_class=1 Jan 14 00:46:51.064879 kernel: SELinux: policy capability always_check_network=0 Jan 14 00:46:51.064892 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 14 00:46:51.064902 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 14 00:46:51.064912 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jan 14 00:46:51.064922 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 14 00:46:51.064932 kernel: SELinux: policy capability userspace_initial_context=0 Jan 14 00:46:51.064942 systemd[1]: Successfully loaded SELinux policy in 84.741ms. Jan 14 00:46:51.064956 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 9.667ms. Jan 14 00:46:51.064968 systemd[1]: systemd 257.9 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +IPE +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -BTF -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jan 14 00:46:51.064980 systemd[1]: Detected virtualization kvm. Jan 14 00:46:51.064993 systemd[1]: Detected architecture x86-64. Jan 14 00:46:51.065003 systemd[1]: Detected first boot. Jan 14 00:46:51.065014 systemd[1]: Initializing machine ID from random generator. Jan 14 00:46:51.068724 zram_generator::config[1138]: No configuration found. Jan 14 00:46:51.068747 kernel: Guest personality initialized and is inactive Jan 14 00:46:51.068759 kernel: VMCI host device registered (name=vmci, major=10, minor=258) Jan 14 00:46:51.068774 kernel: Initialized host personality Jan 14 00:46:51.068784 kernel: NET: Registered PF_VSOCK protocol family Jan 14 00:46:51.068795 systemd[1]: Populated /etc with preset unit settings. Jan 14 00:46:51.068806 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jan 14 00:46:51.068816 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jan 14 00:46:51.068827 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jan 14 00:46:51.068845 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jan 14 00:46:51.068859 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jan 14 00:46:51.068870 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jan 14 00:46:51.068881 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jan 14 00:46:51.068892 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jan 14 00:46:51.068903 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jan 14 00:46:51.068913 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jan 14 00:46:51.068926 systemd[1]: Created slice user.slice - User and Session Slice. Jan 14 00:46:51.068937 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 14 00:46:51.068948 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 14 00:46:51.068959 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jan 14 00:46:51.068969 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jan 14 00:46:51.068980 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jan 14 00:46:51.068995 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 14 00:46:51.069009 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jan 14 00:46:51.069020 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 14 00:46:51.069031 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 14 00:46:51.069042 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jan 14 00:46:51.069053 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jan 14 00:46:51.069066 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jan 14 00:46:51.069077 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jan 14 00:46:51.069087 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 14 00:46:51.069098 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 14 00:46:51.069130 systemd[1]: Reached target remote-veritysetup.target - Remote Verity Protected Volumes. Jan 14 00:46:51.069143 systemd[1]: Reached target slices.target - Slice Units. Jan 14 00:46:51.069153 systemd[1]: Reached target swap.target - Swaps. Jan 14 00:46:51.069167 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jan 14 00:46:51.069178 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jan 14 00:46:51.069190 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Jan 14 00:46:51.069201 systemd[1]: Listening on systemd-journald-audit.socket - Journal Audit Socket. Jan 14 00:46:51.069215 systemd[1]: Listening on systemd-mountfsd.socket - DDI File System Mounter Socket. Jan 14 00:46:51.069226 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 14 00:46:51.069237 systemd[1]: Listening on systemd-nsresourced.socket - Namespace Resource Manager Socket. Jan 14 00:46:51.069247 systemd[1]: Listening on systemd-oomd.socket - Userspace Out-Of-Memory (OOM) Killer Socket. Jan 14 00:46:51.069259 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 14 00:46:51.069270 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 14 00:46:51.069283 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jan 14 00:46:51.069294 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jan 14 00:46:51.069305 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jan 14 00:46:51.069316 systemd[1]: Mounting media.mount - External Media Directory... Jan 14 00:46:51.069327 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 14 00:46:51.069338 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jan 14 00:46:51.069349 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jan 14 00:46:51.069362 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jan 14 00:46:51.069373 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jan 14 00:46:51.069384 systemd[1]: Reached target machines.target - Containers. Jan 14 00:46:51.069395 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jan 14 00:46:51.069406 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 14 00:46:51.069417 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 14 00:46:51.069428 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jan 14 00:46:51.069441 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 14 00:46:51.069452 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 14 00:46:51.069463 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 14 00:46:51.069474 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jan 14 00:46:51.069485 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 14 00:46:51.069496 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jan 14 00:46:51.069509 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jan 14 00:46:51.069520 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jan 14 00:46:51.069531 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jan 14 00:46:51.069542 systemd[1]: Stopped systemd-fsck-usr.service. Jan 14 00:46:51.069553 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 14 00:46:51.069564 kernel: fuse: init (API version 7.41) Jan 14 00:46:51.069575 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 14 00:46:51.069588 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 14 00:46:51.069599 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 14 00:46:51.069610 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jan 14 00:46:51.069620 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Jan 14 00:46:51.069631 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 14 00:46:51.069642 kernel: ACPI: bus type drm_connector registered Jan 14 00:46:51.069655 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 14 00:46:51.069666 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jan 14 00:46:51.069677 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jan 14 00:46:51.069688 systemd[1]: Mounted media.mount - External Media Directory. Jan 14 00:46:51.069698 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jan 14 00:46:51.069709 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jan 14 00:46:51.069741 systemd-journald[1217]: Collecting audit messages is enabled. Jan 14 00:46:51.069764 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jan 14 00:46:51.069776 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jan 14 00:46:51.069787 systemd-journald[1217]: Journal started Jan 14 00:46:51.069809 systemd-journald[1217]: Runtime Journal (/run/log/journal/222697d5d56049888d1a97ea2860142c) is 8M, max 78.1M, 70.1M free. Jan 14 00:46:50.712000 audit[1]: EVENT_LISTENER pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Jan 14 00:46:50.926000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:50.934000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:50.944000 audit: BPF prog-id=14 op=UNLOAD Jan 14 00:46:50.944000 audit: BPF prog-id=13 op=UNLOAD Jan 14 00:46:50.945000 audit: BPF prog-id=15 op=LOAD Jan 14 00:46:50.945000 audit: BPF prog-id=16 op=LOAD Jan 14 00:46:50.946000 audit: BPF prog-id=17 op=LOAD Jan 14 00:46:51.054000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Jan 14 00:46:51.054000 audit[1217]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=3 a1=7ffeeaee2620 a2=4000 a3=0 items=0 ppid=1 pid=1217 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:51.054000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Jan 14 00:46:51.070000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:50.566764 systemd[1]: Queued start job for default target multi-user.target. Jan 14 00:46:50.593681 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Jan 14 00:46:50.594560 systemd[1]: systemd-journald.service: Deactivated successfully. Jan 14 00:46:51.074626 systemd[1]: Started systemd-journald.service - Journal Service. Jan 14 00:46:51.073000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.075215 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 14 00:46:51.076000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.077850 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jan 14 00:46:51.078056 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jan 14 00:46:51.079000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.079000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.080091 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 14 00:46:51.080466 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 14 00:46:51.081000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.081000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.081762 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 14 00:46:51.082095 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 14 00:46:51.082000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.082000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.083530 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 14 00:46:51.083853 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 14 00:46:51.084000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.084000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.085371 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jan 14 00:46:51.085665 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jan 14 00:46:51.086000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.086000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.087088 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 14 00:46:51.087401 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 14 00:46:51.088000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.088000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.088709 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 14 00:46:51.089000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.090035 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 14 00:46:51.090000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.092081 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jan 14 00:46:51.092000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.093789 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Jan 14 00:46:51.094000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-load-credentials comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.110764 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 14 00:46:51.113683 systemd[1]: Listening on systemd-importd.socket - Disk Image Download Service Socket. Jan 14 00:46:51.114548 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jan 14 00:46:51.114626 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 14 00:46:51.116271 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Jan 14 00:46:51.117422 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 14 00:46:51.117599 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 14 00:46:51.121260 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jan 14 00:46:51.124435 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jan 14 00:46:51.126288 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 14 00:46:51.131394 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jan 14 00:46:51.133217 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 14 00:46:51.134380 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 14 00:46:51.138266 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jan 14 00:46:51.144212 systemd-journald[1217]: Time spent on flushing to /var/log/journal/222697d5d56049888d1a97ea2860142c is 76.467ms for 1112 entries. Jan 14 00:46:51.144212 systemd-journald[1217]: System Journal (/var/log/journal/222697d5d56049888d1a97ea2860142c) is 8M, max 588.1M, 580.1M free. Jan 14 00:46:51.258346 systemd-journald[1217]: Received client request to flush runtime journal. Jan 14 00:46:51.258397 kernel: loop1: detected capacity change from 0 to 111560 Jan 14 00:46:51.258425 kernel: loop2: detected capacity change from 0 to 229808 Jan 14 00:46:51.191000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.208000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.224000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.230000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.232000 audit: BPF prog-id=18 op=LOAD Jan 14 00:46:51.234000 audit: BPF prog-id=19 op=LOAD Jan 14 00:46:51.235000 audit: BPF prog-id=20 op=LOAD Jan 14 00:46:51.239000 audit: BPF prog-id=21 op=LOAD Jan 14 00:46:51.148357 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jan 14 00:46:51.190239 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jan 14 00:46:51.191450 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jan 14 00:46:51.197937 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Jan 14 00:46:51.206716 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 14 00:46:51.223551 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 14 00:46:51.229299 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jan 14 00:46:51.236329 systemd[1]: Starting systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer... Jan 14 00:46:51.241267 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 14 00:46:51.244504 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 14 00:46:51.260000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.256589 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Jan 14 00:46:51.267969 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jan 14 00:46:51.270000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.278000 audit: BPF prog-id=22 op=LOAD Jan 14 00:46:51.278000 audit: BPF prog-id=23 op=LOAD Jan 14 00:46:51.278000 audit: BPF prog-id=24 op=LOAD Jan 14 00:46:51.280328 systemd[1]: Starting systemd-nsresourced.service - Namespace Resource Manager... Jan 14 00:46:51.285000 audit: BPF prog-id=25 op=LOAD Jan 14 00:46:51.285000 audit: BPF prog-id=26 op=LOAD Jan 14 00:46:51.285000 audit: BPF prog-id=27 op=LOAD Jan 14 00:46:51.286370 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jan 14 00:46:51.304139 kernel: loop3: detected capacity change from 0 to 8 Jan 14 00:46:51.324431 systemd-tmpfiles[1269]: ACLs are not supported, ignoring. Jan 14 00:46:51.324447 systemd-tmpfiles[1269]: ACLs are not supported, ignoring. Jan 14 00:46:51.338179 kernel: loop4: detected capacity change from 0 to 50784 Jan 14 00:46:51.340306 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 14 00:46:51.342000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.384152 kernel: loop5: detected capacity change from 0 to 111560 Jan 14 00:46:51.389831 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jan 14 00:46:51.392000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.397837 systemd-nsresourced[1278]: Not setting up BPF subsystem, as functionality has been disabled at compile time. Jan 14 00:46:51.415656 kernel: loop6: detected capacity change from 0 to 229808 Jan 14 00:46:51.415000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-nsresourced comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.413893 systemd[1]: Started systemd-nsresourced.service - Namespace Resource Manager. Jan 14 00:46:51.440129 kernel: loop7: detected capacity change from 0 to 8 Jan 14 00:46:51.458135 kernel: loop1: detected capacity change from 0 to 50784 Jan 14 00:46:51.475191 (sd-merge)[1285]: Using extensions 'containerd-flatcar.raw', 'docker-flatcar.raw', 'kubernetes.raw', 'oem-akamai.raw'. Jan 14 00:46:51.482986 (sd-merge)[1285]: Merged extensions into '/usr'. Jan 14 00:46:51.495362 systemd[1]: Reload requested from client PID 1258 ('systemd-sysext') (unit systemd-sysext.service)... Jan 14 00:46:51.495379 systemd[1]: Reloading... Jan 14 00:46:51.621623 systemd-oomd[1267]: No swap; memory pressure usage will be degraded Jan 14 00:46:51.650176 zram_generator::config[1329]: No configuration found. Jan 14 00:46:51.666939 systemd-resolved[1268]: Positive Trust Anchors: Jan 14 00:46:51.667174 systemd-resolved[1268]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 14 00:46:51.667180 systemd-resolved[1268]: . IN DS 38696 8 2 683d2d0acb8c9b712a1948b27f741219298d0a450d612c483af444a4c0fb2b16 Jan 14 00:46:51.667208 systemd-resolved[1268]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 14 00:46:51.676060 systemd-resolved[1268]: Defaulting to hostname 'linux'. Jan 14 00:46:51.862305 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jan 14 00:46:51.862670 systemd[1]: Reloading finished in 366 ms. Jan 14 00:46:51.894829 systemd[1]: Started systemd-oomd.service - Userspace Out-Of-Memory (OOM) Killer. Jan 14 00:46:51.895000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-oomd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.895881 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 14 00:46:51.896229 kernel: kauditd_printk_skb: 110 callbacks suppressed Jan 14 00:46:51.896260 kernel: audit: type=1130 audit(1768351611.895:145): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-oomd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.904000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.905440 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jan 14 00:46:51.912560 kernel: audit: type=1130 audit(1768351611.904:146): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.912603 kernel: audit: type=1130 audit(1768351611.911:147): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.911000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.912545 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jan 14 00:46:51.925135 kernel: audit: type=1130 audit(1768351611.918:148): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.918000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:51.924332 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 14 00:46:51.927672 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jan 14 00:46:51.929944 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jan 14 00:46:51.942343 systemd[1]: Starting ensure-sysext.service... Jan 14 00:46:51.949000 audit: BPF prog-id=8 op=UNLOAD Jan 14 00:46:51.948354 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 14 00:46:51.952633 kernel: audit: type=1334 audit(1768351611.949:149): prog-id=8 op=UNLOAD Jan 14 00:46:51.956532 kernel: audit: type=1334 audit(1768351611.949:150): prog-id=7 op=UNLOAD Jan 14 00:46:51.949000 audit: BPF prog-id=7 op=UNLOAD Jan 14 00:46:51.955917 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 14 00:46:51.952000 audit: BPF prog-id=28 op=LOAD Jan 14 00:46:51.952000 audit: BPF prog-id=29 op=LOAD Jan 14 00:46:51.957201 kernel: audit: type=1334 audit(1768351611.952:151): prog-id=28 op=LOAD Jan 14 00:46:51.957235 kernel: audit: type=1334 audit(1768351611.952:152): prog-id=29 op=LOAD Jan 14 00:46:51.972000 audit: BPF prog-id=30 op=LOAD Jan 14 00:46:51.976155 kernel: audit: type=1334 audit(1768351611.972:153): prog-id=30 op=LOAD Jan 14 00:46:51.975000 audit: BPF prog-id=15 op=UNLOAD Jan 14 00:46:51.979633 kernel: audit: type=1334 audit(1768351611.975:154): prog-id=15 op=UNLOAD Jan 14 00:46:51.975000 audit: BPF prog-id=31 op=LOAD Jan 14 00:46:51.975000 audit: BPF prog-id=32 op=LOAD Jan 14 00:46:51.975000 audit: BPF prog-id=16 op=UNLOAD Jan 14 00:46:51.975000 audit: BPF prog-id=17 op=UNLOAD Jan 14 00:46:51.976000 audit: BPF prog-id=33 op=LOAD Jan 14 00:46:51.976000 audit: BPF prog-id=21 op=UNLOAD Jan 14 00:46:51.976000 audit: BPF prog-id=34 op=LOAD Jan 14 00:46:51.976000 audit: BPF prog-id=18 op=UNLOAD Jan 14 00:46:51.976000 audit: BPF prog-id=35 op=LOAD Jan 14 00:46:51.976000 audit: BPF prog-id=36 op=LOAD Jan 14 00:46:51.976000 audit: BPF prog-id=19 op=UNLOAD Jan 14 00:46:51.976000 audit: BPF prog-id=20 op=UNLOAD Jan 14 00:46:51.978000 audit: BPF prog-id=37 op=LOAD Jan 14 00:46:51.978000 audit: BPF prog-id=22 op=UNLOAD Jan 14 00:46:51.978000 audit: BPF prog-id=38 op=LOAD Jan 14 00:46:51.978000 audit: BPF prog-id=39 op=LOAD Jan 14 00:46:51.978000 audit: BPF prog-id=23 op=UNLOAD Jan 14 00:46:51.978000 audit: BPF prog-id=24 op=UNLOAD Jan 14 00:46:51.979000 audit: BPF prog-id=40 op=LOAD Jan 14 00:46:51.979000 audit: BPF prog-id=25 op=UNLOAD Jan 14 00:46:51.979000 audit: BPF prog-id=41 op=LOAD Jan 14 00:46:51.980000 audit: BPF prog-id=42 op=LOAD Jan 14 00:46:51.980000 audit: BPF prog-id=26 op=UNLOAD Jan 14 00:46:51.980000 audit: BPF prog-id=27 op=UNLOAD Jan 14 00:46:51.982681 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jan 14 00:46:51.983720 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jan 14 00:46:51.988594 systemd[1]: Reload requested from client PID 1374 ('systemctl') (unit ensure-sysext.service)... Jan 14 00:46:51.988666 systemd[1]: Reloading... Jan 14 00:46:51.997970 systemd-tmpfiles[1375]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Jan 14 00:46:51.998012 systemd-tmpfiles[1375]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Jan 14 00:46:51.999353 systemd-tmpfiles[1375]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jan 14 00:46:52.002689 systemd-tmpfiles[1375]: ACLs are not supported, ignoring. Jan 14 00:46:52.002775 systemd-tmpfiles[1375]: ACLs are not supported, ignoring. Jan 14 00:46:52.017752 systemd-udevd[1376]: Using default interface naming scheme 'v257'. Jan 14 00:46:52.017834 systemd-tmpfiles[1375]: Detected autofs mount point /boot during canonicalization of boot. Jan 14 00:46:52.017843 systemd-tmpfiles[1375]: Skipping /boot Jan 14 00:46:52.035479 systemd-tmpfiles[1375]: Detected autofs mount point /boot during canonicalization of boot. Jan 14 00:46:52.035492 systemd-tmpfiles[1375]: Skipping /boot Jan 14 00:46:52.115144 zram_generator::config[1422]: No configuration found. Jan 14 00:46:52.311163 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Jan 14 00:46:52.343131 kernel: ACPI: button: Power Button [PWRF] Jan 14 00:46:52.375195 kernel: mousedev: PS/2 mouse device common for all mice Jan 14 00:46:52.398695 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Jan 14 00:46:52.399082 systemd[1]: Reloading finished in 409 ms. Jan 14 00:46:52.408219 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Jan 14 00:46:52.415357 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 14 00:46:52.418134 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Jan 14 00:46:52.419000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.419874 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 14 00:46:52.420000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.425000 audit: BPF prog-id=43 op=LOAD Jan 14 00:46:52.427000 audit: BPF prog-id=33 op=UNLOAD Jan 14 00:46:52.428000 audit: BPF prog-id=44 op=LOAD Jan 14 00:46:52.429000 audit: BPF prog-id=40 op=UNLOAD Jan 14 00:46:52.429000 audit: BPF prog-id=45 op=LOAD Jan 14 00:46:52.429000 audit: BPF prog-id=46 op=LOAD Jan 14 00:46:52.429000 audit: BPF prog-id=41 op=UNLOAD Jan 14 00:46:52.429000 audit: BPF prog-id=42 op=UNLOAD Jan 14 00:46:52.431000 audit: BPF prog-id=47 op=LOAD Jan 14 00:46:52.431000 audit: BPF prog-id=37 op=UNLOAD Jan 14 00:46:52.431000 audit: BPF prog-id=48 op=LOAD Jan 14 00:46:52.431000 audit: BPF prog-id=49 op=LOAD Jan 14 00:46:52.431000 audit: BPF prog-id=38 op=UNLOAD Jan 14 00:46:52.431000 audit: BPF prog-id=39 op=UNLOAD Jan 14 00:46:52.432000 audit: BPF prog-id=50 op=LOAD Jan 14 00:46:52.433000 audit: BPF prog-id=34 op=UNLOAD Jan 14 00:46:52.433000 audit: BPF prog-id=51 op=LOAD Jan 14 00:46:52.433000 audit: BPF prog-id=52 op=LOAD Jan 14 00:46:52.433000 audit: BPF prog-id=35 op=UNLOAD Jan 14 00:46:52.433000 audit: BPF prog-id=36 op=UNLOAD Jan 14 00:46:52.436000 audit: BPF prog-id=53 op=LOAD Jan 14 00:46:52.436000 audit: BPF prog-id=30 op=UNLOAD Jan 14 00:46:52.436000 audit: BPF prog-id=54 op=LOAD Jan 14 00:46:52.436000 audit: BPF prog-id=55 op=LOAD Jan 14 00:46:52.437000 audit: BPF prog-id=31 op=UNLOAD Jan 14 00:46:52.437000 audit: BPF prog-id=32 op=UNLOAD Jan 14 00:46:52.437000 audit: BPF prog-id=56 op=LOAD Jan 14 00:46:52.437000 audit: BPF prog-id=57 op=LOAD Jan 14 00:46:52.437000 audit: BPF prog-id=28 op=UNLOAD Jan 14 00:46:52.437000 audit: BPF prog-id=29 op=UNLOAD Jan 14 00:46:52.471906 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 14 00:46:52.473778 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 14 00:46:52.477530 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jan 14 00:46:52.479436 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 14 00:46:52.481613 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 14 00:46:52.488367 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 14 00:46:52.494575 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 14 00:46:52.495904 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 14 00:46:52.496102 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 14 00:46:52.500443 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jan 14 00:46:52.502207 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 14 00:46:52.504792 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jan 14 00:46:52.508000 audit: BPF prog-id=58 op=LOAD Jan 14 00:46:52.513477 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 14 00:46:52.532396 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jan 14 00:46:52.534218 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 14 00:46:52.543402 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 14 00:46:52.543580 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 14 00:46:52.543772 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 14 00:46:52.543941 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 14 00:46:52.544022 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 14 00:46:52.544092 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 14 00:46:52.557987 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 14 00:46:52.558214 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 14 00:46:52.561758 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 14 00:46:52.563705 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 14 00:46:52.563898 systemd[1]: systemd-confext.service - Merge System Configuration Images into /etc/ was skipped because no trigger condition checks were met. Jan 14 00:46:52.563988 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jan 14 00:46:52.564201 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 14 00:46:52.577859 systemd[1]: Finished ensure-sysext.service. Jan 14 00:46:52.579000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ensure-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.582000 audit: BPF prog-id=59 op=LOAD Jan 14 00:46:52.585661 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jan 14 00:46:52.593715 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jan 14 00:46:52.595000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.595664 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 14 00:46:52.596170 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 14 00:46:52.600000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.600000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.603000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.603000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.600653 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 14 00:46:52.602534 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 14 00:46:52.617427 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 14 00:46:52.617461 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 14 00:46:52.619000 audit[1510]: SYSTEM_BOOT pid=1510 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.632947 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jan 14 00:46:52.636000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.647385 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 14 00:46:52.649045 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 14 00:46:52.651000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.651000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.654000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.654000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.651689 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 14 00:46:52.652685 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 14 00:46:52.657377 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 14 00:46:52.701468 kernel: EDAC MC: Ver: 3.0.0 Jan 14 00:46:52.700335 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jan 14 00:46:52.702000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:52.732623 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 14 00:46:52.771675 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Jan 14 00:46:52.778242 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jan 14 00:46:52.789000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Jan 14 00:46:52.789000 audit[1542]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7fff23539c90 a2=420 a3=0 items=0 ppid=1496 pid=1542 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:52.789000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 14 00:46:52.791764 augenrules[1542]: No rules Jan 14 00:46:52.795916 systemd[1]: audit-rules.service: Deactivated successfully. Jan 14 00:46:52.797457 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 14 00:46:52.822951 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jan 14 00:46:52.843810 systemd-networkd[1504]: lo: Link UP Jan 14 00:46:52.843821 systemd-networkd[1504]: lo: Gained carrier Jan 14 00:46:52.862861 systemd-networkd[1504]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 14 00:46:52.863460 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 14 00:46:52.864710 systemd-networkd[1504]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 14 00:46:52.865086 systemd[1]: Reached target network.target - Network. Jan 14 00:46:52.868313 systemd-networkd[1504]: eth0: Link UP Jan 14 00:46:52.869099 systemd-networkd[1504]: eth0: Gained carrier Jan 14 00:46:52.869358 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Jan 14 00:46:52.872227 systemd-networkd[1504]: eth0: Found matching .network file, based on potentially unpredictable interface name: /usr/lib/systemd/network/zz-default.network Jan 14 00:46:52.873575 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jan 14 00:46:52.919890 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Jan 14 00:46:53.069891 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jan 14 00:46:53.096373 systemd[1]: Reached target time-set.target - System Time Set. Jan 14 00:46:53.105703 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 14 00:46:53.218214 ldconfig[1501]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jan 14 00:46:53.223261 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jan 14 00:46:53.226070 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jan 14 00:46:53.250127 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jan 14 00:46:53.251286 systemd[1]: Reached target sysinit.target - System Initialization. Jan 14 00:46:53.252282 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jan 14 00:46:53.253065 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jan 14 00:46:53.253845 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Jan 14 00:46:53.254781 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jan 14 00:46:53.255639 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jan 14 00:46:53.256424 systemd[1]: Started systemd-sysupdate-reboot.timer - Reboot Automatically After System Update. Jan 14 00:46:53.257286 systemd[1]: Started systemd-sysupdate.timer - Automatic System Update. Jan 14 00:46:53.258004 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jan 14 00:46:53.258767 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jan 14 00:46:53.258805 systemd[1]: Reached target paths.target - Path Units. Jan 14 00:46:53.259516 systemd[1]: Reached target timers.target - Timer Units. Jan 14 00:46:53.261638 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jan 14 00:46:53.264328 systemd[1]: Starting docker.socket - Docker Socket for the API... Jan 14 00:46:53.267025 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Jan 14 00:46:53.267958 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Jan 14 00:46:53.268732 systemd[1]: Reached target ssh-access.target - SSH Access Available. Jan 14 00:46:53.281210 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jan 14 00:46:53.282485 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Jan 14 00:46:53.283810 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jan 14 00:46:53.285566 systemd[1]: Reached target sockets.target - Socket Units. Jan 14 00:46:53.286272 systemd[1]: Reached target basic.target - Basic System. Jan 14 00:46:53.286986 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jan 14 00:46:53.287022 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jan 14 00:46:53.288466 systemd[1]: Starting containerd.service - containerd container runtime... Jan 14 00:46:53.292441 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Jan 14 00:46:53.298772 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jan 14 00:46:53.302377 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jan 14 00:46:53.308263 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jan 14 00:46:53.313357 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jan 14 00:46:53.314192 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jan 14 00:46:53.321310 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Jan 14 00:46:53.331324 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jan 14 00:46:53.339422 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jan 14 00:46:53.344411 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jan 14 00:46:53.350844 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jan 14 00:46:53.355380 jq[1568]: false Jan 14 00:46:53.370375 systemd[1]: Starting systemd-logind.service - User Login Management... Jan 14 00:46:53.371411 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jan 14 00:46:53.372272 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jan 14 00:46:53.376937 systemd[1]: Starting update-engine.service - Update Engine... Jan 14 00:46:53.380936 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jan 14 00:46:53.390122 google_oslogin_nss_cache[1570]: oslogin_cache_refresh[1570]: Refreshing passwd entry cache Jan 14 00:46:53.398369 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jan 14 00:46:53.399752 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jan 14 00:46:53.400800 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jan 14 00:46:53.404712 coreos-metadata[1565]: Jan 14 00:46:53.404 INFO Putting http://169.254.169.254/v1/token: Attempt #1 Jan 14 00:46:53.406279 oslogin_cache_refresh[1570]: Refreshing passwd entry cache Jan 14 00:46:53.417506 google_oslogin_nss_cache[1570]: oslogin_cache_refresh[1570]: Failure getting users, quitting Jan 14 00:46:53.417506 google_oslogin_nss_cache[1570]: oslogin_cache_refresh[1570]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Jan 14 00:46:53.417494 oslogin_cache_refresh[1570]: Failure getting users, quitting Jan 14 00:46:53.417612 google_oslogin_nss_cache[1570]: oslogin_cache_refresh[1570]: Refreshing group entry cache Jan 14 00:46:53.417511 oslogin_cache_refresh[1570]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Jan 14 00:46:53.417553 oslogin_cache_refresh[1570]: Refreshing group entry cache Jan 14 00:46:53.418028 google_oslogin_nss_cache[1570]: oslogin_cache_refresh[1570]: Failure getting groups, quitting Jan 14 00:46:53.418028 google_oslogin_nss_cache[1570]: oslogin_cache_refresh[1570]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Jan 14 00:46:53.418018 oslogin_cache_refresh[1570]: Failure getting groups, quitting Jan 14 00:46:53.418029 oslogin_cache_refresh[1570]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Jan 14 00:46:53.420706 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jan 14 00:46:53.426253 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jan 14 00:46:53.431169 extend-filesystems[1569]: Found /dev/sda6 Jan 14 00:46:53.440667 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Jan 14 00:46:53.442197 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Jan 14 00:46:53.451987 update_engine[1581]: I20260114 00:46:53.451910 1581 main.cc:92] Flatcar Update Engine starting Jan 14 00:46:53.455087 jq[1582]: true Jan 14 00:46:53.455459 extend-filesystems[1569]: Found /dev/sda9 Jan 14 00:46:53.468544 extend-filesystems[1569]: Checking size of /dev/sda9 Jan 14 00:46:53.475521 tar[1587]: linux-amd64/LICENSE Jan 14 00:46:53.475738 tar[1587]: linux-amd64/helm Jan 14 00:46:53.493519 systemd[1]: motdgen.service: Deactivated successfully. Jan 14 00:46:53.494945 dbus-daemon[1566]: [system] SELinux support is enabled Jan 14 00:46:53.493843 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jan 14 00:46:53.495764 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jan 14 00:46:53.500064 extend-filesystems[1569]: Resized partition /dev/sda9 Jan 14 00:46:53.501655 update_engine[1581]: I20260114 00:46:53.501058 1581 update_check_scheduler.cc:74] Next update check in 9m41s Jan 14 00:46:53.501270 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jan 14 00:46:53.501292 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jan 14 00:46:53.505151 extend-filesystems[1620]: resize2fs 1.47.3 (8-Jul-2025) Jan 14 00:46:53.502621 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jan 14 00:46:53.502639 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jan 14 00:46:53.512124 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 19377147 blocks Jan 14 00:46:53.514239 systemd[1]: Started update-engine.service - Update Engine. Jan 14 00:46:53.517141 jq[1613]: true Jan 14 00:46:53.518134 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jan 14 00:46:53.585585 sshd_keygen[1588]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jan 14 00:46:53.686090 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jan 14 00:46:53.693448 systemd[1]: Starting issuegen.service - Generate /run/issue... Jan 14 00:46:53.704012 systemd-logind[1578]: Watching system buttons on /dev/input/event2 (Power Button) Jan 14 00:46:53.704047 systemd-logind[1578]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jan 14 00:46:53.704597 bash[1643]: Updated "/home/core/.ssh/authorized_keys" Jan 14 00:46:53.706747 systemd-logind[1578]: New seat seat0. Jan 14 00:46:53.711689 systemd[1]: Started systemd-logind.service - User Login Management. Jan 14 00:46:53.713274 systemd-networkd[1504]: eth0: DHCPv4 address 172.236.110.177/24, gateway 172.236.110.1 acquired from 23.40.197.106 Jan 14 00:46:53.715059 systemd-timesyncd[1514]: Network configuration changed, trying to establish connection. Jan 14 00:46:53.715806 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jan 14 00:46:53.719409 dbus-daemon[1566]: [system] Activating via systemd: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.4' (uid=244 pid=1504 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Jan 14 00:46:53.722298 systemd[1]: Starting sshkeys.service... Jan 14 00:46:53.738771 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Jan 14 00:46:53.753602 systemd[1]: issuegen.service: Deactivated successfully. Jan 14 00:46:53.757612 systemd[1]: Finished issuegen.service - Generate /run/issue. Jan 14 00:46:53.765046 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jan 14 00:46:53.847409 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Jan 14 00:46:53.853263 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Jan 14 00:46:53.859759 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Jan 14 00:46:53.874934 dbus-daemon[1566]: [system] Successfully activated service 'org.freedesktop.hostname1' Jan 14 00:46:53.879298 dbus-daemon[1566]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.7' (uid=0 pid=1657 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Jan 14 00:46:53.895729 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jan 14 00:46:53.905624 systemd[1]: Started getty@tty1.service - Getty on tty1. Jan 14 00:46:53.914242 systemd[1]: Starting polkit.service - Authorization Manager... Jan 14 00:46:53.919233 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jan 14 00:46:53.921429 systemd[1]: Reached target getty.target - Login Prompts. Jan 14 00:46:53.955306 kernel: EXT4-fs (sda9): resized filesystem to 19377147 Jan 14 00:46:53.971989 containerd[1611]: time="2026-01-14T00:46:53Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Jan 14 00:46:53.972660 systemd-timesyncd[1514]: Contacted time server 162.159.200.123:123 (0.flatcar.pool.ntp.org). Jan 14 00:46:53.973026 systemd-timesyncd[1514]: Initial clock synchronization to Wed 2026-01-14 00:46:54.088915 UTC. Jan 14 00:46:53.974237 extend-filesystems[1620]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Jan 14 00:46:53.974237 extend-filesystems[1620]: old_desc_blocks = 1, new_desc_blocks = 10 Jan 14 00:46:53.974237 extend-filesystems[1620]: The filesystem on /dev/sda9 is now 19377147 (4k) blocks long. Jan 14 00:46:53.983244 extend-filesystems[1569]: Resized filesystem in /dev/sda9 Jan 14 00:46:53.980268 systemd[1]: extend-filesystems.service: Deactivated successfully. Jan 14 00:46:53.991334 containerd[1611]: time="2026-01-14T00:46:53.983340936Z" level=info msg="starting containerd" revision=fcd43222d6b07379a4be9786bda52438f0dd16a1 version=v2.1.5 Jan 14 00:46:53.981076 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jan 14 00:46:54.004921 containerd[1611]: time="2026-01-14T00:46:54.003852451Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="7.826µs" Jan 14 00:46:54.004921 containerd[1611]: time="2026-01-14T00:46:54.003881560Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Jan 14 00:46:54.004921 containerd[1611]: time="2026-01-14T00:46:54.003917589Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Jan 14 00:46:54.004921 containerd[1611]: time="2026-01-14T00:46:54.003929872Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Jan 14 00:46:54.004921 containerd[1611]: time="2026-01-14T00:46:54.004101336Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Jan 14 00:46:54.004921 containerd[1611]: time="2026-01-14T00:46:54.004132791Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jan 14 00:46:54.004921 containerd[1611]: time="2026-01-14T00:46:54.004711193Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jan 14 00:46:54.004921 containerd[1611]: time="2026-01-14T00:46:54.004729075Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jan 14 00:46:54.005295 containerd[1611]: time="2026-01-14T00:46:54.005260383Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jan 14 00:46:54.005295 containerd[1611]: time="2026-01-14T00:46:54.005285611Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jan 14 00:46:54.005295 containerd[1611]: time="2026-01-14T00:46:54.005297417Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jan 14 00:46:54.005376 containerd[1611]: time="2026-01-14T00:46:54.005305962Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Jan 14 00:46:54.005521 containerd[1611]: time="2026-01-14T00:46:54.005491873Z" level=info msg="skip loading plugin" error="EROFS unsupported, please `modprobe erofs`: skip plugin" id=io.containerd.snapshotter.v1.erofs type=io.containerd.snapshotter.v1 Jan 14 00:46:54.005521 containerd[1611]: time="2026-01-14T00:46:54.005517762Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Jan 14 00:46:54.005657 containerd[1611]: time="2026-01-14T00:46:54.005608909Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Jan 14 00:46:54.005879 containerd[1611]: time="2026-01-14T00:46:54.005848548Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jan 14 00:46:54.005926 containerd[1611]: time="2026-01-14T00:46:54.005890459Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jan 14 00:46:54.005926 containerd[1611]: time="2026-01-14T00:46:54.005904084Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Jan 14 00:46:54.005926 containerd[1611]: time="2026-01-14T00:46:54.005923765Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Jan 14 00:46:54.008066 containerd[1611]: time="2026-01-14T00:46:54.007599997Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Jan 14 00:46:54.008066 containerd[1611]: time="2026-01-14T00:46:54.007679267Z" level=info msg="metadata content store policy set" policy=shared Jan 14 00:46:54.012325 containerd[1611]: time="2026-01-14T00:46:54.012292240Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Jan 14 00:46:54.012365 containerd[1611]: time="2026-01-14T00:46:54.012340359Z" level=info msg="loading plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Jan 14 00:46:54.012468 containerd[1611]: time="2026-01-14T00:46:54.012426172Z" level=info msg="skip loading plugin" error="could not find mkfs.erofs: exec: \"mkfs.erofs\": executable file not found in $PATH: skip plugin" id=io.containerd.differ.v1.erofs type=io.containerd.differ.v1 Jan 14 00:46:54.012468 containerd[1611]: time="2026-01-14T00:46:54.012446777Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Jan 14 00:46:54.012468 containerd[1611]: time="2026-01-14T00:46:54.012458827Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Jan 14 00:46:54.012468 containerd[1611]: time="2026-01-14T00:46:54.012468652Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Jan 14 00:46:54.012552 containerd[1611]: time="2026-01-14T00:46:54.012478579Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Jan 14 00:46:54.012552 containerd[1611]: time="2026-01-14T00:46:54.012487520Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Jan 14 00:46:54.012552 containerd[1611]: time="2026-01-14T00:46:54.012503187Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Jan 14 00:46:54.012552 containerd[1611]: time="2026-01-14T00:46:54.012513865Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Jan 14 00:46:54.012552 containerd[1611]: time="2026-01-14T00:46:54.012523213Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Jan 14 00:46:54.012552 containerd[1611]: time="2026-01-14T00:46:54.012533302Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Jan 14 00:46:54.012828 containerd[1611]: time="2026-01-14T00:46:54.012738498Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Jan 14 00:46:54.012828 containerd[1611]: time="2026-01-14T00:46:54.012750121Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Jan 14 00:46:54.012890 containerd[1611]: time="2026-01-14T00:46:54.012861833Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Jan 14 00:46:54.012890 containerd[1611]: time="2026-01-14T00:46:54.012888839Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Jan 14 00:46:54.012937 containerd[1611]: time="2026-01-14T00:46:54.012902047Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Jan 14 00:46:54.012937 containerd[1611]: time="2026-01-14T00:46:54.012918862Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Jan 14 00:46:54.012937 containerd[1611]: time="2026-01-14T00:46:54.012929266Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Jan 14 00:46:54.012994 containerd[1611]: time="2026-01-14T00:46:54.012939284Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Jan 14 00:46:54.012994 containerd[1611]: time="2026-01-14T00:46:54.012950105Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Jan 14 00:46:54.012994 containerd[1611]: time="2026-01-14T00:46:54.012960367Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Jan 14 00:46:54.012994 containerd[1611]: time="2026-01-14T00:46:54.012976135Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Jan 14 00:46:54.012994 containerd[1611]: time="2026-01-14T00:46:54.012986255Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Jan 14 00:46:54.012994 containerd[1611]: time="2026-01-14T00:46:54.012995592Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Jan 14 00:46:54.013099 containerd[1611]: time="2026-01-14T00:46:54.013015252Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Jan 14 00:46:54.013099 containerd[1611]: time="2026-01-14T00:46:54.013046160Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Jan 14 00:46:54.013099 containerd[1611]: time="2026-01-14T00:46:54.013057051Z" level=info msg="Start snapshots syncer" Jan 14 00:46:54.013099 containerd[1611]: time="2026-01-14T00:46:54.013078591Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Jan 14 00:46:54.013553 containerd[1611]: time="2026-01-14T00:46:54.013310558Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"cgroupWritable\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"\",\"binDirs\":[\"/opt/cni/bin\"],\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogLineSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Jan 14 00:46:54.013672 containerd[1611]: time="2026-01-14T00:46:54.013560307Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Jan 14 00:46:54.013672 containerd[1611]: time="2026-01-14T00:46:54.013602492Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Jan 14 00:46:54.013720 containerd[1611]: time="2026-01-14T00:46:54.013704521Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Jan 14 00:46:54.013741 containerd[1611]: time="2026-01-14T00:46:54.013724374Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Jan 14 00:46:54.013741 containerd[1611]: time="2026-01-14T00:46:54.013734809Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Jan 14 00:46:54.013775 containerd[1611]: time="2026-01-14T00:46:54.013744095Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Jan 14 00:46:54.013775 containerd[1611]: time="2026-01-14T00:46:54.013755820Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Jan 14 00:46:54.013775 containerd[1611]: time="2026-01-14T00:46:54.013775013Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Jan 14 00:46:54.013848 containerd[1611]: time="2026-01-14T00:46:54.013785366Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Jan 14 00:46:54.013848 containerd[1611]: time="2026-01-14T00:46:54.013795963Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Jan 14 00:46:54.013848 containerd[1611]: time="2026-01-14T00:46:54.013806032Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Jan 14 00:46:54.013848 containerd[1611]: time="2026-01-14T00:46:54.013833393Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jan 14 00:46:54.013848 containerd[1611]: time="2026-01-14T00:46:54.013844123Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jan 14 00:46:54.013848 containerd[1611]: time="2026-01-14T00:46:54.013852545Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jan 14 00:46:54.013964 containerd[1611]: time="2026-01-14T00:46:54.013863874Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jan 14 00:46:54.013964 containerd[1611]: time="2026-01-14T00:46:54.013872104Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Jan 14 00:46:54.013964 containerd[1611]: time="2026-01-14T00:46:54.013881472Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Jan 14 00:46:54.013964 containerd[1611]: time="2026-01-14T00:46:54.013891733Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Jan 14 00:46:54.013964 containerd[1611]: time="2026-01-14T00:46:54.013903103Z" level=info msg="runtime interface created" Jan 14 00:46:54.013964 containerd[1611]: time="2026-01-14T00:46:54.013908671Z" level=info msg="created NRI interface" Jan 14 00:46:54.013964 containerd[1611]: time="2026-01-14T00:46:54.013916474Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Jan 14 00:46:54.013964 containerd[1611]: time="2026-01-14T00:46:54.013926288Z" level=info msg="Connect containerd service" Jan 14 00:46:54.013964 containerd[1611]: time="2026-01-14T00:46:54.013959787Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jan 14 00:46:54.020161 containerd[1611]: time="2026-01-14T00:46:54.019100604Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 14 00:46:54.044688 locksmithd[1622]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jan 14 00:46:54.063945 coreos-metadata[1663]: Jan 14 00:46:54.063 INFO Putting http://169.254.169.254/v1/token: Attempt #1 Jan 14 00:46:54.086602 polkitd[1668]: Started polkitd version 126 Jan 14 00:46:54.100335 polkitd[1668]: Loading rules from directory /etc/polkit-1/rules.d Jan 14 00:46:54.100860 polkitd[1668]: Loading rules from directory /run/polkit-1/rules.d Jan 14 00:46:54.100911 polkitd[1668]: Error opening rules directory: Error opening directory “/run/polkit-1/rules.d”: No such file or directory (g-file-error-quark, 4) Jan 14 00:46:54.101143 polkitd[1668]: Loading rules from directory /usr/local/share/polkit-1/rules.d Jan 14 00:46:54.101171 polkitd[1668]: Error opening rules directory: Error opening directory “/usr/local/share/polkit-1/rules.d”: No such file or directory (g-file-error-quark, 4) Jan 14 00:46:54.101213 polkitd[1668]: Loading rules from directory /usr/share/polkit-1/rules.d Jan 14 00:46:54.103809 polkitd[1668]: Finished loading, compiling and executing 2 rules Jan 14 00:46:54.104323 systemd[1]: Started polkit.service - Authorization Manager. Jan 14 00:46:54.108673 dbus-daemon[1566]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Jan 14 00:46:54.110460 polkitd[1668]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Jan 14 00:46:54.129791 systemd-hostnamed[1657]: Hostname set to <172-236-110-177> (transient) Jan 14 00:46:54.129830 systemd-resolved[1268]: System hostname changed to '172-236-110-177'. Jan 14 00:46:54.137152 containerd[1611]: time="2026-01-14T00:46:54.137087035Z" level=info msg="Start subscribing containerd event" Jan 14 00:46:54.137208 containerd[1611]: time="2026-01-14T00:46:54.137172421Z" level=info msg="Start recovering state" Jan 14 00:46:54.137612 containerd[1611]: time="2026-01-14T00:46:54.137481728Z" level=info msg="Start event monitor" Jan 14 00:46:54.137612 containerd[1611]: time="2026-01-14T00:46:54.137501724Z" level=info msg="Start cni network conf syncer for default" Jan 14 00:46:54.137612 containerd[1611]: time="2026-01-14T00:46:54.137510604Z" level=info msg="Start streaming server" Jan 14 00:46:54.137612 containerd[1611]: time="2026-01-14T00:46:54.137531635Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Jan 14 00:46:54.137612 containerd[1611]: time="2026-01-14T00:46:54.137539174Z" level=info msg="runtime interface starting up..." Jan 14 00:46:54.137612 containerd[1611]: time="2026-01-14T00:46:54.137545016Z" level=info msg="starting plugins..." Jan 14 00:46:54.137612 containerd[1611]: time="2026-01-14T00:46:54.137561750Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Jan 14 00:46:54.137958 containerd[1611]: time="2026-01-14T00:46:54.137777167Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jan 14 00:46:54.137958 containerd[1611]: time="2026-01-14T00:46:54.137838037Z" level=info msg=serving... address=/run/containerd/containerd.sock Jan 14 00:46:54.138009 systemd[1]: Started containerd.service - containerd container runtime. Jan 14 00:46:54.138515 containerd[1611]: time="2026-01-14T00:46:54.138489419Z" level=info msg="containerd successfully booted in 0.168560s" Jan 14 00:46:54.172650 coreos-metadata[1663]: Jan 14 00:46:54.172 INFO Fetching http://169.254.169.254/v1/ssh-keys: Attempt #1 Jan 14 00:46:54.255016 tar[1587]: linux-amd64/README.md Jan 14 00:46:54.274141 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jan 14 00:46:54.306830 coreos-metadata[1663]: Jan 14 00:46:54.306 INFO Fetch successful Jan 14 00:46:54.328620 update-ssh-keys[1706]: Updated "/home/core/.ssh/authorized_keys" Jan 14 00:46:54.330335 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Jan 14 00:46:54.333275 systemd[1]: Finished sshkeys.service. Jan 14 00:46:54.417984 coreos-metadata[1565]: Jan 14 00:46:54.417 INFO Putting http://169.254.169.254/v1/token: Attempt #2 Jan 14 00:46:54.509722 coreos-metadata[1565]: Jan 14 00:46:54.509 INFO Fetching http://169.254.169.254/v1/instance: Attempt #1 Jan 14 00:46:54.699653 coreos-metadata[1565]: Jan 14 00:46:54.699 INFO Fetch successful Jan 14 00:46:54.699933 coreos-metadata[1565]: Jan 14 00:46:54.699 INFO Fetching http://169.254.169.254/v1/network: Attempt #1 Jan 14 00:46:54.741976 systemd-networkd[1504]: eth0: Gained IPv6LL Jan 14 00:46:54.745287 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jan 14 00:46:54.746884 systemd[1]: Reached target network-online.target - Network is Online. Jan 14 00:46:54.750051 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 00:46:54.754420 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jan 14 00:46:54.788267 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jan 14 00:46:54.961343 coreos-metadata[1565]: Jan 14 00:46:54.961 INFO Fetch successful Jan 14 00:46:55.101637 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Jan 14 00:46:55.103614 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jan 14 00:46:55.736832 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 00:46:55.739023 systemd[1]: Reached target multi-user.target - Multi-User System. Jan 14 00:46:55.740445 systemd[1]: Startup finished in 2.955s (kernel) + 6.076s (initrd) + 6.004s (userspace) = 15.036s. Jan 14 00:46:55.749654 (kubelet)[1747]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 14 00:46:56.298846 kubelet[1747]: E0114 00:46:56.298786 1747 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 14 00:46:56.303069 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 14 00:46:56.303308 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 14 00:46:56.303950 systemd[1]: kubelet.service: Consumed 907ms CPU time, 268.1M memory peak. Jan 14 00:46:56.535095 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jan 14 00:46:56.536619 systemd[1]: Started sshd@0-172.236.110.177:22-68.220.241.50:34742.service - OpenSSH per-connection server daemon (68.220.241.50:34742). Jan 14 00:46:56.709690 sshd[1760]: Accepted publickey for core from 68.220.241.50 port 34742 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:46:56.712374 sshd-session[1760]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:46:56.720246 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jan 14 00:46:56.721583 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jan 14 00:46:56.727732 systemd-logind[1578]: New session 1 of user core. Jan 14 00:46:56.742386 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jan 14 00:46:56.745988 systemd[1]: Starting user@500.service - User Manager for UID 500... Jan 14 00:46:56.767698 (systemd)[1766]: pam_unix(systemd-user:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:46:56.778002 systemd-logind[1578]: New session 2 of user core. Jan 14 00:46:56.912274 systemd[1766]: Queued start job for default target default.target. Jan 14 00:46:56.919402 systemd[1766]: Created slice app.slice - User Application Slice. Jan 14 00:46:56.919434 systemd[1766]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of User's Temporary Directories. Jan 14 00:46:56.919448 systemd[1766]: Reached target paths.target - Paths. Jan 14 00:46:56.919501 systemd[1766]: Reached target timers.target - Timers. Jan 14 00:46:56.921543 systemd[1766]: Starting dbus.socket - D-Bus User Message Bus Socket... Jan 14 00:46:56.924266 systemd[1766]: Starting systemd-tmpfiles-setup.service - Create User Files and Directories... Jan 14 00:46:56.936612 systemd[1766]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jan 14 00:46:56.936781 systemd[1766]: Reached target sockets.target - Sockets. Jan 14 00:46:56.938532 systemd[1766]: Finished systemd-tmpfiles-setup.service - Create User Files and Directories. Jan 14 00:46:56.938768 systemd[1766]: Reached target basic.target - Basic System. Jan 14 00:46:56.938904 systemd[1766]: Reached target default.target - Main User Target. Jan 14 00:46:56.939006 systemd[1766]: Startup finished in 153ms. Jan 14 00:46:56.939160 systemd[1]: Started user@500.service - User Manager for UID 500. Jan 14 00:46:56.943402 systemd[1]: Started session-1.scope - Session 1 of User core. Jan 14 00:46:57.034772 systemd[1]: Started sshd@1-172.236.110.177:22-68.220.241.50:34752.service - OpenSSH per-connection server daemon (68.220.241.50:34752). Jan 14 00:46:57.188175 sshd[1780]: Accepted publickey for core from 68.220.241.50 port 34752 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:46:57.189831 sshd-session[1780]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:46:57.196913 systemd-logind[1578]: New session 3 of user core. Jan 14 00:46:57.203292 systemd[1]: Started session-3.scope - Session 3 of User core. Jan 14 00:46:57.259482 sshd[1784]: Connection closed by 68.220.241.50 port 34752 Jan 14 00:46:57.260317 sshd-session[1780]: pam_unix(sshd:session): session closed for user core Jan 14 00:46:57.266428 systemd-logind[1578]: Session 3 logged out. Waiting for processes to exit. Jan 14 00:46:57.267437 systemd[1]: sshd@1-172.236.110.177:22-68.220.241.50:34752.service: Deactivated successfully. Jan 14 00:46:57.270771 systemd[1]: session-3.scope: Deactivated successfully. Jan 14 00:46:57.272629 systemd-logind[1578]: Removed session 3. Jan 14 00:46:57.293455 systemd[1]: Started sshd@2-172.236.110.177:22-68.220.241.50:34754.service - OpenSSH per-connection server daemon (68.220.241.50:34754). Jan 14 00:46:57.453181 sshd[1790]: Accepted publickey for core from 68.220.241.50 port 34754 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:46:57.454910 sshd-session[1790]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:46:57.462339 systemd-logind[1578]: New session 4 of user core. Jan 14 00:46:57.471290 systemd[1]: Started session-4.scope - Session 4 of User core. Jan 14 00:46:57.526507 sshd[1794]: Connection closed by 68.220.241.50 port 34754 Jan 14 00:46:57.527334 sshd-session[1790]: pam_unix(sshd:session): session closed for user core Jan 14 00:46:57.533228 systemd[1]: sshd@2-172.236.110.177:22-68.220.241.50:34754.service: Deactivated successfully. Jan 14 00:46:57.536112 systemd[1]: session-4.scope: Deactivated successfully. Jan 14 00:46:57.537237 systemd-logind[1578]: Session 4 logged out. Waiting for processes to exit. Jan 14 00:46:57.539031 systemd-logind[1578]: Removed session 4. Jan 14 00:46:57.558813 systemd[1]: Started sshd@3-172.236.110.177:22-68.220.241.50:34756.service - OpenSSH per-connection server daemon (68.220.241.50:34756). Jan 14 00:46:57.712715 sshd[1800]: Accepted publickey for core from 68.220.241.50 port 34756 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:46:57.715332 sshd-session[1800]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:46:57.722744 systemd-logind[1578]: New session 5 of user core. Jan 14 00:46:57.732343 systemd[1]: Started session-5.scope - Session 5 of User core. Jan 14 00:46:57.792281 sshd[1804]: Connection closed by 68.220.241.50 port 34756 Jan 14 00:46:57.793342 sshd-session[1800]: pam_unix(sshd:session): session closed for user core Jan 14 00:46:57.800625 systemd-logind[1578]: Session 5 logged out. Waiting for processes to exit. Jan 14 00:46:57.801652 systemd[1]: sshd@3-172.236.110.177:22-68.220.241.50:34756.service: Deactivated successfully. Jan 14 00:46:57.804931 systemd[1]: session-5.scope: Deactivated successfully. Jan 14 00:46:57.807496 systemd-logind[1578]: Removed session 5. Jan 14 00:46:57.830187 systemd[1]: Started sshd@4-172.236.110.177:22-68.220.241.50:34760.service - OpenSSH per-connection server daemon (68.220.241.50:34760). Jan 14 00:46:58.007156 sshd[1810]: Accepted publickey for core from 68.220.241.50 port 34760 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:46:58.008714 sshd-session[1810]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:46:58.014248 systemd-logind[1578]: New session 6 of user core. Jan 14 00:46:58.020258 systemd[1]: Started session-6.scope - Session 6 of User core. Jan 14 00:46:58.068302 sudo[1815]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jan 14 00:46:58.068669 sudo[1815]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 14 00:46:58.082723 sudo[1815]: pam_unix(sudo:session): session closed for user root Jan 14 00:46:58.104772 sshd[1814]: Connection closed by 68.220.241.50 port 34760 Jan 14 00:46:58.105310 sshd-session[1810]: pam_unix(sshd:session): session closed for user core Jan 14 00:46:58.110835 systemd-logind[1578]: Session 6 logged out. Waiting for processes to exit. Jan 14 00:46:58.111160 systemd[1]: sshd@4-172.236.110.177:22-68.220.241.50:34760.service: Deactivated successfully. Jan 14 00:46:58.113691 systemd[1]: session-6.scope: Deactivated successfully. Jan 14 00:46:58.115636 systemd-logind[1578]: Removed session 6. Jan 14 00:46:58.140217 systemd[1]: Started sshd@5-172.236.110.177:22-68.220.241.50:34768.service - OpenSSH per-connection server daemon (68.220.241.50:34768). Jan 14 00:46:58.324145 sshd[1822]: Accepted publickey for core from 68.220.241.50 port 34768 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:46:58.325997 sshd-session[1822]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:46:58.332322 systemd-logind[1578]: New session 7 of user core. Jan 14 00:46:58.341261 systemd[1]: Started session-7.scope - Session 7 of User core. Jan 14 00:46:58.393699 sudo[1828]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jan 14 00:46:58.394548 sudo[1828]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 14 00:46:58.398170 sudo[1828]: pam_unix(sudo:session): session closed for user root Jan 14 00:46:58.406366 sudo[1827]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Jan 14 00:46:58.406721 sudo[1827]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 14 00:46:58.416517 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 14 00:46:58.458000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Jan 14 00:46:58.460074 augenrules[1852]: No rules Jan 14 00:46:58.460528 kernel: kauditd_printk_skb: 74 callbacks suppressed Jan 14 00:46:58.460557 kernel: audit: type=1305 audit(1768351618.458:227): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Jan 14 00:46:58.466073 systemd[1]: audit-rules.service: Deactivated successfully. Jan 14 00:46:58.458000 audit[1852]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffe8566e7f0 a2=420 a3=0 items=0 ppid=1833 pid=1852 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:58.466389 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 14 00:46:58.471533 sudo[1827]: pam_unix(sudo:session): session closed for user root Jan 14 00:46:58.458000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 14 00:46:58.475659 kernel: audit: type=1300 audit(1768351618.458:227): arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffe8566e7f0 a2=420 a3=0 items=0 ppid=1833 pid=1852 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/bin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:58.475695 kernel: audit: type=1327 audit(1768351618.458:227): proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jan 14 00:46:58.466000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:58.480338 kernel: audit: type=1130 audit(1768351618.466:228): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:58.485151 kernel: audit: type=1131 audit(1768351618.466:229): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:58.466000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:58.470000 audit[1827]: USER_END pid=1827 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 00:46:58.493217 kernel: audit: type=1106 audit(1768351618.470:230): pid=1827 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 00:46:58.500137 kernel: audit: type=1104 audit(1768351618.470:231): pid=1827 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 00:46:58.470000 audit[1827]: CRED_DISP pid=1827 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 00:46:58.500218 sshd[1826]: Connection closed by 68.220.241.50 port 34768 Jan 14 00:46:58.500567 sshd-session[1822]: pam_unix(sshd:session): session closed for user core Jan 14 00:46:58.512698 kernel: audit: type=1106 audit(1768351618.502:232): pid=1822 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:46:58.502000 audit[1822]: USER_END pid=1822 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:46:58.505487 systemd[1]: sshd@5-172.236.110.177:22-68.220.241.50:34768.service: Deactivated successfully. Jan 14 00:46:58.507663 systemd-logind[1578]: Session 7 logged out. Waiting for processes to exit. Jan 14 00:46:58.508772 systemd[1]: session-7.scope: Deactivated successfully. Jan 14 00:46:58.511565 systemd-logind[1578]: Removed session 7. Jan 14 00:46:58.502000 audit[1822]: CRED_DISP pid=1822 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:46:58.505000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@5-172.236.110.177:22-68.220.241.50:34768 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:58.521481 kernel: audit: type=1104 audit(1768351618.502:233): pid=1822 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:46:58.521524 kernel: audit: type=1131 audit(1768351618.505:234): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@5-172.236.110.177:22-68.220.241.50:34768 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:58.538000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-172.236.110.177:22-68.220.241.50:34778 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:58.539646 systemd[1]: Started sshd@6-172.236.110.177:22-68.220.241.50:34778.service - OpenSSH per-connection server daemon (68.220.241.50:34778). Jan 14 00:46:58.699000 audit[1861]: USER_ACCT pid=1861 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:46:58.700831 sshd[1861]: Accepted publickey for core from 68.220.241.50 port 34778 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:46:58.701000 audit[1861]: CRED_ACQ pid=1861 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:46:58.701000 audit[1861]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffce185d590 a2=3 a3=0 items=0 ppid=1 pid=1861 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=8 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:58.701000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:46:58.702874 sshd-session[1861]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:46:58.709176 systemd-logind[1578]: New session 8 of user core. Jan 14 00:46:58.714247 systemd[1]: Started session-8.scope - Session 8 of User core. Jan 14 00:46:58.717000 audit[1861]: USER_START pid=1861 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:46:58.719000 audit[1865]: CRED_ACQ pid=1865 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:46:58.750000 audit[1866]: USER_ACCT pid=1866 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_unix,pam_faillock acct="core" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 00:46:58.752257 sudo[1866]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jan 14 00:46:58.751000 audit[1866]: CRED_REFR pid=1866 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 00:46:58.752635 sudo[1866]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 14 00:46:58.751000 audit[1866]: USER_START pid=1866 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 00:46:59.132290 systemd[1]: Starting docker.service - Docker Application Container Engine... Jan 14 00:46:59.160438 (dockerd)[1884]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jan 14 00:46:59.421780 dockerd[1884]: time="2026-01-14T00:46:59.421513534Z" level=info msg="Starting up" Jan 14 00:46:59.423642 dockerd[1884]: time="2026-01-14T00:46:59.423619972Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Jan 14 00:46:59.436204 dockerd[1884]: time="2026-01-14T00:46:59.436160926Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Jan 14 00:46:59.528871 dockerd[1884]: time="2026-01-14T00:46:59.528838015Z" level=info msg="Loading containers: start." Jan 14 00:46:59.541428 kernel: Initializing XFRM netlink socket Jan 14 00:46:59.606000 audit[1933]: NETFILTER_CFG table=nat:2 family=2 entries=2 op=nft_register_chain pid=1933 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.606000 audit[1933]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffda3afa4f0 a2=0 a3=0 items=0 ppid=1884 pid=1933 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.606000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Jan 14 00:46:59.609000 audit[1935]: NETFILTER_CFG table=filter:3 family=2 entries=2 op=nft_register_chain pid=1935 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.609000 audit[1935]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffcbe1088a0 a2=0 a3=0 items=0 ppid=1884 pid=1935 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.609000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Jan 14 00:46:59.611000 audit[1937]: NETFILTER_CFG table=filter:4 family=2 entries=1 op=nft_register_chain pid=1937 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.611000 audit[1937]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd4b7c7780 a2=0 a3=0 items=0 ppid=1884 pid=1937 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.611000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Jan 14 00:46:59.614000 audit[1939]: NETFILTER_CFG table=filter:5 family=2 entries=1 op=nft_register_chain pid=1939 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.614000 audit[1939]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff3ccbcf30 a2=0 a3=0 items=0 ppid=1884 pid=1939 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.614000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D425249444745 Jan 14 00:46:59.616000 audit[1941]: NETFILTER_CFG table=filter:6 family=2 entries=1 op=nft_register_chain pid=1941 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.616000 audit[1941]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffc0f5b7000 a2=0 a3=0 items=0 ppid=1884 pid=1941 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.616000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D4354 Jan 14 00:46:59.619000 audit[1943]: NETFILTER_CFG table=filter:7 family=2 entries=1 op=nft_register_chain pid=1943 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.619000 audit[1943]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7fff9f70f7e0 a2=0 a3=0 items=0 ppid=1884 pid=1943 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.619000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 14 00:46:59.623000 audit[1945]: NETFILTER_CFG table=filter:8 family=2 entries=1 op=nft_register_chain pid=1945 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.623000 audit[1945]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7fffcc164490 a2=0 a3=0 items=0 ppid=1884 pid=1945 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.623000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 14 00:46:59.626000 audit[1947]: NETFILTER_CFG table=nat:9 family=2 entries=2 op=nft_register_chain pid=1947 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.626000 audit[1947]: SYSCALL arch=c000003e syscall=46 success=yes exit=384 a0=3 a1=7ffe04459630 a2=0 a3=0 items=0 ppid=1884 pid=1947 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.626000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Jan 14 00:46:59.655000 audit[1950]: NETFILTER_CFG table=nat:10 family=2 entries=2 op=nft_register_chain pid=1950 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.655000 audit[1950]: SYSCALL arch=c000003e syscall=46 success=yes exit=472 a0=3 a1=7fffd8988b90 a2=0 a3=0 items=0 ppid=1884 pid=1950 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.655000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003132372E302E302E302F38 Jan 14 00:46:59.659000 audit[1952]: NETFILTER_CFG table=filter:11 family=2 entries=2 op=nft_register_chain pid=1952 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.659000 audit[1952]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffeef98a7b0 a2=0 a3=0 items=0 ppid=1884 pid=1952 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.659000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D464F5257415244 Jan 14 00:46:59.662000 audit[1954]: NETFILTER_CFG table=filter:12 family=2 entries=1 op=nft_register_rule pid=1954 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.662000 audit[1954]: SYSCALL arch=c000003e syscall=46 success=yes exit=236 a0=3 a1=7ffd9f7575b0 a2=0 a3=0 items=0 ppid=1884 pid=1954 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.662000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D425249444745 Jan 14 00:46:59.664000 audit[1956]: NETFILTER_CFG table=filter:13 family=2 entries=1 op=nft_register_rule pid=1956 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.664000 audit[1956]: SYSCALL arch=c000003e syscall=46 success=yes exit=248 a0=3 a1=7ffd62c663b0 a2=0 a3=0 items=0 ppid=1884 pid=1956 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.664000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 14 00:46:59.667000 audit[1958]: NETFILTER_CFG table=filter:14 family=2 entries=1 op=nft_register_rule pid=1958 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.667000 audit[1958]: SYSCALL arch=c000003e syscall=46 success=yes exit=232 a0=3 a1=7ffd723b8800 a2=0 a3=0 items=0 ppid=1884 pid=1958 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.667000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D4354 Jan 14 00:46:59.709000 audit[1988]: NETFILTER_CFG table=nat:15 family=10 entries=2 op=nft_register_chain pid=1988 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.709000 audit[1988]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffec9dd8420 a2=0 a3=0 items=0 ppid=1884 pid=1988 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.709000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Jan 14 00:46:59.712000 audit[1990]: NETFILTER_CFG table=filter:16 family=10 entries=2 op=nft_register_chain pid=1990 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.712000 audit[1990]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffd6ae90800 a2=0 a3=0 items=0 ppid=1884 pid=1990 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.712000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Jan 14 00:46:59.714000 audit[1992]: NETFILTER_CFG table=filter:17 family=10 entries=1 op=nft_register_chain pid=1992 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.714000 audit[1992]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffed93f6300 a2=0 a3=0 items=0 ppid=1884 pid=1992 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.714000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D464F5257415244 Jan 14 00:46:59.716000 audit[1994]: NETFILTER_CFG table=filter:18 family=10 entries=1 op=nft_register_chain pid=1994 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.716000 audit[1994]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe97912390 a2=0 a3=0 items=0 ppid=1884 pid=1994 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.716000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D425249444745 Jan 14 00:46:59.719000 audit[1996]: NETFILTER_CFG table=filter:19 family=10 entries=1 op=nft_register_chain pid=1996 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.719000 audit[1996]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7fffcfd9db20 a2=0 a3=0 items=0 ppid=1884 pid=1996 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.719000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D4354 Jan 14 00:46:59.721000 audit[1998]: NETFILTER_CFG table=filter:20 family=10 entries=1 op=nft_register_chain pid=1998 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.721000 audit[1998]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffccb272500 a2=0 a3=0 items=0 ppid=1884 pid=1998 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.721000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 14 00:46:59.723000 audit[2000]: NETFILTER_CFG table=filter:21 family=10 entries=1 op=nft_register_chain pid=2000 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.723000 audit[2000]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7ffeda23f440 a2=0 a3=0 items=0 ppid=1884 pid=2000 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.723000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 14 00:46:59.726000 audit[2002]: NETFILTER_CFG table=nat:22 family=10 entries=2 op=nft_register_chain pid=2002 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.726000 audit[2002]: SYSCALL arch=c000003e syscall=46 success=yes exit=384 a0=3 a1=7fff80a4d130 a2=0 a3=0 items=0 ppid=1884 pid=2002 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.726000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Jan 14 00:46:59.729000 audit[2004]: NETFILTER_CFG table=nat:23 family=10 entries=2 op=nft_register_chain pid=2004 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.729000 audit[2004]: SYSCALL arch=c000003e syscall=46 success=yes exit=484 a0=3 a1=7ffff75514d0 a2=0 a3=0 items=0 ppid=1884 pid=2004 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.729000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003A3A312F313238 Jan 14 00:46:59.732000 audit[2006]: NETFILTER_CFG table=filter:24 family=10 entries=2 op=nft_register_chain pid=2006 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.732000 audit[2006]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffd7fbd7470 a2=0 a3=0 items=0 ppid=1884 pid=2006 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.732000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D464F5257415244 Jan 14 00:46:59.734000 audit[2008]: NETFILTER_CFG table=filter:25 family=10 entries=1 op=nft_register_rule pid=2008 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.734000 audit[2008]: SYSCALL arch=c000003e syscall=46 success=yes exit=236 a0=3 a1=7ffe75320500 a2=0 a3=0 items=0 ppid=1884 pid=2008 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.734000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D425249444745 Jan 14 00:46:59.737000 audit[2010]: NETFILTER_CFG table=filter:26 family=10 entries=1 op=nft_register_rule pid=2010 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.737000 audit[2010]: SYSCALL arch=c000003e syscall=46 success=yes exit=248 a0=3 a1=7fff8a4008d0 a2=0 a3=0 items=0 ppid=1884 pid=2010 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.737000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Jan 14 00:46:59.740000 audit[2012]: NETFILTER_CFG table=filter:27 family=10 entries=1 op=nft_register_rule pid=2012 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.740000 audit[2012]: SYSCALL arch=c000003e syscall=46 success=yes exit=232 a0=3 a1=7fff16dcb480 a2=0 a3=0 items=0 ppid=1884 pid=2012 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.740000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900444F434B45522D464F5257415244002D6A00444F434B45522D4354 Jan 14 00:46:59.745000 audit[2017]: NETFILTER_CFG table=filter:28 family=2 entries=1 op=nft_register_chain pid=2017 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.745000 audit[2017]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffeeae22f90 a2=0 a3=0 items=0 ppid=1884 pid=2017 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.745000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Jan 14 00:46:59.748000 audit[2019]: NETFILTER_CFG table=filter:29 family=2 entries=1 op=nft_register_rule pid=2019 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.748000 audit[2019]: SYSCALL arch=c000003e syscall=46 success=yes exit=212 a0=3 a1=7fffc188f820 a2=0 a3=0 items=0 ppid=1884 pid=2019 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.748000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E Jan 14 00:46:59.751000 audit[2021]: NETFILTER_CFG table=filter:30 family=2 entries=1 op=nft_register_rule pid=2021 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.751000 audit[2021]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7ffe7dc12e50 a2=0 a3=0 items=0 ppid=1884 pid=2021 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.751000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Jan 14 00:46:59.753000 audit[2023]: NETFILTER_CFG table=filter:31 family=10 entries=1 op=nft_register_chain pid=2023 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.753000 audit[2023]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffe386435e0 a2=0 a3=0 items=0 ppid=1884 pid=2023 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.753000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Jan 14 00:46:59.755000 audit[2025]: NETFILTER_CFG table=filter:32 family=10 entries=1 op=nft_register_rule pid=2025 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.755000 audit[2025]: SYSCALL arch=c000003e syscall=46 success=yes exit=212 a0=3 a1=7ffd53de47c0 a2=0 a3=0 items=0 ppid=1884 pid=2025 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.755000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E Jan 14 00:46:59.758000 audit[2027]: NETFILTER_CFG table=filter:33 family=10 entries=1 op=nft_register_rule pid=2027 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:46:59.758000 audit[2027]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7ffe81ab1260 a2=0 a3=0 items=0 ppid=1884 pid=2027 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.758000 audit: PROCTITLE proctitle=2F7573722F62696E2F6970367461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Jan 14 00:46:59.777000 audit[2031]: NETFILTER_CFG table=nat:34 family=2 entries=2 op=nft_register_chain pid=2031 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.777000 audit[2031]: SYSCALL arch=c000003e syscall=46 success=yes exit=520 a0=3 a1=7ffecad35b50 a2=0 a3=0 items=0 ppid=1884 pid=2031 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.777000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900504F5354524F5554494E47002D73003137322E31372E302E302F31360000002D6F00646F636B657230002D6A004D415351554552414445 Jan 14 00:46:59.783000 audit[2033]: NETFILTER_CFG table=nat:35 family=2 entries=1 op=nft_register_rule pid=2033 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.783000 audit[2033]: SYSCALL arch=c000003e syscall=46 success=yes exit=288 a0=3 a1=7ffe4c1e7af0 a2=0 a3=0 items=0 ppid=1884 pid=2033 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.783000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D74006E6174002D4900444F434B4552002D6900646F636B657230002D6A0052455455524E Jan 14 00:46:59.794000 audit[2041]: NETFILTER_CFG table=filter:36 family=2 entries=1 op=nft_register_rule pid=2041 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.794000 audit[2041]: SYSCALL arch=c000003e syscall=46 success=yes exit=300 a0=3 a1=7ffe4af9f720 a2=0 a3=0 items=0 ppid=1884 pid=2041 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.794000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D464F5257415244002D6900646F636B657230002D6A00414343455054 Jan 14 00:46:59.805000 audit[2047]: NETFILTER_CFG table=filter:37 family=2 entries=1 op=nft_register_rule pid=2047 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.805000 audit[2047]: SYSCALL arch=c000003e syscall=46 success=yes exit=376 a0=3 a1=7ffdc7eb5840 a2=0 a3=0 items=0 ppid=1884 pid=2047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.805000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45520000002D6900646F636B657230002D6F00646F636B657230002D6A0044524F50 Jan 14 00:46:59.808000 audit[2049]: NETFILTER_CFG table=filter:38 family=2 entries=1 op=nft_register_rule pid=2049 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.808000 audit[2049]: SYSCALL arch=c000003e syscall=46 success=yes exit=512 a0=3 a1=7ffee7a93e90 a2=0 a3=0 items=0 ppid=1884 pid=2049 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.808000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D4354002D6F00646F636B657230002D6D00636F6E6E747261636B002D2D637473746174650052454C415445442C45535441424C4953484544002D6A00414343455054 Jan 14 00:46:59.811000 audit[2051]: NETFILTER_CFG table=filter:39 family=2 entries=1 op=nft_register_rule pid=2051 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.811000 audit[2051]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7fff1ce97460 a2=0 a3=0 items=0 ppid=1884 pid=2051 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.811000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D425249444745002D6F00646F636B657230002D6A00444F434B4552 Jan 14 00:46:59.814000 audit[2053]: NETFILTER_CFG table=filter:40 family=2 entries=1 op=nft_register_rule pid=2053 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.814000 audit[2053]: SYSCALL arch=c000003e syscall=46 success=yes exit=428 a0=3 a1=7ffc63e7f2c0 a2=0 a3=0 items=0 ppid=1884 pid=2053 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.814000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4100444F434B45522D49534F4C4154494F4E2D53544147452D31002D6900646F636B6572300000002D6F00646F636B657230002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D32 Jan 14 00:46:59.817000 audit[2055]: NETFILTER_CFG table=filter:41 family=2 entries=1 op=nft_register_rule pid=2055 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:46:59.817000 audit[2055]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffd4c53cb70 a2=0 a3=0 items=0 ppid=1884 pid=2055 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:46:59.817000 audit: PROCTITLE proctitle=2F7573722F62696E2F69707461626C6573002D2D77616974002D740066696C746572002D4900444F434B45522D49534F4C4154494F4E2D53544147452D32002D6F00646F636B657230002D6A0044524F50 Jan 14 00:46:59.818589 systemd-networkd[1504]: docker0: Link UP Jan 14 00:46:59.822579 dockerd[1884]: time="2026-01-14T00:46:59.822552908Z" level=info msg="Loading containers: done." Jan 14 00:46:59.839408 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck527771583-merged.mount: Deactivated successfully. Jan 14 00:46:59.841761 dockerd[1884]: time="2026-01-14T00:46:59.841728536Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jan 14 00:46:59.841832 dockerd[1884]: time="2026-01-14T00:46:59.841791651Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Jan 14 00:46:59.842073 dockerd[1884]: time="2026-01-14T00:46:59.842050531Z" level=info msg="Initializing buildkit" Jan 14 00:46:59.864234 dockerd[1884]: time="2026-01-14T00:46:59.864213627Z" level=info msg="Completed buildkit initialization" Jan 14 00:46:59.870552 dockerd[1884]: time="2026-01-14T00:46:59.870514633Z" level=info msg="Daemon has completed initialization" Jan 14 00:46:59.870703 systemd[1]: Started docker.service - Docker Application Container Engine. Jan 14 00:46:59.870000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=docker comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:46:59.871796 dockerd[1884]: time="2026-01-14T00:46:59.871337086Z" level=info msg="API listen on /run/docker.sock" Jan 14 00:47:00.846055 containerd[1611]: time="2026-01-14T00:47:00.845961085Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.7\"" Jan 14 00:47:01.941879 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4214932066.mount: Deactivated successfully. Jan 14 00:47:02.792140 containerd[1611]: time="2026-01-14T00:47:02.791585129Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:02.792856 containerd[1611]: time="2026-01-14T00:47:02.792619180Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.7: active requests=0, bytes read=28445968" Jan 14 00:47:02.794154 containerd[1611]: time="2026-01-14T00:47:02.793468168Z" level=info msg="ImageCreate event name:\"sha256:021d1ceeffb11df7a9fb9adfa0ad0a30dcd13cb3d630022066f184cdcb93731b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:02.796388 containerd[1611]: time="2026-01-14T00:47:02.796354881Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:9585226cb85d1dc0f0ef5f7a75f04e4bc91ddd82de249533bd293aa3cf958dab\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:02.797548 containerd[1611]: time="2026-01-14T00:47:02.797513947Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.7\" with image id \"sha256:021d1ceeffb11df7a9fb9adfa0ad0a30dcd13cb3d630022066f184cdcb93731b\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.7\", repo digest \"registry.k8s.io/kube-apiserver@sha256:9585226cb85d1dc0f0ef5f7a75f04e4bc91ddd82de249533bd293aa3cf958dab\", size \"30111311\" in 1.951480418s" Jan 14 00:47:02.797594 containerd[1611]: time="2026-01-14T00:47:02.797551905Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.7\" returns image reference \"sha256:021d1ceeffb11df7a9fb9adfa0ad0a30dcd13cb3d630022066f184cdcb93731b\"" Jan 14 00:47:02.798237 containerd[1611]: time="2026-01-14T00:47:02.798214924Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.7\"" Jan 14 00:47:04.303282 containerd[1611]: time="2026-01-14T00:47:04.303218928Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:04.304321 containerd[1611]: time="2026-01-14T00:47:04.304125362Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.7: active requests=0, bytes read=26008626" Jan 14 00:47:04.305824 containerd[1611]: time="2026-01-14T00:47:04.305798312Z" level=info msg="ImageCreate event name:\"sha256:29c7cab9d8e681d047281fd3711baf13c28f66923480fb11c8f22ddb7ca742d1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:04.309181 containerd[1611]: time="2026-01-14T00:47:04.309159156Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:f69d77ca0626b5a4b7b432c18de0952941181db7341c80eb89731f46d1d0c230\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:04.309977 containerd[1611]: time="2026-01-14T00:47:04.309947846Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.7\" with image id \"sha256:29c7cab9d8e681d047281fd3711baf13c28f66923480fb11c8f22ddb7ca742d1\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.7\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:f69d77ca0626b5a4b7b432c18de0952941181db7341c80eb89731f46d1d0c230\", size \"27673815\" in 1.511707685s" Jan 14 00:47:04.310019 containerd[1611]: time="2026-01-14T00:47:04.309980031Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.7\" returns image reference \"sha256:29c7cab9d8e681d047281fd3711baf13c28f66923480fb11c8f22ddb7ca742d1\"" Jan 14 00:47:04.310563 containerd[1611]: time="2026-01-14T00:47:04.310536567Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.7\"" Jan 14 00:47:05.671215 containerd[1611]: time="2026-01-14T00:47:05.671146610Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:05.672721 containerd[1611]: time="2026-01-14T00:47:05.672283005Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.7: active requests=0, bytes read=20149965" Jan 14 00:47:05.673216 containerd[1611]: time="2026-01-14T00:47:05.673182007Z" level=info msg="ImageCreate event name:\"sha256:f457f6fcd712acb5b9beef873f6f4a4869182f9eb52ea6e24824fd4ac4eed393\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:05.675600 containerd[1611]: time="2026-01-14T00:47:05.675571848Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:21bda321d8b4d48eb059fbc1593203d55d8b3bc7acd0584e04e55504796d78d0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:05.676695 containerd[1611]: time="2026-01-14T00:47:05.676674107Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.7\" with image id \"sha256:f457f6fcd712acb5b9beef873f6f4a4869182f9eb52ea6e24824fd4ac4eed393\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.7\", repo digest \"registry.k8s.io/kube-scheduler@sha256:21bda321d8b4d48eb059fbc1593203d55d8b3bc7acd0584e04e55504796d78d0\", size \"21815154\" in 1.366107598s" Jan 14 00:47:05.676781 containerd[1611]: time="2026-01-14T00:47:05.676765764Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.7\" returns image reference \"sha256:f457f6fcd712acb5b9beef873f6f4a4869182f9eb52ea6e24824fd4ac4eed393\"" Jan 14 00:47:05.677480 containerd[1611]: time="2026-01-14T00:47:05.677454133Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.7\"" Jan 14 00:47:06.417708 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jan 14 00:47:06.419919 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 00:47:06.627355 kernel: kauditd_printk_skb: 132 callbacks suppressed Jan 14 00:47:06.627632 kernel: audit: type=1130 audit(1768351626.625:285): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:47:06.625000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:47:06.626299 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 00:47:06.645932 (kubelet)[2170]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 14 00:47:06.705862 kubelet[2170]: E0114 00:47:06.705691 2170 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 14 00:47:06.714935 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 14 00:47:06.715147 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 14 00:47:06.714000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 00:47:06.715946 systemd[1]: kubelet.service: Consumed 206ms CPU time, 110.5M memory peak. Jan 14 00:47:06.723152 kernel: audit: type=1131 audit(1768351626.714:286): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 00:47:06.974981 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2493709863.mount: Deactivated successfully. Jan 14 00:47:07.347753 containerd[1611]: time="2026-01-14T00:47:07.347434227Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:07.348925 containerd[1611]: time="2026-01-14T00:47:07.348536769Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.7: active requests=0, bytes read=20340589" Jan 14 00:47:07.350092 containerd[1611]: time="2026-01-14T00:47:07.350055224Z" level=info msg="ImageCreate event name:\"sha256:0929027b17fc30cb9de279f3bdba4e130b991a1dab7978a7db2e5feb2091853c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:07.352096 containerd[1611]: time="2026-01-14T00:47:07.352064510Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:ec25702b19026e9c0d339bc1c3bd231435a59f28b5fccb21e1b1078a357380f5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:07.352717 containerd[1611]: time="2026-01-14T00:47:07.352683071Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.7\" with image id \"sha256:0929027b17fc30cb9de279f3bdba4e130b991a1dab7978a7db2e5feb2091853c\", repo tag \"registry.k8s.io/kube-proxy:v1.33.7\", repo digest \"registry.k8s.io/kube-proxy@sha256:ec25702b19026e9c0d339bc1c3bd231435a59f28b5fccb21e1b1078a357380f5\", size \"31929115\" in 1.675194811s" Jan 14 00:47:07.352811 containerd[1611]: time="2026-01-14T00:47:07.352794715Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.7\" returns image reference \"sha256:0929027b17fc30cb9de279f3bdba4e130b991a1dab7978a7db2e5feb2091853c\"" Jan 14 00:47:07.353487 containerd[1611]: time="2026-01-14T00:47:07.353403690Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Jan 14 00:47:08.057867 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1763535490.mount: Deactivated successfully. Jan 14 00:47:08.764577 containerd[1611]: time="2026-01-14T00:47:08.764290941Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:08.765591 containerd[1611]: time="2026-01-14T00:47:08.765496592Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20128467" Jan 14 00:47:08.766249 containerd[1611]: time="2026-01-14T00:47:08.766191976Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:08.768865 containerd[1611]: time="2026-01-14T00:47:08.768821635Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:08.770094 containerd[1611]: time="2026-01-14T00:47:08.769961964Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 1.416389293s" Jan 14 00:47:08.770094 containerd[1611]: time="2026-01-14T00:47:08.769999878Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Jan 14 00:47:08.771217 containerd[1611]: time="2026-01-14T00:47:08.771184326Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Jan 14 00:47:09.445239 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1528444369.mount: Deactivated successfully. Jan 14 00:47:09.450058 containerd[1611]: time="2026-01-14T00:47:09.449345056Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 14 00:47:09.450058 containerd[1611]: time="2026-01-14T00:47:09.450035013Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" Jan 14 00:47:09.450572 containerd[1611]: time="2026-01-14T00:47:09.450548641Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 14 00:47:09.452692 containerd[1611]: time="2026-01-14T00:47:09.452669274Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 14 00:47:09.453652 containerd[1611]: time="2026-01-14T00:47:09.453601354Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 682.387909ms" Jan 14 00:47:09.453706 containerd[1611]: time="2026-01-14T00:47:09.453656062Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Jan 14 00:47:09.454803 containerd[1611]: time="2026-01-14T00:47:09.454765214Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Jan 14 00:47:10.108081 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2861763128.mount: Deactivated successfully. Jan 14 00:47:11.974198 containerd[1611]: time="2026-01-14T00:47:11.974094881Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:11.975581 containerd[1611]: time="2026-01-14T00:47:11.975488572Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58133605" Jan 14 00:47:11.976179 containerd[1611]: time="2026-01-14T00:47:11.976141311Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:11.979468 containerd[1611]: time="2026-01-14T00:47:11.979417111Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:11.980681 containerd[1611]: time="2026-01-14T00:47:11.980536038Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 2.525741145s" Jan 14 00:47:11.980681 containerd[1611]: time="2026-01-14T00:47:11.980563574Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Jan 14 00:47:15.596994 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 00:47:15.596000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:47:15.597940 systemd[1]: kubelet.service: Consumed 206ms CPU time, 110.5M memory peak. Jan 14 00:47:15.605526 kernel: audit: type=1130 audit(1768351635.596:287): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:47:15.605589 kernel: audit: type=1131 audit(1768351635.596:288): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:47:15.596000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:47:15.603322 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 00:47:15.644251 systemd[1]: Reload requested from client PID 2321 ('systemctl') (unit session-8.scope)... Jan 14 00:47:15.644271 systemd[1]: Reloading... Jan 14 00:47:15.812155 zram_generator::config[2383]: No configuration found. Jan 14 00:47:16.024284 systemd[1]: Reloading finished in 379 ms. Jan 14 00:47:16.066000 audit: BPF prog-id=67 op=LOAD Jan 14 00:47:16.071489 kernel: audit: type=1334 audit(1768351636.066:289): prog-id=67 op=LOAD Jan 14 00:47:16.069000 audit: BPF prog-id=59 op=UNLOAD Jan 14 00:47:16.071000 audit: BPF prog-id=68 op=LOAD Jan 14 00:47:16.076264 kernel: audit: type=1334 audit(1768351636.069:290): prog-id=59 op=UNLOAD Jan 14 00:47:16.076310 kernel: audit: type=1334 audit(1768351636.071:291): prog-id=68 op=LOAD Jan 14 00:47:16.071000 audit: BPF prog-id=50 op=UNLOAD Jan 14 00:47:16.078686 kernel: audit: type=1334 audit(1768351636.071:292): prog-id=50 op=UNLOAD Jan 14 00:47:16.071000 audit: BPF prog-id=69 op=LOAD Jan 14 00:47:16.083152 kernel: audit: type=1334 audit(1768351636.071:293): prog-id=69 op=LOAD Jan 14 00:47:16.083205 kernel: audit: type=1334 audit(1768351636.071:294): prog-id=70 op=LOAD Jan 14 00:47:16.071000 audit: BPF prog-id=70 op=LOAD Jan 14 00:47:16.071000 audit: BPF prog-id=51 op=UNLOAD Jan 14 00:47:16.090148 kernel: audit: type=1334 audit(1768351636.071:295): prog-id=51 op=UNLOAD Jan 14 00:47:16.071000 audit: BPF prog-id=52 op=UNLOAD Jan 14 00:47:16.081000 audit: BPF prog-id=71 op=LOAD Jan 14 00:47:16.081000 audit: BPF prog-id=58 op=UNLOAD Jan 14 00:47:16.083000 audit: BPF prog-id=72 op=LOAD Jan 14 00:47:16.083000 audit: BPF prog-id=43 op=UNLOAD Jan 14 00:47:16.084000 audit: BPF prog-id=73 op=LOAD Jan 14 00:47:16.084000 audit: BPF prog-id=47 op=UNLOAD Jan 14 00:47:16.084000 audit: BPF prog-id=74 op=LOAD Jan 14 00:47:16.084000 audit: BPF prog-id=75 op=LOAD Jan 14 00:47:16.084000 audit: BPF prog-id=48 op=UNLOAD Jan 14 00:47:16.084000 audit: BPF prog-id=49 op=UNLOAD Jan 14 00:47:16.084000 audit: BPF prog-id=76 op=LOAD Jan 14 00:47:16.084000 audit: BPF prog-id=77 op=LOAD Jan 14 00:47:16.084000 audit: BPF prog-id=56 op=UNLOAD Jan 14 00:47:16.084000 audit: BPF prog-id=57 op=UNLOAD Jan 14 00:47:16.085000 audit: BPF prog-id=78 op=LOAD Jan 14 00:47:16.085000 audit: BPF prog-id=44 op=UNLOAD Jan 14 00:47:16.085000 audit: BPF prog-id=79 op=LOAD Jan 14 00:47:16.093192 kernel: audit: type=1334 audit(1768351636.071:296): prog-id=52 op=UNLOAD Jan 14 00:47:16.085000 audit: BPF prog-id=80 op=LOAD Jan 14 00:47:16.085000 audit: BPF prog-id=45 op=UNLOAD Jan 14 00:47:16.085000 audit: BPF prog-id=46 op=UNLOAD Jan 14 00:47:16.086000 audit: BPF prog-id=81 op=LOAD Jan 14 00:47:16.086000 audit: BPF prog-id=63 op=UNLOAD Jan 14 00:47:16.086000 audit: BPF prog-id=82 op=LOAD Jan 14 00:47:16.086000 audit: BPF prog-id=83 op=LOAD Jan 14 00:47:16.086000 audit: BPF prog-id=64 op=UNLOAD Jan 14 00:47:16.086000 audit: BPF prog-id=65 op=UNLOAD Jan 14 00:47:16.088000 audit: BPF prog-id=84 op=LOAD Jan 14 00:47:16.088000 audit: BPF prog-id=66 op=UNLOAD Jan 14 00:47:16.089000 audit: BPF prog-id=85 op=LOAD Jan 14 00:47:16.089000 audit: BPF prog-id=53 op=UNLOAD Jan 14 00:47:16.089000 audit: BPF prog-id=86 op=LOAD Jan 14 00:47:16.089000 audit: BPF prog-id=87 op=LOAD Jan 14 00:47:16.089000 audit: BPF prog-id=54 op=UNLOAD Jan 14 00:47:16.089000 audit: BPF prog-id=55 op=UNLOAD Jan 14 00:47:16.093000 audit: BPF prog-id=88 op=LOAD Jan 14 00:47:16.093000 audit: BPF prog-id=60 op=UNLOAD Jan 14 00:47:16.093000 audit: BPF prog-id=89 op=LOAD Jan 14 00:47:16.093000 audit: BPF prog-id=90 op=LOAD Jan 14 00:47:16.093000 audit: BPF prog-id=61 op=UNLOAD Jan 14 00:47:16.093000 audit: BPF prog-id=62 op=UNLOAD Jan 14 00:47:16.110321 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jan 14 00:47:16.110420 systemd[1]: kubelet.service: Failed with result 'signal'. Jan 14 00:47:16.111029 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 00:47:16.111190 systemd[1]: kubelet.service: Consumed 153ms CPU time, 98.5M memory peak. Jan 14 00:47:16.110000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Jan 14 00:47:16.115825 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 00:47:16.289382 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 00:47:16.288000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:47:16.303471 (kubelet)[2421]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 14 00:47:16.352168 kubelet[2421]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 14 00:47:16.354142 kubelet[2421]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jan 14 00:47:16.354142 kubelet[2421]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 14 00:47:16.354142 kubelet[2421]: I0114 00:47:16.352699 2421 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 14 00:47:16.523047 kubelet[2421]: I0114 00:47:16.523004 2421 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Jan 14 00:47:16.523221 kubelet[2421]: I0114 00:47:16.523205 2421 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 14 00:47:16.523801 kubelet[2421]: I0114 00:47:16.523781 2421 server.go:956] "Client rotation is on, will bootstrap in background" Jan 14 00:47:16.554580 kubelet[2421]: I0114 00:47:16.553701 2421 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 14 00:47:16.555149 kubelet[2421]: E0114 00:47:16.555084 2421 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://172.236.110.177:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.236.110.177:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jan 14 00:47:16.566460 kubelet[2421]: I0114 00:47:16.566366 2421 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 14 00:47:16.573046 kubelet[2421]: I0114 00:47:16.573018 2421 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 14 00:47:16.573478 kubelet[2421]: I0114 00:47:16.573417 2421 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 14 00:47:16.573734 kubelet[2421]: I0114 00:47:16.573461 2421 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"172-236-110-177","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 14 00:47:16.573734 kubelet[2421]: I0114 00:47:16.573713 2421 topology_manager.go:138] "Creating topology manager with none policy" Jan 14 00:47:16.573734 kubelet[2421]: I0114 00:47:16.573729 2421 container_manager_linux.go:303] "Creating device plugin manager" Jan 14 00:47:16.573997 kubelet[2421]: I0114 00:47:16.573973 2421 state_mem.go:36] "Initialized new in-memory state store" Jan 14 00:47:16.578177 kubelet[2421]: I0114 00:47:16.577717 2421 kubelet.go:480] "Attempting to sync node with API server" Jan 14 00:47:16.578177 kubelet[2421]: I0114 00:47:16.577747 2421 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 14 00:47:16.578177 kubelet[2421]: I0114 00:47:16.577783 2421 kubelet.go:386] "Adding apiserver pod source" Jan 14 00:47:16.580314 kubelet[2421]: I0114 00:47:16.580292 2421 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 14 00:47:16.588718 kubelet[2421]: E0114 00:47:16.588085 2421 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://172.236.110.177:6443/api/v1/nodes?fieldSelector=metadata.name%3D172-236-110-177&limit=500&resourceVersion=0\": dial tcp 172.236.110.177:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jan 14 00:47:16.588718 kubelet[2421]: I0114 00:47:16.588284 2421 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.1.5" apiVersion="v1" Jan 14 00:47:16.589027 kubelet[2421]: I0114 00:47:16.588975 2421 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Jan 14 00:47:16.590398 kubelet[2421]: W0114 00:47:16.590349 2421 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jan 14 00:47:16.600143 kubelet[2421]: I0114 00:47:16.599530 2421 watchdog_linux.go:99] "Systemd watchdog is not enabled" Jan 14 00:47:16.600143 kubelet[2421]: I0114 00:47:16.599592 2421 server.go:1289] "Started kubelet" Jan 14 00:47:16.603612 kubelet[2421]: I0114 00:47:16.603591 2421 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 14 00:47:16.604976 kubelet[2421]: I0114 00:47:16.604931 2421 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Jan 14 00:47:16.608017 kubelet[2421]: E0114 00:47:16.607986 2421 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://172.236.110.177:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.236.110.177:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jan 14 00:47:16.611930 kubelet[2421]: I0114 00:47:16.611852 2421 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 14 00:47:16.612687 kubelet[2421]: I0114 00:47:16.612649 2421 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 14 00:47:16.613037 kubelet[2421]: I0114 00:47:16.612999 2421 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 14 00:47:16.616148 kubelet[2421]: I0114 00:47:16.615830 2421 volume_manager.go:297] "Starting Kubelet Volume Manager" Jan 14 00:47:16.616148 kubelet[2421]: E0114 00:47:16.616062 2421 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172-236-110-177\" not found" Jan 14 00:47:16.616720 kubelet[2421]: I0114 00:47:16.616688 2421 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Jan 14 00:47:16.616778 kubelet[2421]: I0114 00:47:16.616771 2421 reconciler.go:26] "Reconciler: start to sync state" Jan 14 00:47:16.617000 audit[2435]: NETFILTER_CFG table=mangle:42 family=2 entries=2 op=nft_register_chain pid=2435 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:16.617000 audit[2435]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffc7fa11380 a2=0 a3=0 items=0 ppid=2421 pid=2435 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:16.617000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Jan 14 00:47:16.619000 audit[2438]: NETFILTER_CFG table=filter:43 family=2 entries=1 op=nft_register_chain pid=2438 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:16.619000 audit[2438]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff9a0e98b0 a2=0 a3=0 items=0 ppid=2421 pid=2438 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:16.619000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 Jan 14 00:47:16.621520 kubelet[2421]: E0114 00:47:16.621478 2421 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://172.236.110.177:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.236.110.177:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jan 14 00:47:16.621636 kubelet[2421]: E0114 00:47:16.621571 2421 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.236.110.177:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172-236-110-177?timeout=10s\": dial tcp 172.236.110.177:6443: connect: connection refused" interval="200ms" Jan 14 00:47:16.623855 kubelet[2421]: E0114 00:47:16.621664 2421 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.236.110.177:6443/api/v1/namespaces/default/events\": dial tcp 172.236.110.177:6443: connect: connection refused" event="&Event{ObjectMeta:{172-236-110-177.188a7271c17f45b2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:172-236-110-177,UID:172-236-110-177,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:172-236-110-177,},FirstTimestamp:2026-01-14 00:47:16.599555506 +0000 UTC m=+0.291600986,LastTimestamp:2026-01-14 00:47:16.599555506 +0000 UTC m=+0.291600986,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172-236-110-177,}" Jan 14 00:47:16.623982 kubelet[2421]: I0114 00:47:16.623879 2421 server.go:317] "Adding debug handlers to kubelet server" Jan 14 00:47:16.624638 kubelet[2421]: I0114 00:47:16.624384 2421 factory.go:223] Registration of the systemd container factory successfully Jan 14 00:47:16.624638 kubelet[2421]: I0114 00:47:16.624504 2421 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 14 00:47:16.626000 audit[2440]: NETFILTER_CFG table=filter:44 family=2 entries=2 op=nft_register_chain pid=2440 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:16.627993 kubelet[2421]: I0114 00:47:16.627928 2421 factory.go:223] Registration of the containerd container factory successfully Jan 14 00:47:16.628187 kubelet[2421]: E0114 00:47:16.628171 2421 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 14 00:47:16.626000 audit[2440]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffc7e245cb0 a2=0 a3=0 items=0 ppid=2421 pid=2440 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:16.626000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 14 00:47:16.636000 audit[2445]: NETFILTER_CFG table=filter:45 family=2 entries=2 op=nft_register_chain pid=2445 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:16.636000 audit[2445]: SYSCALL arch=c000003e syscall=46 success=yes exit=340 a0=3 a1=7ffd7932ffa0 a2=0 a3=0 items=0 ppid=2421 pid=2445 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:16.636000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 14 00:47:16.647000 audit[2448]: NETFILTER_CFG table=filter:46 family=2 entries=1 op=nft_register_rule pid=2448 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:16.647000 audit[2448]: SYSCALL arch=c000003e syscall=46 success=yes exit=924 a0=3 a1=7fff23e18450 a2=0 a3=0 items=0 ppid=2421 pid=2448 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:16.647000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D41004B5542452D4649524557414C4C002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E7400626C6F636B20696E636F6D696E67206C6F63616C6E657420636F6E6E656374696F6E73002D2D647374003132372E302E302E302F38 Jan 14 00:47:16.649072 kubelet[2421]: I0114 00:47:16.649034 2421 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Jan 14 00:47:16.649000 audit[2450]: NETFILTER_CFG table=mangle:47 family=2 entries=1 op=nft_register_chain pid=2450 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:16.649000 audit[2450]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffc60c5c970 a2=0 a3=0 items=0 ppid=2421 pid=2450 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:16.649000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Jan 14 00:47:16.651000 audit[2453]: NETFILTER_CFG table=nat:48 family=2 entries=1 op=nft_register_chain pid=2453 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:16.651000 audit[2453]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fffa424a900 a2=0 a3=0 items=0 ppid=2421 pid=2453 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:16.651000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Jan 14 00:47:16.653883 kubelet[2421]: I0114 00:47:16.653861 2421 cpu_manager.go:221] "Starting CPU manager" policy="none" Jan 14 00:47:16.653883 kubelet[2421]: I0114 00:47:16.653882 2421 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jan 14 00:47:16.653946 kubelet[2421]: I0114 00:47:16.653902 2421 state_mem.go:36] "Initialized new in-memory state store" Jan 14 00:47:16.654000 audit[2454]: NETFILTER_CFG table=filter:49 family=2 entries=1 op=nft_register_chain pid=2454 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:16.654000 audit[2454]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffdf11ba4f0 a2=0 a3=0 items=0 ppid=2421 pid=2454 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:16.654000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Jan 14 00:47:16.654000 audit[2452]: NETFILTER_CFG table=mangle:50 family=10 entries=2 op=nft_register_chain pid=2452 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:16.654000 audit[2452]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffc60bb4530 a2=0 a3=0 items=0 ppid=2421 pid=2452 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:16.654000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Jan 14 00:47:16.655970 kubelet[2421]: I0114 00:47:16.655948 2421 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Jan 14 00:47:16.656014 kubelet[2421]: I0114 00:47:16.655989 2421 status_manager.go:230] "Starting to sync pod status with apiserver" Jan 14 00:47:16.655000 audit[2455]: NETFILTER_CFG table=mangle:51 family=10 entries=1 op=nft_register_chain pid=2455 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:16.657180 kubelet[2421]: I0114 00:47:16.656442 2421 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jan 14 00:47:16.655000 audit[2455]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffd58de6cc0 a2=0 a3=0 items=0 ppid=2421 pid=2455 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:16.655000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Jan 14 00:47:16.658181 kubelet[2421]: I0114 00:47:16.658165 2421 kubelet.go:2436] "Starting kubelet main sync loop" Jan 14 00:47:16.658497 kubelet[2421]: E0114 00:47:16.658295 2421 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 14 00:47:16.658497 kubelet[2421]: E0114 00:47:16.658098 2421 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://172.236.110.177:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.236.110.177:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 14 00:47:16.658497 kubelet[2421]: I0114 00:47:16.658379 2421 policy_none.go:49] "None policy: Start" Jan 14 00:47:16.658497 kubelet[2421]: I0114 00:47:16.658439 2421 memory_manager.go:186] "Starting memorymanager" policy="None" Jan 14 00:47:16.658497 kubelet[2421]: I0114 00:47:16.658455 2421 state_mem.go:35] "Initializing new in-memory state store" Jan 14 00:47:16.660000 audit[2456]: NETFILTER_CFG table=nat:52 family=10 entries=1 op=nft_register_chain pid=2456 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:16.660000 audit[2456]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffeb2524a40 a2=0 a3=0 items=0 ppid=2421 pid=2456 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:16.660000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Jan 14 00:47:16.662000 audit[2457]: NETFILTER_CFG table=filter:53 family=10 entries=1 op=nft_register_chain pid=2457 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:16.662000 audit[2457]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffc345c2280 a2=0 a3=0 items=0 ppid=2421 pid=2457 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/bin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:16.662000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Jan 14 00:47:16.668036 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jan 14 00:47:16.677945 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jan 14 00:47:16.698813 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jan 14 00:47:16.700883 kubelet[2421]: E0114 00:47:16.700830 2421 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Jan 14 00:47:16.701137 kubelet[2421]: I0114 00:47:16.701008 2421 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 14 00:47:16.701137 kubelet[2421]: I0114 00:47:16.701024 2421 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 14 00:47:16.701715 kubelet[2421]: I0114 00:47:16.701694 2421 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 14 00:47:16.703293 kubelet[2421]: E0114 00:47:16.703278 2421 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jan 14 00:47:16.703436 kubelet[2421]: E0114 00:47:16.703419 2421 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"172-236-110-177\" not found" Jan 14 00:47:16.773776 systemd[1]: Created slice kubepods-burstable-pod8bbf6446f3f4d6255fd3ad0a0405f236.slice - libcontainer container kubepods-burstable-pod8bbf6446f3f4d6255fd3ad0a0405f236.slice. Jan 14 00:47:16.786618 kubelet[2421]: E0114 00:47:16.786487 2421 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-236-110-177\" not found" node="172-236-110-177" Jan 14 00:47:16.789834 systemd[1]: Created slice kubepods-burstable-pod3ae0af65e6e9e55461a48715990c01e7.slice - libcontainer container kubepods-burstable-pod3ae0af65e6e9e55461a48715990c01e7.slice. Jan 14 00:47:16.802428 kubelet[2421]: E0114 00:47:16.801649 2421 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-236-110-177\" not found" node="172-236-110-177" Jan 14 00:47:16.807162 kubelet[2421]: I0114 00:47:16.806243 2421 kubelet_node_status.go:75] "Attempting to register node" node="172-236-110-177" Jan 14 00:47:16.807162 kubelet[2421]: E0114 00:47:16.806591 2421 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.236.110.177:6443/api/v1/nodes\": dial tcp 172.236.110.177:6443: connect: connection refused" node="172-236-110-177" Jan 14 00:47:16.810672 systemd[1]: Created slice kubepods-burstable-podf6a3c8bb6d69c8c4320f01549746566a.slice - libcontainer container kubepods-burstable-podf6a3c8bb6d69c8c4320f01549746566a.slice. Jan 14 00:47:16.814513 kubelet[2421]: E0114 00:47:16.814488 2421 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-236-110-177\" not found" node="172-236-110-177" Jan 14 00:47:16.822060 kubelet[2421]: E0114 00:47:16.822016 2421 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.236.110.177:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172-236-110-177?timeout=10s\": dial tcp 172.236.110.177:6443: connect: connection refused" interval="400ms" Jan 14 00:47:16.917768 kubelet[2421]: I0114 00:47:16.917667 2421 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f6a3c8bb6d69c8c4320f01549746566a-k8s-certs\") pod \"kube-apiserver-172-236-110-177\" (UID: \"f6a3c8bb6d69c8c4320f01549746566a\") " pod="kube-system/kube-apiserver-172-236-110-177" Jan 14 00:47:16.917768 kubelet[2421]: I0114 00:47:16.917742 2421 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8bbf6446f3f4d6255fd3ad0a0405f236-ca-certs\") pod \"kube-controller-manager-172-236-110-177\" (UID: \"8bbf6446f3f4d6255fd3ad0a0405f236\") " pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:16.917768 kubelet[2421]: I0114 00:47:16.917782 2421 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8bbf6446f3f4d6255fd3ad0a0405f236-usr-share-ca-certificates\") pod \"kube-controller-manager-172-236-110-177\" (UID: \"8bbf6446f3f4d6255fd3ad0a0405f236\") " pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:16.918044 kubelet[2421]: I0114 00:47:16.917822 2421 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f6a3c8bb6d69c8c4320f01549746566a-ca-certs\") pod \"kube-apiserver-172-236-110-177\" (UID: \"f6a3c8bb6d69c8c4320f01549746566a\") " pod="kube-system/kube-apiserver-172-236-110-177" Jan 14 00:47:16.918044 kubelet[2421]: I0114 00:47:16.917850 2421 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f6a3c8bb6d69c8c4320f01549746566a-usr-share-ca-certificates\") pod \"kube-apiserver-172-236-110-177\" (UID: \"f6a3c8bb6d69c8c4320f01549746566a\") " pod="kube-system/kube-apiserver-172-236-110-177" Jan 14 00:47:16.918044 kubelet[2421]: I0114 00:47:16.917884 2421 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/8bbf6446f3f4d6255fd3ad0a0405f236-flexvolume-dir\") pod \"kube-controller-manager-172-236-110-177\" (UID: \"8bbf6446f3f4d6255fd3ad0a0405f236\") " pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:16.918044 kubelet[2421]: I0114 00:47:16.917905 2421 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8bbf6446f3f4d6255fd3ad0a0405f236-k8s-certs\") pod \"kube-controller-manager-172-236-110-177\" (UID: \"8bbf6446f3f4d6255fd3ad0a0405f236\") " pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:16.918044 kubelet[2421]: I0114 00:47:16.917932 2421 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8bbf6446f3f4d6255fd3ad0a0405f236-kubeconfig\") pod \"kube-controller-manager-172-236-110-177\" (UID: \"8bbf6446f3f4d6255fd3ad0a0405f236\") " pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:16.918219 kubelet[2421]: I0114 00:47:16.917952 2421 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/3ae0af65e6e9e55461a48715990c01e7-kubeconfig\") pod \"kube-scheduler-172-236-110-177\" (UID: \"3ae0af65e6e9e55461a48715990c01e7\") " pod="kube-system/kube-scheduler-172-236-110-177" Jan 14 00:47:17.010137 kubelet[2421]: I0114 00:47:17.010041 2421 kubelet_node_status.go:75] "Attempting to register node" node="172-236-110-177" Jan 14 00:47:17.010706 kubelet[2421]: E0114 00:47:17.010659 2421 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.236.110.177:6443/api/v1/nodes\": dial tcp 172.236.110.177:6443: connect: connection refused" node="172-236-110-177" Jan 14 00:47:17.089243 kubelet[2421]: E0114 00:47:17.087929 2421 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:17.090546 containerd[1611]: time="2026-01-14T00:47:17.090270454Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-172-236-110-177,Uid:8bbf6446f3f4d6255fd3ad0a0405f236,Namespace:kube-system,Attempt:0,}" Jan 14 00:47:17.102756 kubelet[2421]: E0114 00:47:17.102636 2421 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:17.104255 containerd[1611]: time="2026-01-14T00:47:17.104159501Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-172-236-110-177,Uid:3ae0af65e6e9e55461a48715990c01e7,Namespace:kube-system,Attempt:0,}" Jan 14 00:47:17.115875 kubelet[2421]: E0114 00:47:17.115831 2421 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:17.116302 containerd[1611]: time="2026-01-14T00:47:17.116280177Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-172-236-110-177,Uid:f6a3c8bb6d69c8c4320f01549746566a,Namespace:kube-system,Attempt:0,}" Jan 14 00:47:17.213142 containerd[1611]: time="2026-01-14T00:47:17.212861247Z" level=info msg="connecting to shim 2a26505f4183945daf9f1bceacb2cd11dbd324f37305a70db4b19964066d740c" address="unix:///run/containerd/s/2a87a64a0c81b43875ef460052fe9bb0e33b6ecfa31c66d27e8a2ed196e36ad4" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:47:17.222433 kubelet[2421]: E0114 00:47:17.222400 2421 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.236.110.177:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172-236-110-177?timeout=10s\": dial tcp 172.236.110.177:6443: connect: connection refused" interval="800ms" Jan 14 00:47:17.228146 containerd[1611]: time="2026-01-14T00:47:17.227272871Z" level=info msg="connecting to shim cfddbdad0a89c13bc1e11c07f1396bd1437f8dec977cfba7bed4ee93826eb4df" address="unix:///run/containerd/s/2431b978943890a23fd50b031b4f3994c50fcd7d6313d1c1de77403ea8b73ad8" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:47:17.248492 containerd[1611]: time="2026-01-14T00:47:17.248460308Z" level=info msg="connecting to shim d1f6835108a3019ca54fb4685084e4889cbdaa910583795c01d98fa0dd9c1519" address="unix:///run/containerd/s/4ea470c45a4002b98c3756d4f05672a7276baa73fae8fdb39a95ae294b09b6ab" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:47:17.263424 systemd[1]: Started cri-containerd-2a26505f4183945daf9f1bceacb2cd11dbd324f37305a70db4b19964066d740c.scope - libcontainer container 2a26505f4183945daf9f1bceacb2cd11dbd324f37305a70db4b19964066d740c. Jan 14 00:47:17.282258 systemd[1]: Started cri-containerd-cfddbdad0a89c13bc1e11c07f1396bd1437f8dec977cfba7bed4ee93826eb4df.scope - libcontainer container cfddbdad0a89c13bc1e11c07f1396bd1437f8dec977cfba7bed4ee93826eb4df. Jan 14 00:47:17.293000 audit: BPF prog-id=91 op=LOAD Jan 14 00:47:17.294000 audit: BPF prog-id=92 op=LOAD Jan 14 00:47:17.294000 audit[2500]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=2466 pid=2500 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.294000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3261323635303566343138333934356461663966316263656163623263 Jan 14 00:47:17.297000 audit: BPF prog-id=92 op=UNLOAD Jan 14 00:47:17.297000 audit[2500]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2466 pid=2500 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.297000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3261323635303566343138333934356461663966316263656163623263 Jan 14 00:47:17.297000 audit: BPF prog-id=93 op=LOAD Jan 14 00:47:17.297000 audit[2500]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=2466 pid=2500 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.297000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3261323635303566343138333934356461663966316263656163623263 Jan 14 00:47:17.298000 audit: BPF prog-id=94 op=LOAD Jan 14 00:47:17.298000 audit[2500]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=2466 pid=2500 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.298000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3261323635303566343138333934356461663966316263656163623263 Jan 14 00:47:17.298000 audit: BPF prog-id=94 op=UNLOAD Jan 14 00:47:17.298000 audit[2500]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2466 pid=2500 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.298000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3261323635303566343138333934356461663966316263656163623263 Jan 14 00:47:17.298000 audit: BPF prog-id=93 op=UNLOAD Jan 14 00:47:17.298000 audit[2500]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2466 pid=2500 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.298000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3261323635303566343138333934356461663966316263656163623263 Jan 14 00:47:17.298000 audit: BPF prog-id=95 op=LOAD Jan 14 00:47:17.298000 audit[2500]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=2466 pid=2500 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.298000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3261323635303566343138333934356461663966316263656163623263 Jan 14 00:47:17.302242 systemd[1]: Started cri-containerd-d1f6835108a3019ca54fb4685084e4889cbdaa910583795c01d98fa0dd9c1519.scope - libcontainer container d1f6835108a3019ca54fb4685084e4889cbdaa910583795c01d98fa0dd9c1519. Jan 14 00:47:17.306000 audit: BPF prog-id=96 op=LOAD Jan 14 00:47:17.307000 audit: BPF prog-id=97 op=LOAD Jan 14 00:47:17.307000 audit[2512]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=2487 pid=2512 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.307000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6366646462646164306138396331336263316531316330376631333936 Jan 14 00:47:17.307000 audit: BPF prog-id=97 op=UNLOAD Jan 14 00:47:17.307000 audit[2512]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2487 pid=2512 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.307000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6366646462646164306138396331336263316531316330376631333936 Jan 14 00:47:17.308000 audit: BPF prog-id=98 op=LOAD Jan 14 00:47:17.308000 audit[2512]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=2487 pid=2512 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.308000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6366646462646164306138396331336263316531316330376631333936 Jan 14 00:47:17.308000 audit: BPF prog-id=99 op=LOAD Jan 14 00:47:17.308000 audit[2512]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=2487 pid=2512 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.308000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6366646462646164306138396331336263316531316330376631333936 Jan 14 00:47:17.309000 audit: BPF prog-id=99 op=UNLOAD Jan 14 00:47:17.309000 audit[2512]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2487 pid=2512 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.309000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6366646462646164306138396331336263316531316330376631333936 Jan 14 00:47:17.309000 audit: BPF prog-id=98 op=UNLOAD Jan 14 00:47:17.309000 audit[2512]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2487 pid=2512 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.309000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6366646462646164306138396331336263316531316330376631333936 Jan 14 00:47:17.309000 audit: BPF prog-id=100 op=LOAD Jan 14 00:47:17.309000 audit[2512]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=2487 pid=2512 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.309000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6366646462646164306138396331336263316531316330376631333936 Jan 14 00:47:17.318000 audit: BPF prog-id=101 op=LOAD Jan 14 00:47:17.321000 audit: BPF prog-id=102 op=LOAD Jan 14 00:47:17.321000 audit[2541]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=2510 pid=2541 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.321000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431663638333531303861333031396361353466623436383530383465 Jan 14 00:47:17.321000 audit: BPF prog-id=102 op=UNLOAD Jan 14 00:47:17.321000 audit[2541]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2510 pid=2541 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.321000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431663638333531303861333031396361353466623436383530383465 Jan 14 00:47:17.321000 audit: BPF prog-id=103 op=LOAD Jan 14 00:47:17.321000 audit[2541]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=2510 pid=2541 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.321000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431663638333531303861333031396361353466623436383530383465 Jan 14 00:47:17.322000 audit: BPF prog-id=104 op=LOAD Jan 14 00:47:17.322000 audit[2541]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=2510 pid=2541 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.322000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431663638333531303861333031396361353466623436383530383465 Jan 14 00:47:17.322000 audit: BPF prog-id=104 op=UNLOAD Jan 14 00:47:17.322000 audit[2541]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2510 pid=2541 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.322000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431663638333531303861333031396361353466623436383530383465 Jan 14 00:47:17.322000 audit: BPF prog-id=103 op=UNLOAD Jan 14 00:47:17.322000 audit[2541]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2510 pid=2541 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.322000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431663638333531303861333031396361353466623436383530383465 Jan 14 00:47:17.322000 audit: BPF prog-id=105 op=LOAD Jan 14 00:47:17.322000 audit[2541]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=2510 pid=2541 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.322000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6431663638333531303861333031396361353466623436383530383465 Jan 14 00:47:17.372149 containerd[1611]: time="2026-01-14T00:47:17.370517544Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-172-236-110-177,Uid:8bbf6446f3f4d6255fd3ad0a0405f236,Namespace:kube-system,Attempt:0,} returns sandbox id \"2a26505f4183945daf9f1bceacb2cd11dbd324f37305a70db4b19964066d740c\"" Jan 14 00:47:17.373322 kubelet[2421]: E0114 00:47:17.373009 2421 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:17.378422 containerd[1611]: time="2026-01-14T00:47:17.378387699Z" level=info msg="CreateContainer within sandbox \"2a26505f4183945daf9f1bceacb2cd11dbd324f37305a70db4b19964066d740c\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jan 14 00:47:17.390465 containerd[1611]: time="2026-01-14T00:47:17.390330944Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-172-236-110-177,Uid:3ae0af65e6e9e55461a48715990c01e7,Namespace:kube-system,Attempt:0,} returns sandbox id \"cfddbdad0a89c13bc1e11c07f1396bd1437f8dec977cfba7bed4ee93826eb4df\"" Jan 14 00:47:17.392154 kubelet[2421]: E0114 00:47:17.392013 2421 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:17.394408 containerd[1611]: time="2026-01-14T00:47:17.394373620Z" level=info msg="Container 70777344d79354dcad4a7606120b1adafab8aef901ebdf1188e2549766777e50: CDI devices from CRI Config.CDIDevices: []" Jan 14 00:47:17.394964 containerd[1611]: time="2026-01-14T00:47:17.394945394Z" level=info msg="CreateContainer within sandbox \"cfddbdad0a89c13bc1e11c07f1396bd1437f8dec977cfba7bed4ee93826eb4df\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jan 14 00:47:17.403887 containerd[1611]: time="2026-01-14T00:47:17.403789731Z" level=info msg="CreateContainer within sandbox \"2a26505f4183945daf9f1bceacb2cd11dbd324f37305a70db4b19964066d740c\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"70777344d79354dcad4a7606120b1adafab8aef901ebdf1188e2549766777e50\"" Jan 14 00:47:17.404435 containerd[1611]: time="2026-01-14T00:47:17.404417176Z" level=info msg="StartContainer for \"70777344d79354dcad4a7606120b1adafab8aef901ebdf1188e2549766777e50\"" Jan 14 00:47:17.404988 containerd[1611]: time="2026-01-14T00:47:17.404970096Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-172-236-110-177,Uid:f6a3c8bb6d69c8c4320f01549746566a,Namespace:kube-system,Attempt:0,} returns sandbox id \"d1f6835108a3019ca54fb4685084e4889cbdaa910583795c01d98fa0dd9c1519\"" Jan 14 00:47:17.405993 containerd[1611]: time="2026-01-14T00:47:17.405851839Z" level=info msg="connecting to shim 70777344d79354dcad4a7606120b1adafab8aef901ebdf1188e2549766777e50" address="unix:///run/containerd/s/2a87a64a0c81b43875ef460052fe9bb0e33b6ecfa31c66d27e8a2ed196e36ad4" protocol=ttrpc version=3 Jan 14 00:47:17.407600 kubelet[2421]: E0114 00:47:17.407556 2421 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:17.410454 containerd[1611]: time="2026-01-14T00:47:17.410350665Z" level=info msg="CreateContainer within sandbox \"d1f6835108a3019ca54fb4685084e4889cbdaa910583795c01d98fa0dd9c1519\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jan 14 00:47:17.410754 containerd[1611]: time="2026-01-14T00:47:17.410356480Z" level=info msg="Container 33b9cd5af421d6ef461e7485cfaa38aeb9f5876eb9c899c66cbcd0d834905761: CDI devices from CRI Config.CDIDevices: []" Jan 14 00:47:17.413530 kubelet[2421]: I0114 00:47:17.413504 2421 kubelet_node_status.go:75] "Attempting to register node" node="172-236-110-177" Jan 14 00:47:17.413807 kubelet[2421]: E0114 00:47:17.413782 2421 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.236.110.177:6443/api/v1/nodes\": dial tcp 172.236.110.177:6443: connect: connection refused" node="172-236-110-177" Jan 14 00:47:17.417980 containerd[1611]: time="2026-01-14T00:47:17.417957574Z" level=info msg="CreateContainer within sandbox \"cfddbdad0a89c13bc1e11c07f1396bd1437f8dec977cfba7bed4ee93826eb4df\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"33b9cd5af421d6ef461e7485cfaa38aeb9f5876eb9c899c66cbcd0d834905761\"" Jan 14 00:47:17.419757 containerd[1611]: time="2026-01-14T00:47:17.419708142Z" level=info msg="StartContainer for \"33b9cd5af421d6ef461e7485cfaa38aeb9f5876eb9c899c66cbcd0d834905761\"" Jan 14 00:47:17.420783 containerd[1611]: time="2026-01-14T00:47:17.420550286Z" level=info msg="connecting to shim 33b9cd5af421d6ef461e7485cfaa38aeb9f5876eb9c899c66cbcd0d834905761" address="unix:///run/containerd/s/2431b978943890a23fd50b031b4f3994c50fcd7d6313d1c1de77403ea8b73ad8" protocol=ttrpc version=3 Jan 14 00:47:17.425209 containerd[1611]: time="2026-01-14T00:47:17.425174685Z" level=info msg="Container 13607a672ba3765c9021be12cc8c88c54bf8076c503ceaabddd17db47c7f9a27: CDI devices from CRI Config.CDIDevices: []" Jan 14 00:47:17.434681 containerd[1611]: time="2026-01-14T00:47:17.434649088Z" level=info msg="CreateContainer within sandbox \"d1f6835108a3019ca54fb4685084e4889cbdaa910583795c01d98fa0dd9c1519\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"13607a672ba3765c9021be12cc8c88c54bf8076c503ceaabddd17db47c7f9a27\"" Jan 14 00:47:17.434959 containerd[1611]: time="2026-01-14T00:47:17.434940875Z" level=info msg="StartContainer for \"13607a672ba3765c9021be12cc8c88c54bf8076c503ceaabddd17db47c7f9a27\"" Jan 14 00:47:17.438136 containerd[1611]: time="2026-01-14T00:47:17.438091110Z" level=info msg="connecting to shim 13607a672ba3765c9021be12cc8c88c54bf8076c503ceaabddd17db47c7f9a27" address="unix:///run/containerd/s/4ea470c45a4002b98c3756d4f05672a7276baa73fae8fdb39a95ae294b09b6ab" protocol=ttrpc version=3 Jan 14 00:47:17.438329 systemd[1]: Started cri-containerd-70777344d79354dcad4a7606120b1adafab8aef901ebdf1188e2549766777e50.scope - libcontainer container 70777344d79354dcad4a7606120b1adafab8aef901ebdf1188e2549766777e50. Jan 14 00:47:17.454259 systemd[1]: Started cri-containerd-33b9cd5af421d6ef461e7485cfaa38aeb9f5876eb9c899c66cbcd0d834905761.scope - libcontainer container 33b9cd5af421d6ef461e7485cfaa38aeb9f5876eb9c899c66cbcd0d834905761. Jan 14 00:47:17.482256 systemd[1]: Started cri-containerd-13607a672ba3765c9021be12cc8c88c54bf8076c503ceaabddd17db47c7f9a27.scope - libcontainer container 13607a672ba3765c9021be12cc8c88c54bf8076c503ceaabddd17db47c7f9a27. Jan 14 00:47:17.486000 audit: BPF prog-id=106 op=LOAD Jan 14 00:47:17.487000 audit: BPF prog-id=107 op=LOAD Jan 14 00:47:17.487000 audit[2611]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=2487 pid=2611 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.487000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333623963643561663432316436656634363165373438356366616133 Jan 14 00:47:17.487000 audit: BPF prog-id=107 op=UNLOAD Jan 14 00:47:17.487000 audit[2611]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2487 pid=2611 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.487000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333623963643561663432316436656634363165373438356366616133 Jan 14 00:47:17.487000 audit: BPF prog-id=108 op=LOAD Jan 14 00:47:17.487000 audit[2611]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=2487 pid=2611 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.487000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333623963643561663432316436656634363165373438356366616133 Jan 14 00:47:17.488000 audit: BPF prog-id=109 op=LOAD Jan 14 00:47:17.488000 audit[2611]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=2487 pid=2611 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.488000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333623963643561663432316436656634363165373438356366616133 Jan 14 00:47:17.488000 audit: BPF prog-id=109 op=UNLOAD Jan 14 00:47:17.488000 audit[2611]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2487 pid=2611 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.488000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333623963643561663432316436656634363165373438356366616133 Jan 14 00:47:17.488000 audit: BPF prog-id=108 op=UNLOAD Jan 14 00:47:17.488000 audit[2611]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2487 pid=2611 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.488000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333623963643561663432316436656634363165373438356366616133 Jan 14 00:47:17.488000 audit: BPF prog-id=110 op=LOAD Jan 14 00:47:17.488000 audit[2611]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=2487 pid=2611 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.488000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333623963643561663432316436656634363165373438356366616133 Jan 14 00:47:17.489000 audit: BPF prog-id=111 op=LOAD Jan 14 00:47:17.489000 audit: BPF prog-id=112 op=LOAD Jan 14 00:47:17.489000 audit[2599]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=2466 pid=2599 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.489000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3730373737333434643739333534646361643461373630363132306231 Jan 14 00:47:17.489000 audit: BPF prog-id=112 op=UNLOAD Jan 14 00:47:17.489000 audit[2599]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2466 pid=2599 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.489000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3730373737333434643739333534646361643461373630363132306231 Jan 14 00:47:17.491000 audit: BPF prog-id=113 op=LOAD Jan 14 00:47:17.491000 audit[2599]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=2466 pid=2599 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.491000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3730373737333434643739333534646361643461373630363132306231 Jan 14 00:47:17.491000 audit: BPF prog-id=114 op=LOAD Jan 14 00:47:17.491000 audit[2599]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=2466 pid=2599 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.491000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3730373737333434643739333534646361643461373630363132306231 Jan 14 00:47:17.492000 audit: BPF prog-id=114 op=UNLOAD Jan 14 00:47:17.492000 audit[2599]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2466 pid=2599 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.492000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3730373737333434643739333534646361643461373630363132306231 Jan 14 00:47:17.492000 audit: BPF prog-id=113 op=UNLOAD Jan 14 00:47:17.492000 audit[2599]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2466 pid=2599 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.492000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3730373737333434643739333534646361643461373630363132306231 Jan 14 00:47:17.492000 audit: BPF prog-id=115 op=LOAD Jan 14 00:47:17.492000 audit[2599]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=2466 pid=2599 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.492000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3730373737333434643739333534646361643461373630363132306231 Jan 14 00:47:17.510000 audit: BPF prog-id=116 op=LOAD Jan 14 00:47:17.510000 audit: BPF prog-id=117 op=LOAD Jan 14 00:47:17.510000 audit[2624]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000128238 a2=98 a3=0 items=0 ppid=2510 pid=2624 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.510000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3133363037613637326261333736356339303231626531326363386338 Jan 14 00:47:17.510000 audit: BPF prog-id=117 op=UNLOAD Jan 14 00:47:17.510000 audit[2624]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2510 pid=2624 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.510000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3133363037613637326261333736356339303231626531326363386338 Jan 14 00:47:17.510000 audit: BPF prog-id=118 op=LOAD Jan 14 00:47:17.510000 audit[2624]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000128488 a2=98 a3=0 items=0 ppid=2510 pid=2624 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.510000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3133363037613637326261333736356339303231626531326363386338 Jan 14 00:47:17.510000 audit: BPF prog-id=119 op=LOAD Jan 14 00:47:17.510000 audit[2624]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000128218 a2=98 a3=0 items=0 ppid=2510 pid=2624 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.510000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3133363037613637326261333736356339303231626531326363386338 Jan 14 00:47:17.510000 audit: BPF prog-id=119 op=UNLOAD Jan 14 00:47:17.510000 audit[2624]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2510 pid=2624 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.510000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3133363037613637326261333736356339303231626531326363386338 Jan 14 00:47:17.511000 audit: BPF prog-id=118 op=UNLOAD Jan 14 00:47:17.511000 audit[2624]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2510 pid=2624 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.511000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3133363037613637326261333736356339303231626531326363386338 Jan 14 00:47:17.511000 audit: BPF prog-id=120 op=LOAD Jan 14 00:47:17.511000 audit[2624]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001286e8 a2=98 a3=0 items=0 ppid=2510 pid=2624 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:17.511000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3133363037613637326261333736356339303231626531326363386338 Jan 14 00:47:17.539206 containerd[1611]: time="2026-01-14T00:47:17.539062915Z" level=info msg="StartContainer for \"33b9cd5af421d6ef461e7485cfaa38aeb9f5876eb9c899c66cbcd0d834905761\" returns successfully" Jan 14 00:47:17.587904 containerd[1611]: time="2026-01-14T00:47:17.587841146Z" level=info msg="StartContainer for \"70777344d79354dcad4a7606120b1adafab8aef901ebdf1188e2549766777e50\" returns successfully" Jan 14 00:47:17.589281 containerd[1611]: time="2026-01-14T00:47:17.589261570Z" level=info msg="StartContainer for \"13607a672ba3765c9021be12cc8c88c54bf8076c503ceaabddd17db47c7f9a27\" returns successfully" Jan 14 00:47:17.661749 kubelet[2421]: E0114 00:47:17.661590 2421 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://172.236.110.177:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.236.110.177:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jan 14 00:47:17.676509 kubelet[2421]: E0114 00:47:17.676490 2421 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-236-110-177\" not found" node="172-236-110-177" Jan 14 00:47:17.676791 kubelet[2421]: E0114 00:47:17.676753 2421 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:17.678049 kubelet[2421]: E0114 00:47:17.678023 2421 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-236-110-177\" not found" node="172-236-110-177" Jan 14 00:47:17.678407 kubelet[2421]: E0114 00:47:17.678193 2421 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-236-110-177\" not found" node="172-236-110-177" Jan 14 00:47:17.678509 kubelet[2421]: E0114 00:47:17.678496 2421 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:17.679343 kubelet[2421]: E0114 00:47:17.679330 2421 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:18.217182 kubelet[2421]: I0114 00:47:18.216570 2421 kubelet_node_status.go:75] "Attempting to register node" node="172-236-110-177" Jan 14 00:47:18.681710 kubelet[2421]: E0114 00:47:18.681271 2421 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-236-110-177\" not found" node="172-236-110-177" Jan 14 00:47:18.681710 kubelet[2421]: E0114 00:47:18.681373 2421 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:18.682358 kubelet[2421]: E0114 00:47:18.682343 2421 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"172-236-110-177\" not found" node="172-236-110-177" Jan 14 00:47:18.682601 kubelet[2421]: E0114 00:47:18.682589 2421 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:19.001070 kubelet[2421]: E0114 00:47:19.000512 2421 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"172-236-110-177\" not found" node="172-236-110-177" Jan 14 00:47:19.084340 kubelet[2421]: I0114 00:47:19.084241 2421 kubelet_node_status.go:78] "Successfully registered node" node="172-236-110-177" Jan 14 00:47:19.084340 kubelet[2421]: E0114 00:47:19.084288 2421 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"172-236-110-177\": node \"172-236-110-177\" not found" Jan 14 00:47:19.116510 kubelet[2421]: I0114 00:47:19.116467 2421 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:19.133157 kubelet[2421]: E0114 00:47:19.132589 2421 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{172-236-110-177.188a7271c17f45b2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:172-236-110-177,UID:172-236-110-177,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:172-236-110-177,},FirstTimestamp:2026-01-14 00:47:16.599555506 +0000 UTC m=+0.291600986,LastTimestamp:2026-01-14 00:47:16.599555506 +0000 UTC m=+0.291600986,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172-236-110-177,}" Jan 14 00:47:19.174187 kubelet[2421]: E0114 00:47:19.174070 2421 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-172-236-110-177\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:19.174187 kubelet[2421]: I0114 00:47:19.174126 2421 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-172-236-110-177" Jan 14 00:47:19.175956 kubelet[2421]: E0114 00:47:19.175739 2421 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-172-236-110-177\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-172-236-110-177" Jan 14 00:47:19.175956 kubelet[2421]: I0114 00:47:19.175762 2421 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-172-236-110-177" Jan 14 00:47:19.177981 kubelet[2421]: E0114 00:47:19.177948 2421 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-172-236-110-177\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-172-236-110-177" Jan 14 00:47:19.597252 kubelet[2421]: I0114 00:47:19.597189 2421 apiserver.go:52] "Watching apiserver" Jan 14 00:47:19.616894 kubelet[2421]: I0114 00:47:19.616839 2421 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Jan 14 00:47:20.769058 systemd[1]: Reload requested from client PID 2704 ('systemctl') (unit session-8.scope)... Jan 14 00:47:20.769382 systemd[1]: Reloading... Jan 14 00:47:20.885161 zram_generator::config[2760]: No configuration found. Jan 14 00:47:21.116708 systemd[1]: Reloading finished in 346 ms. Jan 14 00:47:21.148540 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 00:47:21.169604 systemd[1]: kubelet.service: Deactivated successfully. Jan 14 00:47:21.169959 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 00:47:21.173231 kernel: kauditd_printk_skb: 210 callbacks suppressed Jan 14 00:47:21.173281 kernel: audit: type=1131 audit(1768351641.168:399): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:47:21.168000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:47:21.170025 systemd[1]: kubelet.service: Consumed 732ms CPU time, 130M memory peak. Jan 14 00:47:21.179571 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 14 00:47:21.178000 audit: BPF prog-id=121 op=LOAD Jan 14 00:47:21.184802 kernel: audit: type=1334 audit(1768351641.178:400): prog-id=121 op=LOAD Jan 14 00:47:21.184843 kernel: audit: type=1334 audit(1768351641.179:401): prog-id=84 op=UNLOAD Jan 14 00:47:21.179000 audit: BPF prog-id=84 op=UNLOAD Jan 14 00:47:21.187266 kernel: audit: type=1334 audit(1768351641.181:402): prog-id=122 op=LOAD Jan 14 00:47:21.181000 audit: BPF prog-id=122 op=LOAD Jan 14 00:47:21.189558 kernel: audit: type=1334 audit(1768351641.181:403): prog-id=81 op=UNLOAD Jan 14 00:47:21.181000 audit: BPF prog-id=81 op=UNLOAD Jan 14 00:47:21.191792 kernel: audit: type=1334 audit(1768351641.181:404): prog-id=123 op=LOAD Jan 14 00:47:21.181000 audit: BPF prog-id=123 op=LOAD Jan 14 00:47:21.193965 kernel: audit: type=1334 audit(1768351641.181:405): prog-id=124 op=LOAD Jan 14 00:47:21.181000 audit: BPF prog-id=124 op=LOAD Jan 14 00:47:21.181000 audit: BPF prog-id=82 op=UNLOAD Jan 14 00:47:21.194648 kernel: audit: type=1334 audit(1768351641.181:406): prog-id=82 op=UNLOAD Jan 14 00:47:21.198129 kernel: audit: type=1334 audit(1768351641.181:407): prog-id=83 op=UNLOAD Jan 14 00:47:21.181000 audit: BPF prog-id=83 op=UNLOAD Jan 14 00:47:21.182000 audit: BPF prog-id=125 op=LOAD Jan 14 00:47:21.182000 audit: BPF prog-id=67 op=UNLOAD Jan 14 00:47:21.185000 audit: BPF prog-id=126 op=LOAD Jan 14 00:47:21.185000 audit: BPF prog-id=72 op=UNLOAD Jan 14 00:47:21.186000 audit: BPF prog-id=127 op=LOAD Jan 14 00:47:21.186000 audit: BPF prog-id=128 op=LOAD Jan 14 00:47:21.186000 audit: BPF prog-id=76 op=UNLOAD Jan 14 00:47:21.186000 audit: BPF prog-id=77 op=UNLOAD Jan 14 00:47:21.188000 audit: BPF prog-id=129 op=LOAD Jan 14 00:47:21.188000 audit: BPF prog-id=85 op=UNLOAD Jan 14 00:47:21.188000 audit: BPF prog-id=130 op=LOAD Jan 14 00:47:21.202263 kernel: audit: type=1334 audit(1768351641.182:408): prog-id=125 op=LOAD Jan 14 00:47:21.188000 audit: BPF prog-id=131 op=LOAD Jan 14 00:47:21.188000 audit: BPF prog-id=86 op=UNLOAD Jan 14 00:47:21.188000 audit: BPF prog-id=87 op=UNLOAD Jan 14 00:47:21.189000 audit: BPF prog-id=132 op=LOAD Jan 14 00:47:21.189000 audit: BPF prog-id=78 op=UNLOAD Jan 14 00:47:21.189000 audit: BPF prog-id=133 op=LOAD Jan 14 00:47:21.189000 audit: BPF prog-id=134 op=LOAD Jan 14 00:47:21.189000 audit: BPF prog-id=79 op=UNLOAD Jan 14 00:47:21.189000 audit: BPF prog-id=80 op=UNLOAD Jan 14 00:47:21.196000 audit: BPF prog-id=135 op=LOAD Jan 14 00:47:21.196000 audit: BPF prog-id=68 op=UNLOAD Jan 14 00:47:21.196000 audit: BPF prog-id=136 op=LOAD Jan 14 00:47:21.196000 audit: BPF prog-id=137 op=LOAD Jan 14 00:47:21.196000 audit: BPF prog-id=69 op=UNLOAD Jan 14 00:47:21.196000 audit: BPF prog-id=70 op=UNLOAD Jan 14 00:47:21.199000 audit: BPF prog-id=138 op=LOAD Jan 14 00:47:21.199000 audit: BPF prog-id=88 op=UNLOAD Jan 14 00:47:21.199000 audit: BPF prog-id=139 op=LOAD Jan 14 00:47:21.199000 audit: BPF prog-id=140 op=LOAD Jan 14 00:47:21.199000 audit: BPF prog-id=89 op=UNLOAD Jan 14 00:47:21.199000 audit: BPF prog-id=90 op=UNLOAD Jan 14 00:47:21.200000 audit: BPF prog-id=141 op=LOAD Jan 14 00:47:21.200000 audit: BPF prog-id=71 op=UNLOAD Jan 14 00:47:21.202000 audit: BPF prog-id=142 op=LOAD Jan 14 00:47:21.202000 audit: BPF prog-id=73 op=UNLOAD Jan 14 00:47:21.202000 audit: BPF prog-id=143 op=LOAD Jan 14 00:47:21.202000 audit: BPF prog-id=144 op=LOAD Jan 14 00:47:21.202000 audit: BPF prog-id=74 op=UNLOAD Jan 14 00:47:21.202000 audit: BPF prog-id=75 op=UNLOAD Jan 14 00:47:21.386852 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 14 00:47:21.386000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:47:21.399407 (kubelet)[2802]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 14 00:47:21.442338 kubelet[2802]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 14 00:47:21.442338 kubelet[2802]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jan 14 00:47:21.442338 kubelet[2802]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 14 00:47:21.442338 kubelet[2802]: I0114 00:47:21.441919 2802 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 14 00:47:21.453333 kubelet[2802]: I0114 00:47:21.452927 2802 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Jan 14 00:47:21.453333 kubelet[2802]: I0114 00:47:21.452944 2802 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 14 00:47:21.453440 kubelet[2802]: I0114 00:47:21.453340 2802 server.go:956] "Client rotation is on, will bootstrap in background" Jan 14 00:47:21.454534 kubelet[2802]: I0114 00:47:21.454507 2802 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Jan 14 00:47:21.457693 kubelet[2802]: I0114 00:47:21.457679 2802 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 14 00:47:21.461163 kubelet[2802]: I0114 00:47:21.461146 2802 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 14 00:47:21.469051 kubelet[2802]: I0114 00:47:21.467420 2802 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 14 00:47:21.469051 kubelet[2802]: I0114 00:47:21.467626 2802 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 14 00:47:21.469051 kubelet[2802]: I0114 00:47:21.467642 2802 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"172-236-110-177","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 14 00:47:21.469051 kubelet[2802]: I0114 00:47:21.467818 2802 topology_manager.go:138] "Creating topology manager with none policy" Jan 14 00:47:21.469429 kubelet[2802]: I0114 00:47:21.467827 2802 container_manager_linux.go:303] "Creating device plugin manager" Jan 14 00:47:21.469429 kubelet[2802]: I0114 00:47:21.467870 2802 state_mem.go:36] "Initialized new in-memory state store" Jan 14 00:47:21.469429 kubelet[2802]: I0114 00:47:21.468241 2802 kubelet.go:480] "Attempting to sync node with API server" Jan 14 00:47:21.469429 kubelet[2802]: I0114 00:47:21.468252 2802 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 14 00:47:21.469429 kubelet[2802]: I0114 00:47:21.468274 2802 kubelet.go:386] "Adding apiserver pod source" Jan 14 00:47:21.469429 kubelet[2802]: I0114 00:47:21.468287 2802 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 14 00:47:21.470510 kubelet[2802]: I0114 00:47:21.470493 2802 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.1.5" apiVersion="v1" Jan 14 00:47:21.471267 kubelet[2802]: I0114 00:47:21.471253 2802 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Jan 14 00:47:21.474333 kubelet[2802]: I0114 00:47:21.474320 2802 watchdog_linux.go:99] "Systemd watchdog is not enabled" Jan 14 00:47:21.474422 kubelet[2802]: I0114 00:47:21.474412 2802 server.go:1289] "Started kubelet" Jan 14 00:47:21.476829 kubelet[2802]: I0114 00:47:21.476816 2802 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 14 00:47:21.490212 kubelet[2802]: I0114 00:47:21.489044 2802 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 14 00:47:21.492130 kubelet[2802]: I0114 00:47:21.476839 2802 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Jan 14 00:47:21.492130 kubelet[2802]: I0114 00:47:21.492088 2802 server.go:317] "Adding debug handlers to kubelet server" Jan 14 00:47:21.493159 kubelet[2802]: I0114 00:47:21.492908 2802 volume_manager.go:297] "Starting Kubelet Volume Manager" Jan 14 00:47:21.493159 kubelet[2802]: E0114 00:47:21.493044 2802 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172-236-110-177\" not found" Jan 14 00:47:21.495324 kubelet[2802]: I0114 00:47:21.476867 2802 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 14 00:47:21.495324 kubelet[2802]: I0114 00:47:21.494870 2802 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 14 00:47:21.495396 kubelet[2802]: I0114 00:47:21.495339 2802 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Jan 14 00:47:21.495458 kubelet[2802]: I0114 00:47:21.495442 2802 reconciler.go:26] "Reconciler: start to sync state" Jan 14 00:47:21.497238 kubelet[2802]: I0114 00:47:21.497212 2802 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Jan 14 00:47:21.499132 kubelet[2802]: I0114 00:47:21.498332 2802 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Jan 14 00:47:21.499132 kubelet[2802]: I0114 00:47:21.498357 2802 status_manager.go:230] "Starting to sync pod status with apiserver" Jan 14 00:47:21.499132 kubelet[2802]: I0114 00:47:21.498371 2802 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jan 14 00:47:21.499132 kubelet[2802]: I0114 00:47:21.498377 2802 kubelet.go:2436] "Starting kubelet main sync loop" Jan 14 00:47:21.499132 kubelet[2802]: E0114 00:47:21.498415 2802 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 14 00:47:21.503542 kubelet[2802]: I0114 00:47:21.503527 2802 factory.go:223] Registration of the systemd container factory successfully Jan 14 00:47:21.505212 kubelet[2802]: I0114 00:47:21.505183 2802 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 14 00:47:21.507590 kubelet[2802]: E0114 00:47:21.507567 2802 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 14 00:47:21.507990 kubelet[2802]: I0114 00:47:21.507953 2802 factory.go:223] Registration of the containerd container factory successfully Jan 14 00:47:21.557223 kubelet[2802]: I0114 00:47:21.556994 2802 cpu_manager.go:221] "Starting CPU manager" policy="none" Jan 14 00:47:21.557390 kubelet[2802]: I0114 00:47:21.557376 2802 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jan 14 00:47:21.557466 kubelet[2802]: I0114 00:47:21.557457 2802 state_mem.go:36] "Initialized new in-memory state store" Jan 14 00:47:21.557614 kubelet[2802]: I0114 00:47:21.557601 2802 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jan 14 00:47:21.557682 kubelet[2802]: I0114 00:47:21.557660 2802 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jan 14 00:47:21.557724 kubelet[2802]: I0114 00:47:21.557717 2802 policy_none.go:49] "None policy: Start" Jan 14 00:47:21.557765 kubelet[2802]: I0114 00:47:21.557757 2802 memory_manager.go:186] "Starting memorymanager" policy="None" Jan 14 00:47:21.557812 kubelet[2802]: I0114 00:47:21.557804 2802 state_mem.go:35] "Initializing new in-memory state store" Jan 14 00:47:21.557943 kubelet[2802]: I0114 00:47:21.557932 2802 state_mem.go:75] "Updated machine memory state" Jan 14 00:47:21.562759 kubelet[2802]: E0114 00:47:21.562740 2802 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Jan 14 00:47:21.563164 kubelet[2802]: I0114 00:47:21.563144 2802 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 14 00:47:21.563419 kubelet[2802]: I0114 00:47:21.563158 2802 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 14 00:47:21.563636 kubelet[2802]: I0114 00:47:21.563616 2802 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 14 00:47:21.564999 kubelet[2802]: E0114 00:47:21.564960 2802 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jan 14 00:47:21.599981 kubelet[2802]: I0114 00:47:21.599672 2802 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:21.599981 kubelet[2802]: I0114 00:47:21.599713 2802 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-172-236-110-177" Jan 14 00:47:21.599981 kubelet[2802]: I0114 00:47:21.599871 2802 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-172-236-110-177" Jan 14 00:47:21.666159 kubelet[2802]: I0114 00:47:21.665477 2802 kubelet_node_status.go:75] "Attempting to register node" node="172-236-110-177" Jan 14 00:47:21.672453 kubelet[2802]: I0114 00:47:21.672416 2802 kubelet_node_status.go:124] "Node was previously registered" node="172-236-110-177" Jan 14 00:47:21.672507 kubelet[2802]: I0114 00:47:21.672477 2802 kubelet_node_status.go:78] "Successfully registered node" node="172-236-110-177" Jan 14 00:47:21.697361 kubelet[2802]: I0114 00:47:21.697323 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8bbf6446f3f4d6255fd3ad0a0405f236-ca-certs\") pod \"kube-controller-manager-172-236-110-177\" (UID: \"8bbf6446f3f4d6255fd3ad0a0405f236\") " pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:21.697361 kubelet[2802]: I0114 00:47:21.697359 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/8bbf6446f3f4d6255fd3ad0a0405f236-flexvolume-dir\") pod \"kube-controller-manager-172-236-110-177\" (UID: \"8bbf6446f3f4d6255fd3ad0a0405f236\") " pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:21.697361 kubelet[2802]: I0114 00:47:21.697376 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8bbf6446f3f4d6255fd3ad0a0405f236-k8s-certs\") pod \"kube-controller-manager-172-236-110-177\" (UID: \"8bbf6446f3f4d6255fd3ad0a0405f236\") " pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:21.697361 kubelet[2802]: I0114 00:47:21.697391 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8bbf6446f3f4d6255fd3ad0a0405f236-kubeconfig\") pod \"kube-controller-manager-172-236-110-177\" (UID: \"8bbf6446f3f4d6255fd3ad0a0405f236\") " pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:21.697645 kubelet[2802]: I0114 00:47:21.697408 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8bbf6446f3f4d6255fd3ad0a0405f236-usr-share-ca-certificates\") pod \"kube-controller-manager-172-236-110-177\" (UID: \"8bbf6446f3f4d6255fd3ad0a0405f236\") " pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:21.697645 kubelet[2802]: I0114 00:47:21.697423 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/3ae0af65e6e9e55461a48715990c01e7-kubeconfig\") pod \"kube-scheduler-172-236-110-177\" (UID: \"3ae0af65e6e9e55461a48715990c01e7\") " pod="kube-system/kube-scheduler-172-236-110-177" Jan 14 00:47:21.697645 kubelet[2802]: I0114 00:47:21.697438 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f6a3c8bb6d69c8c4320f01549746566a-ca-certs\") pod \"kube-apiserver-172-236-110-177\" (UID: \"f6a3c8bb6d69c8c4320f01549746566a\") " pod="kube-system/kube-apiserver-172-236-110-177" Jan 14 00:47:21.697645 kubelet[2802]: I0114 00:47:21.697467 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f6a3c8bb6d69c8c4320f01549746566a-k8s-certs\") pod \"kube-apiserver-172-236-110-177\" (UID: \"f6a3c8bb6d69c8c4320f01549746566a\") " pod="kube-system/kube-apiserver-172-236-110-177" Jan 14 00:47:21.697645 kubelet[2802]: I0114 00:47:21.697481 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f6a3c8bb6d69c8c4320f01549746566a-usr-share-ca-certificates\") pod \"kube-apiserver-172-236-110-177\" (UID: \"f6a3c8bb6d69c8c4320f01549746566a\") " pod="kube-system/kube-apiserver-172-236-110-177" Jan 14 00:47:21.907792 kubelet[2802]: E0114 00:47:21.907686 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:21.908130 kubelet[2802]: E0114 00:47:21.908088 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:21.908431 kubelet[2802]: E0114 00:47:21.908410 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:22.470877 kubelet[2802]: I0114 00:47:22.470748 2802 apiserver.go:52] "Watching apiserver" Jan 14 00:47:22.496039 kubelet[2802]: I0114 00:47:22.496006 2802 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Jan 14 00:47:22.539983 kubelet[2802]: I0114 00:47:22.539944 2802 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-172-236-110-177" Jan 14 00:47:22.540534 kubelet[2802]: I0114 00:47:22.540519 2802 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:22.542261 kubelet[2802]: I0114 00:47:22.542247 2802 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-172-236-110-177" Jan 14 00:47:22.547330 kubelet[2802]: E0114 00:47:22.546637 2802 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-172-236-110-177\" already exists" pod="kube-system/kube-apiserver-172-236-110-177" Jan 14 00:47:22.547330 kubelet[2802]: E0114 00:47:22.546759 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:22.547330 kubelet[2802]: E0114 00:47:22.546636 2802 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-172-236-110-177\" already exists" pod="kube-system/kube-controller-manager-172-236-110-177" Jan 14 00:47:22.547756 kubelet[2802]: E0114 00:47:22.547743 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:22.548176 kubelet[2802]: E0114 00:47:22.548162 2802 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-172-236-110-177\" already exists" pod="kube-system/kube-scheduler-172-236-110-177" Jan 14 00:47:22.548527 kubelet[2802]: E0114 00:47:22.548514 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:22.561164 kubelet[2802]: I0114 00:47:22.561132 2802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-172-236-110-177" podStartSLOduration=1.5610867160000002 podStartE2EDuration="1.561086716s" podCreationTimestamp="2026-01-14 00:47:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-14 00:47:22.554396374 +0000 UTC m=+1.149492788" watchObservedRunningTime="2026-01-14 00:47:22.561086716 +0000 UTC m=+1.156183130" Jan 14 00:47:22.567585 kubelet[2802]: I0114 00:47:22.567546 2802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-172-236-110-177" podStartSLOduration=1.567538557 podStartE2EDuration="1.567538557s" podCreationTimestamp="2026-01-14 00:47:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-14 00:47:22.56133669 +0000 UTC m=+1.156433114" watchObservedRunningTime="2026-01-14 00:47:22.567538557 +0000 UTC m=+1.162634971" Jan 14 00:47:22.575228 kubelet[2802]: I0114 00:47:22.575192 2802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-172-236-110-177" podStartSLOduration=1.575184411 podStartE2EDuration="1.575184411s" podCreationTimestamp="2026-01-14 00:47:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-14 00:47:22.567803598 +0000 UTC m=+1.162900012" watchObservedRunningTime="2026-01-14 00:47:22.575184411 +0000 UTC m=+1.170280825" Jan 14 00:47:23.543341 kubelet[2802]: E0114 00:47:23.542889 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:23.543341 kubelet[2802]: E0114 00:47:23.542979 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:23.543341 kubelet[2802]: E0114 00:47:23.543315 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:24.164551 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Jan 14 00:47:24.163000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hostnamed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:47:24.182000 audit: BPF prog-id=122 op=UNLOAD Jan 14 00:47:25.453902 kubelet[2802]: E0114 00:47:25.453864 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:26.955569 kubelet[2802]: I0114 00:47:26.955535 2802 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jan 14 00:47:26.956168 containerd[1611]: time="2026-01-14T00:47:26.955879944Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jan 14 00:47:26.956424 kubelet[2802]: I0114 00:47:26.956401 2802 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jan 14 00:47:27.524009 kubelet[2802]: E0114 00:47:27.523287 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:27.551272 kubelet[2802]: E0114 00:47:27.551237 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:28.081917 systemd[1]: Created slice kubepods-besteffort-podbd1ee5cf_1728_44da_9a75_5fb88c305faa.slice - libcontainer container kubepods-besteffort-podbd1ee5cf_1728_44da_9a75_5fb88c305faa.slice. Jan 14 00:47:28.139776 kubelet[2802]: I0114 00:47:28.139724 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/bd1ee5cf-1728-44da-9a75-5fb88c305faa-kube-proxy\") pod \"kube-proxy-k8h5v\" (UID: \"bd1ee5cf-1728-44da-9a75-5fb88c305faa\") " pod="kube-system/kube-proxy-k8h5v" Jan 14 00:47:28.139776 kubelet[2802]: I0114 00:47:28.139770 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd1ee5cf-1728-44da-9a75-5fb88c305faa-lib-modules\") pod \"kube-proxy-k8h5v\" (UID: \"bd1ee5cf-1728-44da-9a75-5fb88c305faa\") " pod="kube-system/kube-proxy-k8h5v" Jan 14 00:47:28.142152 kubelet[2802]: I0114 00:47:28.139792 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/bd1ee5cf-1728-44da-9a75-5fb88c305faa-xtables-lock\") pod \"kube-proxy-k8h5v\" (UID: \"bd1ee5cf-1728-44da-9a75-5fb88c305faa\") " pod="kube-system/kube-proxy-k8h5v" Jan 14 00:47:28.142152 kubelet[2802]: I0114 00:47:28.139807 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g58n7\" (UniqueName: \"kubernetes.io/projected/bd1ee5cf-1728-44da-9a75-5fb88c305faa-kube-api-access-g58n7\") pod \"kube-proxy-k8h5v\" (UID: \"bd1ee5cf-1728-44da-9a75-5fb88c305faa\") " pod="kube-system/kube-proxy-k8h5v" Jan 14 00:47:28.220183 systemd[1]: Created slice kubepods-besteffort-pod77a91f56_ee0d_4fca_9d72_88b0d190f818.slice - libcontainer container kubepods-besteffort-pod77a91f56_ee0d_4fca_9d72_88b0d190f818.slice. Jan 14 00:47:28.241091 kubelet[2802]: I0114 00:47:28.240661 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp574\" (UniqueName: \"kubernetes.io/projected/77a91f56-ee0d-4fca-9d72-88b0d190f818-kube-api-access-wp574\") pod \"tigera-operator-7dcd859c48-dkzxt\" (UID: \"77a91f56-ee0d-4fca-9d72-88b0d190f818\") " pod="tigera-operator/tigera-operator-7dcd859c48-dkzxt" Jan 14 00:47:28.241091 kubelet[2802]: I0114 00:47:28.240690 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/77a91f56-ee0d-4fca-9d72-88b0d190f818-var-lib-calico\") pod \"tigera-operator-7dcd859c48-dkzxt\" (UID: \"77a91f56-ee0d-4fca-9d72-88b0d190f818\") " pod="tigera-operator/tigera-operator-7dcd859c48-dkzxt" Jan 14 00:47:28.390988 kubelet[2802]: E0114 00:47:28.390857 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:28.393780 containerd[1611]: time="2026-01-14T00:47:28.393723125Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-k8h5v,Uid:bd1ee5cf-1728-44da-9a75-5fb88c305faa,Namespace:kube-system,Attempt:0,}" Jan 14 00:47:28.414507 containerd[1611]: time="2026-01-14T00:47:28.414452146Z" level=info msg="connecting to shim 6e7a31512819570ab1c9ccb84591e8e8410c846a3e6cc4c0db904931c110d22d" address="unix:///run/containerd/s/71d34e4f247be7666acfec8a5bbb8a6519c2d43b435cb20b78d113646d86d2ea" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:47:28.453274 systemd[1]: Started cri-containerd-6e7a31512819570ab1c9ccb84591e8e8410c846a3e6cc4c0db904931c110d22d.scope - libcontainer container 6e7a31512819570ab1c9ccb84591e8e8410c846a3e6cc4c0db904931c110d22d. Jan 14 00:47:28.464000 audit: BPF prog-id=145 op=LOAD Jan 14 00:47:28.469248 kernel: kauditd_printk_skb: 42 callbacks suppressed Jan 14 00:47:28.469323 kernel: audit: type=1334 audit(1768351648.464:451): prog-id=145 op=LOAD Jan 14 00:47:28.471629 kernel: audit: type=1334 audit(1768351648.464:452): prog-id=146 op=LOAD Jan 14 00:47:28.464000 audit: BPF prog-id=146 op=LOAD Jan 14 00:47:28.464000 audit[2873]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=2862 pid=2873 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.473552 kernel: audit: type=1300 audit(1768351648.464:452): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=2862 pid=2873 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.483419 kernel: audit: type=1327 audit(1768351648.464:452): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665376133313531323831393537306162316339636362383435393165 Jan 14 00:47:28.464000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665376133313531323831393537306162316339636362383435393165 Jan 14 00:47:28.489182 kernel: audit: type=1334 audit(1768351648.465:453): prog-id=146 op=UNLOAD Jan 14 00:47:28.465000 audit: BPF prog-id=146 op=UNLOAD Jan 14 00:47:28.498134 kernel: audit: type=1300 audit(1768351648.465:453): arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2862 pid=2873 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.465000 audit[2873]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2862 pid=2873 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.465000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665376133313531323831393537306162316339636362383435393165 Jan 14 00:47:28.500258 kernel: audit: type=1327 audit(1768351648.465:453): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665376133313531323831393537306162316339636362383435393165 Jan 14 00:47:28.465000 audit: BPF prog-id=147 op=LOAD Jan 14 00:47:28.465000 audit[2873]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=2862 pid=2873 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.510553 kernel: audit: type=1334 audit(1768351648.465:454): prog-id=147 op=LOAD Jan 14 00:47:28.510695 kernel: audit: type=1300 audit(1768351648.465:454): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=2862 pid=2873 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.512381 containerd[1611]: time="2026-01-14T00:47:28.512338303Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-k8h5v,Uid:bd1ee5cf-1728-44da-9a75-5fb88c305faa,Namespace:kube-system,Attempt:0,} returns sandbox id \"6e7a31512819570ab1c9ccb84591e8e8410c846a3e6cc4c0db904931c110d22d\"" Jan 14 00:47:28.513640 kubelet[2802]: E0114 00:47:28.513595 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:28.517673 containerd[1611]: time="2026-01-14T00:47:28.517218875Z" level=info msg="CreateContainer within sandbox \"6e7a31512819570ab1c9ccb84591e8e8410c846a3e6cc4c0db904931c110d22d\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jan 14 00:47:28.465000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665376133313531323831393537306162316339636362383435393165 Jan 14 00:47:28.526217 containerd[1611]: time="2026-01-14T00:47:28.526099380Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-dkzxt,Uid:77a91f56-ee0d-4fca-9d72-88b0d190f818,Namespace:tigera-operator,Attempt:0,}" Jan 14 00:47:28.527131 kernel: audit: type=1327 audit(1768351648.465:454): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665376133313531323831393537306162316339636362383435393165 Jan 14 00:47:28.465000 audit: BPF prog-id=148 op=LOAD Jan 14 00:47:28.465000 audit[2873]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=2862 pid=2873 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.465000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665376133313531323831393537306162316339636362383435393165 Jan 14 00:47:28.465000 audit: BPF prog-id=148 op=UNLOAD Jan 14 00:47:28.465000 audit[2873]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2862 pid=2873 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.465000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665376133313531323831393537306162316339636362383435393165 Jan 14 00:47:28.465000 audit: BPF prog-id=147 op=UNLOAD Jan 14 00:47:28.465000 audit[2873]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2862 pid=2873 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.465000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665376133313531323831393537306162316339636362383435393165 Jan 14 00:47:28.465000 audit: BPF prog-id=149 op=LOAD Jan 14 00:47:28.465000 audit[2873]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=2862 pid=2873 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.465000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3665376133313531323831393537306162316339636362383435393165 Jan 14 00:47:28.543564 containerd[1611]: time="2026-01-14T00:47:28.543472799Z" level=info msg="Container 17cb69a2a9a3e6692f45343b376891b7e646c675d15dadcb3822d6277649d8cd: CDI devices from CRI Config.CDIDevices: []" Jan 14 00:47:28.554897 containerd[1611]: time="2026-01-14T00:47:28.554795220Z" level=info msg="CreateContainer within sandbox \"6e7a31512819570ab1c9ccb84591e8e8410c846a3e6cc4c0db904931c110d22d\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"17cb69a2a9a3e6692f45343b376891b7e646c675d15dadcb3822d6277649d8cd\"" Jan 14 00:47:28.556131 containerd[1611]: time="2026-01-14T00:47:28.556095216Z" level=info msg="StartContainer for \"17cb69a2a9a3e6692f45343b376891b7e646c675d15dadcb3822d6277649d8cd\"" Jan 14 00:47:28.557937 containerd[1611]: time="2026-01-14T00:47:28.557898261Z" level=info msg="connecting to shim 17cb69a2a9a3e6692f45343b376891b7e646c675d15dadcb3822d6277649d8cd" address="unix:///run/containerd/s/71d34e4f247be7666acfec8a5bbb8a6519c2d43b435cb20b78d113646d86d2ea" protocol=ttrpc version=3 Jan 14 00:47:28.566259 containerd[1611]: time="2026-01-14T00:47:28.565398668Z" level=info msg="connecting to shim 6b939fa46103d5087de09dbfce6028b2150265f05a06ee8b81f81af302a2d066" address="unix:///run/containerd/s/1449628e5bc16bbcc6dd3bb66104ce9621782c57b3fc8600745d93c52146c676" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:47:28.586499 systemd[1]: Started cri-containerd-17cb69a2a9a3e6692f45343b376891b7e646c675d15dadcb3822d6277649d8cd.scope - libcontainer container 17cb69a2a9a3e6692f45343b376891b7e646c675d15dadcb3822d6277649d8cd. Jan 14 00:47:28.602410 systemd[1]: Started cri-containerd-6b939fa46103d5087de09dbfce6028b2150265f05a06ee8b81f81af302a2d066.scope - libcontainer container 6b939fa46103d5087de09dbfce6028b2150265f05a06ee8b81f81af302a2d066. Jan 14 00:47:28.623000 audit: BPF prog-id=150 op=LOAD Jan 14 00:47:28.624000 audit: BPF prog-id=151 op=LOAD Jan 14 00:47:28.624000 audit[2931]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000138238 a2=98 a3=0 items=0 ppid=2908 pid=2931 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.624000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3662393339666134363130336435303837646530396462666365363032 Jan 14 00:47:28.624000 audit: BPF prog-id=151 op=UNLOAD Jan 14 00:47:28.624000 audit[2931]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2908 pid=2931 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.624000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3662393339666134363130336435303837646530396462666365363032 Jan 14 00:47:28.625000 audit: BPF prog-id=152 op=LOAD Jan 14 00:47:28.625000 audit[2931]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000138488 a2=98 a3=0 items=0 ppid=2908 pid=2931 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.625000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3662393339666134363130336435303837646530396462666365363032 Jan 14 00:47:28.625000 audit: BPF prog-id=153 op=LOAD Jan 14 00:47:28.625000 audit[2931]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000138218 a2=98 a3=0 items=0 ppid=2908 pid=2931 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.625000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3662393339666134363130336435303837646530396462666365363032 Jan 14 00:47:28.625000 audit: BPF prog-id=153 op=UNLOAD Jan 14 00:47:28.625000 audit[2931]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2908 pid=2931 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.625000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3662393339666134363130336435303837646530396462666365363032 Jan 14 00:47:28.625000 audit: BPF prog-id=152 op=UNLOAD Jan 14 00:47:28.625000 audit[2931]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2908 pid=2931 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.625000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3662393339666134363130336435303837646530396462666365363032 Jan 14 00:47:28.625000 audit: BPF prog-id=154 op=LOAD Jan 14 00:47:28.625000 audit[2931]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001386e8 a2=98 a3=0 items=0 ppid=2908 pid=2931 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.625000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3662393339666134363130336435303837646530396462666365363032 Jan 14 00:47:28.644000 audit: BPF prog-id=155 op=LOAD Jan 14 00:47:28.644000 audit[2903]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000228488 a2=98 a3=0 items=0 ppid=2862 pid=2903 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.644000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3137636236396132613961336536363932663435333433623337363839 Jan 14 00:47:28.645000 audit: BPF prog-id=156 op=LOAD Jan 14 00:47:28.645000 audit[2903]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000228218 a2=98 a3=0 items=0 ppid=2862 pid=2903 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.645000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3137636236396132613961336536363932663435333433623337363839 Jan 14 00:47:28.645000 audit: BPF prog-id=156 op=UNLOAD Jan 14 00:47:28.645000 audit[2903]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=2862 pid=2903 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.645000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3137636236396132613961336536363932663435333433623337363839 Jan 14 00:47:28.645000 audit: BPF prog-id=155 op=UNLOAD Jan 14 00:47:28.645000 audit[2903]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=2862 pid=2903 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.645000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3137636236396132613961336536363932663435333433623337363839 Jan 14 00:47:28.645000 audit: BPF prog-id=157 op=LOAD Jan 14 00:47:28.645000 audit[2903]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0002286e8 a2=98 a3=0 items=0 ppid=2862 pid=2903 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.645000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3137636236396132613961336536363932663435333433623337363839 Jan 14 00:47:28.682866 containerd[1611]: time="2026-01-14T00:47:28.682826627Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7dcd859c48-dkzxt,Uid:77a91f56-ee0d-4fca-9d72-88b0d190f818,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"6b939fa46103d5087de09dbfce6028b2150265f05a06ee8b81f81af302a2d066\"" Jan 14 00:47:28.686739 containerd[1611]: time="2026-01-14T00:47:28.686704560Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\"" Jan 14 00:47:28.687770 containerd[1611]: time="2026-01-14T00:47:28.687740994Z" level=info msg="StartContainer for \"17cb69a2a9a3e6692f45343b376891b7e646c675d15dadcb3822d6277649d8cd\" returns successfully" Jan 14 00:47:28.704787 kubelet[2802]: E0114 00:47:28.704754 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:28.842000 audit[3010]: NETFILTER_CFG table=mangle:54 family=10 entries=1 op=nft_register_chain pid=3010 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:28.842000 audit[3010]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffd761be200 a2=0 a3=7ffd761be1ec items=0 ppid=2943 pid=3010 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.842000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Jan 14 00:47:28.845000 audit[3012]: NETFILTER_CFG table=nat:55 family=10 entries=1 op=nft_register_chain pid=3012 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:28.845000 audit[3012]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffffcd369f0 a2=0 a3=7ffffcd369dc items=0 ppid=2943 pid=3012 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.845000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Jan 14 00:47:28.845000 audit[3011]: NETFILTER_CFG table=mangle:56 family=2 entries=1 op=nft_register_chain pid=3011 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.845000 audit[3011]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffc09bf8d80 a2=0 a3=7ffc09bf8d6c items=0 ppid=2943 pid=3011 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.845000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Jan 14 00:47:28.846000 audit[3013]: NETFILTER_CFG table=filter:57 family=10 entries=1 op=nft_register_chain pid=3013 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:28.846000 audit[3013]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffc2f011620 a2=0 a3=7ffc2f01160c items=0 ppid=2943 pid=3013 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.846000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Jan 14 00:47:28.848000 audit[3015]: NETFILTER_CFG table=nat:58 family=2 entries=1 op=nft_register_chain pid=3015 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.848000 audit[3015]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffcd9d8c8c0 a2=0 a3=7ffcd9d8c8ac items=0 ppid=2943 pid=3015 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.848000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Jan 14 00:47:28.850000 audit[3017]: NETFILTER_CFG table=filter:59 family=2 entries=1 op=nft_register_chain pid=3017 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.850000 audit[3017]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffc59bc1060 a2=0 a3=7ffc59bc104c items=0 ppid=2943 pid=3017 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.850000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Jan 14 00:47:28.953000 audit[3019]: NETFILTER_CFG table=filter:60 family=2 entries=1 op=nft_register_chain pid=3019 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.953000 audit[3019]: SYSCALL arch=c000003e syscall=46 success=yes exit=108 a0=3 a1=7ffddd5721e0 a2=0 a3=7ffddd5721cc items=0 ppid=2943 pid=3019 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.953000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 Jan 14 00:47:28.957000 audit[3021]: NETFILTER_CFG table=filter:61 family=2 entries=1 op=nft_register_rule pid=3021 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.957000 audit[3021]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffdacb4e610 a2=0 a3=7ffdacb4e5fc items=0 ppid=2943 pid=3021 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.957000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C652073657276696365 Jan 14 00:47:28.962000 audit[3024]: NETFILTER_CFG table=filter:62 family=2 entries=1 op=nft_register_rule pid=3024 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.962000 audit[3024]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffc6df7a4a0 a2=0 a3=7ffc6df7a48c items=0 ppid=2943 pid=3024 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.962000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669 Jan 14 00:47:28.964000 audit[3025]: NETFILTER_CFG table=filter:63 family=2 entries=1 op=nft_register_chain pid=3025 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.964000 audit[3025]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe222c6330 a2=0 a3=7ffe222c631c items=0 ppid=2943 pid=3025 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.964000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4E4F4445504F525453002D740066696C746572 Jan 14 00:47:28.968000 audit[3027]: NETFILTER_CFG table=filter:64 family=2 entries=1 op=nft_register_rule pid=3027 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.968000 audit[3027]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffcafee1410 a2=0 a3=7ffcafee13fc items=0 ppid=2943 pid=3027 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.968000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 Jan 14 00:47:28.969000 audit[3028]: NETFILTER_CFG table=filter:65 family=2 entries=1 op=nft_register_chain pid=3028 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.969000 audit[3028]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff0f0b6fa0 a2=0 a3=7fff0f0b6f8c items=0 ppid=2943 pid=3028 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.969000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D740066696C746572 Jan 14 00:47:28.973000 audit[3030]: NETFILTER_CFG table=filter:66 family=2 entries=1 op=nft_register_rule pid=3030 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.973000 audit[3030]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffd69eccec0 a2=0 a3=7ffd69ecceac items=0 ppid=2943 pid=3030 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.973000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D Jan 14 00:47:28.978000 audit[3033]: NETFILTER_CFG table=filter:67 family=2 entries=1 op=nft_register_rule pid=3033 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.978000 audit[3033]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffe96fa8110 a2=0 a3=7ffe96fa80fc items=0 ppid=2943 pid=3033 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.978000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D53 Jan 14 00:47:28.980000 audit[3034]: NETFILTER_CFG table=filter:68 family=2 entries=1 op=nft_register_chain pid=3034 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.980000 audit[3034]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd207075b0 a2=0 a3=7ffd2070759c items=0 ppid=2943 pid=3034 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.980000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D464F5257415244002D740066696C746572 Jan 14 00:47:28.983000 audit[3036]: NETFILTER_CFG table=filter:69 family=2 entries=1 op=nft_register_rule pid=3036 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.983000 audit[3036]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffdb6765440 a2=0 a3=7ffdb676542c items=0 ppid=2943 pid=3036 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.983000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 Jan 14 00:47:28.985000 audit[3037]: NETFILTER_CFG table=filter:70 family=2 entries=1 op=nft_register_chain pid=3037 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.985000 audit[3037]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffe12106760 a2=0 a3=7ffe1210674c items=0 ppid=2943 pid=3037 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.985000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 Jan 14 00:47:28.988000 audit[3039]: NETFILTER_CFG table=filter:71 family=2 entries=1 op=nft_register_rule pid=3039 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.988000 audit[3039]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffe67f43da0 a2=0 a3=7ffe67f43d8c items=0 ppid=2943 pid=3039 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.988000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Jan 14 00:47:28.994000 audit[3042]: NETFILTER_CFG table=filter:72 family=2 entries=1 op=nft_register_rule pid=3042 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:28.994000 audit[3042]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffdbe312490 a2=0 a3=7ffdbe31247c items=0 ppid=2943 pid=3042 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:28.994000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Jan 14 00:47:29.000000 audit[3045]: NETFILTER_CFG table=filter:73 family=2 entries=1 op=nft_register_rule pid=3045 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:29.000000 audit[3045]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7fffcc570610 a2=0 a3=7fffcc5705fc items=0 ppid=2943 pid=3045 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.000000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D Jan 14 00:47:29.001000 audit[3046]: NETFILTER_CFG table=nat:74 family=2 entries=1 op=nft_register_chain pid=3046 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:29.001000 audit[3046]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffd78468160 a2=0 a3=7ffd7846814c items=0 ppid=2943 pid=3046 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.001000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D74006E6174 Jan 14 00:47:29.005000 audit[3048]: NETFILTER_CFG table=nat:75 family=2 entries=1 op=nft_register_rule pid=3048 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:29.005000 audit[3048]: SYSCALL arch=c000003e syscall=46 success=yes exit=524 a0=3 a1=7ffdd6ed4510 a2=0 a3=7ffdd6ed44fc items=0 ppid=2943 pid=3048 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.005000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 14 00:47:29.010000 audit[3051]: NETFILTER_CFG table=nat:76 family=2 entries=1 op=nft_register_rule pid=3051 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:29.010000 audit[3051]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffd6d63c630 a2=0 a3=7ffd6d63c61c items=0 ppid=2943 pid=3051 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.010000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 14 00:47:29.012000 audit[3052]: NETFILTER_CFG table=nat:77 family=2 entries=1 op=nft_register_chain pid=3052 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:29.012000 audit[3052]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd2024b660 a2=0 a3=7ffd2024b64c items=0 ppid=2943 pid=3052 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.012000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Jan 14 00:47:29.016000 audit[3054]: NETFILTER_CFG table=nat:78 family=2 entries=1 op=nft_register_rule pid=3054 subj=system_u:system_r:kernel_t:s0 comm="iptables" Jan 14 00:47:29.016000 audit[3054]: SYSCALL arch=c000003e syscall=46 success=yes exit=532 a0=3 a1=7fffcb0b2d80 a2=0 a3=7fffcb0b2d6c items=0 ppid=2943 pid=3054 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.016000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 Jan 14 00:47:29.049000 audit[3060]: NETFILTER_CFG table=filter:79 family=2 entries=8 op=nft_register_rule pid=3060 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:29.049000 audit[3060]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7fff82f9d610 a2=0 a3=7fff82f9d5fc items=0 ppid=2943 pid=3060 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.049000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:29.057000 audit[3060]: NETFILTER_CFG table=nat:80 family=2 entries=14 op=nft_register_chain pid=3060 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:29.057000 audit[3060]: SYSCALL arch=c000003e syscall=46 success=yes exit=5508 a0=3 a1=7fff82f9d610 a2=0 a3=7fff82f9d5fc items=0 ppid=2943 pid=3060 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.057000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:29.059000 audit[3065]: NETFILTER_CFG table=filter:81 family=10 entries=1 op=nft_register_chain pid=3065 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.059000 audit[3065]: SYSCALL arch=c000003e syscall=46 success=yes exit=108 a0=3 a1=7ffff8675dc0 a2=0 a3=7ffff8675dac items=0 ppid=2943 pid=3065 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.059000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 Jan 14 00:47:29.064000 audit[3067]: NETFILTER_CFG table=filter:82 family=10 entries=2 op=nft_register_chain pid=3067 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.064000 audit[3067]: SYSCALL arch=c000003e syscall=46 success=yes exit=836 a0=3 a1=7ffc91baed60 a2=0 a3=7ffc91baed4c items=0 ppid=2943 pid=3067 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.064000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C6520736572766963 Jan 14 00:47:29.069000 audit[3070]: NETFILTER_CFG table=filter:83 family=10 entries=1 op=nft_register_rule pid=3070 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.069000 audit[3070]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffc6324f080 a2=0 a3=7ffc6324f06c items=0 ppid=2943 pid=3070 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.069000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C652073657276 Jan 14 00:47:29.071000 audit[3071]: NETFILTER_CFG table=filter:84 family=10 entries=1 op=nft_register_chain pid=3071 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.071000 audit[3071]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd5b532620 a2=0 a3=7ffd5b53260c items=0 ppid=2943 pid=3071 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.071000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4E4F4445504F525453002D740066696C746572 Jan 14 00:47:29.075000 audit[3073]: NETFILTER_CFG table=filter:85 family=10 entries=1 op=nft_register_rule pid=3073 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.075000 audit[3073]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffed877b330 a2=0 a3=7ffed877b31c items=0 ppid=2943 pid=3073 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.075000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 Jan 14 00:47:29.077000 audit[3074]: NETFILTER_CFG table=filter:86 family=10 entries=1 op=nft_register_chain pid=3074 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.077000 audit[3074]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc42a63ae0 a2=0 a3=7ffc42a63acc items=0 ppid=2943 pid=3074 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.077000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D740066696C746572 Jan 14 00:47:29.080000 audit[3076]: NETFILTER_CFG table=filter:87 family=10 entries=1 op=nft_register_rule pid=3076 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.080000 audit[3076]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffce5df05c0 a2=0 a3=7ffce5df05ac items=0 ppid=2943 pid=3076 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.080000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B554245 Jan 14 00:47:29.086000 audit[3079]: NETFILTER_CFG table=filter:88 family=10 entries=2 op=nft_register_chain pid=3079 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.086000 audit[3079]: SYSCALL arch=c000003e syscall=46 success=yes exit=828 a0=3 a1=7ffd28daf490 a2=0 a3=7ffd28daf47c items=0 ppid=2943 pid=3079 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.086000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D Jan 14 00:47:29.088000 audit[3080]: NETFILTER_CFG table=filter:89 family=10 entries=1 op=nft_register_chain pid=3080 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.088000 audit[3080]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffcba2611c0 a2=0 a3=7ffcba2611ac items=0 ppid=2943 pid=3080 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.088000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D464F5257415244002D740066696C746572 Jan 14 00:47:29.092000 audit[3082]: NETFILTER_CFG table=filter:90 family=10 entries=1 op=nft_register_rule pid=3082 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.092000 audit[3082]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffeba7b9380 a2=0 a3=7ffeba7b936c items=0 ppid=2943 pid=3082 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.092000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 Jan 14 00:47:29.094000 audit[3083]: NETFILTER_CFG table=filter:91 family=10 entries=1 op=nft_register_chain pid=3083 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.094000 audit[3083]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffd1c94ab00 a2=0 a3=7ffd1c94aaec items=0 ppid=2943 pid=3083 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.094000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 Jan 14 00:47:29.097000 audit[3085]: NETFILTER_CFG table=filter:92 family=10 entries=1 op=nft_register_rule pid=3085 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.097000 audit[3085]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffc55debbb0 a2=0 a3=7ffc55debb9c items=0 ppid=2943 pid=3085 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.097000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Jan 14 00:47:29.101000 audit[3088]: NETFILTER_CFG table=filter:93 family=10 entries=1 op=nft_register_rule pid=3088 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.101000 audit[3088]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffd1f1216b0 a2=0 a3=7ffd1f12169c items=0 ppid=2943 pid=3088 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.101000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D Jan 14 00:47:29.106000 audit[3091]: NETFILTER_CFG table=filter:94 family=10 entries=1 op=nft_register_rule pid=3091 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.106000 audit[3091]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffd7a6618e0 a2=0 a3=7ffd7a6618cc items=0 ppid=2943 pid=3091 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.106000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C Jan 14 00:47:29.108000 audit[3092]: NETFILTER_CFG table=nat:95 family=10 entries=1 op=nft_register_chain pid=3092 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.108000 audit[3092]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffe0df84f60 a2=0 a3=7ffe0df84f4c items=0 ppid=2943 pid=3092 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.108000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D74006E6174 Jan 14 00:47:29.111000 audit[3094]: NETFILTER_CFG table=nat:96 family=10 entries=1 op=nft_register_rule pid=3094 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.111000 audit[3094]: SYSCALL arch=c000003e syscall=46 success=yes exit=524 a0=3 a1=7fff6a24f010 a2=0 a3=7fff6a24effc items=0 ppid=2943 pid=3094 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.111000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 14 00:47:29.117000 audit[3097]: NETFILTER_CFG table=nat:97 family=10 entries=1 op=nft_register_rule pid=3097 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.117000 audit[3097]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffcee919e60 a2=0 a3=7ffcee919e4c items=0 ppid=2943 pid=3097 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.117000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Jan 14 00:47:29.118000 audit[3098]: NETFILTER_CFG table=nat:98 family=10 entries=1 op=nft_register_chain pid=3098 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.118000 audit[3098]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff7c66bad0 a2=0 a3=7fff7c66babc items=0 ppid=2943 pid=3098 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.118000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Jan 14 00:47:29.123000 audit[3100]: NETFILTER_CFG table=nat:99 family=10 entries=2 op=nft_register_chain pid=3100 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.123000 audit[3100]: SYSCALL arch=c000003e syscall=46 success=yes exit=612 a0=3 a1=7ffd511e84d0 a2=0 a3=7ffd511e84bc items=0 ppid=2943 pid=3100 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.123000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 Jan 14 00:47:29.125000 audit[3101]: NETFILTER_CFG table=filter:100 family=10 entries=1 op=nft_register_chain pid=3101 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.125000 audit[3101]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd7712e750 a2=0 a3=7ffd7712e73c items=0 ppid=2943 pid=3101 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.125000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 Jan 14 00:47:29.130000 audit[3103]: NETFILTER_CFG table=filter:101 family=10 entries=1 op=nft_register_rule pid=3103 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.130000 audit[3103]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7ffd9e0d11a0 a2=0 a3=7ffd9e0d118c items=0 ppid=2943 pid=3103 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.130000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 14 00:47:29.136000 audit[3106]: NETFILTER_CFG table=filter:102 family=10 entries=1 op=nft_register_rule pid=3106 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Jan 14 00:47:29.136000 audit[3106]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7ffe50433a80 a2=0 a3=7ffe50433a6c items=0 ppid=2943 pid=3106 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.136000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Jan 14 00:47:29.141000 audit[3108]: NETFILTER_CFG table=filter:103 family=10 entries=3 op=nft_register_rule pid=3108 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" Jan 14 00:47:29.141000 audit[3108]: SYSCALL arch=c000003e syscall=46 success=yes exit=2088 a0=3 a1=7ffec0e144a0 a2=0 a3=7ffec0e1448c items=0 ppid=2943 pid=3108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.141000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:29.142000 audit[3108]: NETFILTER_CFG table=nat:104 family=10 entries=7 op=nft_register_chain pid=3108 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" Jan 14 00:47:29.142000 audit[3108]: SYSCALL arch=c000003e syscall=46 success=yes exit=2056 a0=3 a1=7ffec0e144a0 a2=0 a3=7ffec0e1448c items=0 ppid=2943 pid=3108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:29.142000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:29.255359 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount397835572.mount: Deactivated successfully. Jan 14 00:47:29.561916 kubelet[2802]: E0114 00:47:29.561801 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:29.564940 kubelet[2802]: E0114 00:47:29.564908 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:29.583319 kubelet[2802]: I0114 00:47:29.582822 2802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-k8h5v" podStartSLOduration=1.582808623 podStartE2EDuration="1.582808623s" podCreationTimestamp="2026-01-14 00:47:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-14 00:47:29.57315398 +0000 UTC m=+8.168250394" watchObservedRunningTime="2026-01-14 00:47:29.582808623 +0000 UTC m=+8.177905037" Jan 14 00:47:29.795953 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2866679198.mount: Deactivated successfully. Jan 14 00:47:30.565843 kubelet[2802]: E0114 00:47:30.565710 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:31.032455 containerd[1611]: time="2026-01-14T00:47:31.032409844Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:31.033446 containerd[1611]: time="2026-01-14T00:47:31.033207580Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.7: active requests=0, bytes read=23558205" Jan 14 00:47:31.033916 containerd[1611]: time="2026-01-14T00:47:31.033885661Z" level=info msg="ImageCreate event name:\"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:31.036015 containerd[1611]: time="2026-01-14T00:47:31.035986841Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:31.036543 containerd[1611]: time="2026-01-14T00:47:31.036510031Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.7\" with image id \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\", repo tag \"quay.io/tigera/operator:v1.38.7\", repo digest \"quay.io/tigera/operator@sha256:1b629a1403f5b6d7243f7dd523d04b8a50352a33c1d4d6970b6002a8733acf2e\", size \"25057686\" in 2.349667489s" Jan 14 00:47:31.036584 containerd[1611]: time="2026-01-14T00:47:31.036543195Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.7\" returns image reference \"sha256:f2c1be207523e593db82e3b8cf356a12f3ad8d1aad2225f8114b2cf9d6486cf1\"" Jan 14 00:47:31.041540 containerd[1611]: time="2026-01-14T00:47:31.041496467Z" level=info msg="CreateContainer within sandbox \"6b939fa46103d5087de09dbfce6028b2150265f05a06ee8b81f81af302a2d066\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Jan 14 00:47:31.051142 containerd[1611]: time="2026-01-14T00:47:31.049938054Z" level=info msg="Container 4ff6da8fbea0ac41dfd4c3a822e0a627c6acd2c8f78c9fa2e35c6f985945090a: CDI devices from CRI Config.CDIDevices: []" Jan 14 00:47:31.053855 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2940682642.mount: Deactivated successfully. Jan 14 00:47:31.065918 containerd[1611]: time="2026-01-14T00:47:31.065892714Z" level=info msg="CreateContainer within sandbox \"6b939fa46103d5087de09dbfce6028b2150265f05a06ee8b81f81af302a2d066\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"4ff6da8fbea0ac41dfd4c3a822e0a627c6acd2c8f78c9fa2e35c6f985945090a\"" Jan 14 00:47:31.066509 containerd[1611]: time="2026-01-14T00:47:31.066434216Z" level=info msg="StartContainer for \"4ff6da8fbea0ac41dfd4c3a822e0a627c6acd2c8f78c9fa2e35c6f985945090a\"" Jan 14 00:47:31.068133 containerd[1611]: time="2026-01-14T00:47:31.068058472Z" level=info msg="connecting to shim 4ff6da8fbea0ac41dfd4c3a822e0a627c6acd2c8f78c9fa2e35c6f985945090a" address="unix:///run/containerd/s/1449628e5bc16bbcc6dd3bb66104ce9621782c57b3fc8600745d93c52146c676" protocol=ttrpc version=3 Jan 14 00:47:31.098299 systemd[1]: Started cri-containerd-4ff6da8fbea0ac41dfd4c3a822e0a627c6acd2c8f78c9fa2e35c6f985945090a.scope - libcontainer container 4ff6da8fbea0ac41dfd4c3a822e0a627c6acd2c8f78c9fa2e35c6f985945090a. Jan 14 00:47:31.112000 audit: BPF prog-id=158 op=LOAD Jan 14 00:47:31.113000 audit: BPF prog-id=159 op=LOAD Jan 14 00:47:31.113000 audit[3117]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8238 a2=98 a3=0 items=0 ppid=2908 pid=3117 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:31.113000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3466663664613866626561306163343164666434633361383232653061 Jan 14 00:47:31.113000 audit: BPF prog-id=159 op=UNLOAD Jan 14 00:47:31.113000 audit[3117]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2908 pid=3117 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:31.113000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3466663664613866626561306163343164666434633361383232653061 Jan 14 00:47:31.113000 audit: BPF prog-id=160 op=LOAD Jan 14 00:47:31.113000 audit[3117]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a8488 a2=98 a3=0 items=0 ppid=2908 pid=3117 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:31.113000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3466663664613866626561306163343164666434633361383232653061 Jan 14 00:47:31.113000 audit: BPF prog-id=161 op=LOAD Jan 14 00:47:31.113000 audit[3117]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c0001a8218 a2=98 a3=0 items=0 ppid=2908 pid=3117 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:31.113000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3466663664613866626561306163343164666434633361383232653061 Jan 14 00:47:31.113000 audit: BPF prog-id=161 op=UNLOAD Jan 14 00:47:31.113000 audit[3117]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=2908 pid=3117 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:31.113000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3466663664613866626561306163343164666434633361383232653061 Jan 14 00:47:31.113000 audit: BPF prog-id=160 op=UNLOAD Jan 14 00:47:31.113000 audit[3117]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=2908 pid=3117 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:31.113000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3466663664613866626561306163343164666434633361383232653061 Jan 14 00:47:31.113000 audit: BPF prog-id=162 op=LOAD Jan 14 00:47:31.113000 audit[3117]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001a86e8 a2=98 a3=0 items=0 ppid=2908 pid=3117 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:31.113000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3466663664613866626561306163343164666434633361383232653061 Jan 14 00:47:31.133398 containerd[1611]: time="2026-01-14T00:47:31.133359419Z" level=info msg="StartContainer for \"4ff6da8fbea0ac41dfd4c3a822e0a627c6acd2c8f78c9fa2e35c6f985945090a\" returns successfully" Jan 14 00:47:31.580987 kubelet[2802]: I0114 00:47:31.580812 2802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7dcd859c48-dkzxt" podStartSLOduration=1.228656743 podStartE2EDuration="3.580789698s" podCreationTimestamp="2026-01-14 00:47:28 +0000 UTC" firstStartedPulling="2026-01-14 00:47:28.685312451 +0000 UTC m=+7.280408865" lastFinishedPulling="2026-01-14 00:47:31.037445406 +0000 UTC m=+9.632541820" observedRunningTime="2026-01-14 00:47:31.58072841 +0000 UTC m=+10.175824824" watchObservedRunningTime="2026-01-14 00:47:31.580789698 +0000 UTC m=+10.175886112" Jan 14 00:47:35.457660 kubelet[2802]: E0114 00:47:35.457194 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:36.715234 sudo[1866]: pam_unix(sudo:session): session closed for user root Jan 14 00:47:36.718400 kernel: kauditd_printk_skb: 224 callbacks suppressed Jan 14 00:47:36.718481 kernel: audit: type=1106 audit(1768351656.714:531): pid=1866 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 00:47:36.714000 audit[1866]: USER_END pid=1866 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_umask,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 00:47:36.731980 kernel: audit: type=1104 audit(1768351656.714:532): pid=1866 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 00:47:36.714000 audit[1866]: CRED_DISP pid=1866 uid=500 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Jan 14 00:47:36.739134 sshd[1865]: Connection closed by 68.220.241.50 port 34778 Jan 14 00:47:36.739757 sshd-session[1861]: pam_unix(sshd:session): session closed for user core Jan 14 00:47:36.741000 audit[1861]: USER_END pid=1861 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:47:36.762311 kernel: audit: type=1106 audit(1768351656.741:533): pid=1861 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:47:36.762377 kernel: audit: type=1104 audit(1768351656.741:534): pid=1861 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:47:36.741000 audit[1861]: CRED_DISP pid=1861 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:47:36.761597 systemd[1]: sshd@6-172.236.110.177:22-68.220.241.50:34778.service: Deactivated successfully. Jan 14 00:47:36.772171 kernel: audit: type=1131 audit(1768351656.761:535): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-172.236.110.177:22-68.220.241.50:34778 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:47:36.761000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-172.236.110.177:22-68.220.241.50:34778 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:47:36.769483 systemd[1]: session-8.scope: Deactivated successfully. Jan 14 00:47:36.770407 systemd[1]: session-8.scope: Consumed 5.528s CPU time, 233M memory peak. Jan 14 00:47:36.774911 systemd-logind[1578]: Session 8 logged out. Waiting for processes to exit. Jan 14 00:47:36.776373 systemd-logind[1578]: Removed session 8. Jan 14 00:47:37.593000 audit[3200]: NETFILTER_CFG table=filter:105 family=2 entries=14 op=nft_register_rule pid=3200 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:37.623250 kernel: audit: type=1325 audit(1768351657.593:536): table=filter:105 family=2 entries=14 op=nft_register_rule pid=3200 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:37.593000 audit[3200]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7fff9ac65a40 a2=0 a3=7fff9ac65a2c items=0 ppid=2943 pid=3200 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:37.644289 kernel: audit: type=1300 audit(1768351657.593:536): arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7fff9ac65a40 a2=0 a3=7fff9ac65a2c items=0 ppid=2943 pid=3200 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:37.593000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:37.656129 kernel: audit: type=1327 audit(1768351657.593:536): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:37.633000 audit[3200]: NETFILTER_CFG table=nat:106 family=2 entries=12 op=nft_register_rule pid=3200 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:37.663007 kernel: audit: type=1325 audit(1768351657.633:537): table=nat:106 family=2 entries=12 op=nft_register_rule pid=3200 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:37.633000 audit[3200]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7fff9ac65a40 a2=0 a3=0 items=0 ppid=2943 pid=3200 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:37.633000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:37.673130 kernel: audit: type=1300 audit(1768351657.633:537): arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7fff9ac65a40 a2=0 a3=0 items=0 ppid=2943 pid=3200 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:37.716000 audit[3202]: NETFILTER_CFG table=filter:107 family=2 entries=15 op=nft_register_rule pid=3202 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:37.716000 audit[3202]: SYSCALL arch=c000003e syscall=46 success=yes exit=5992 a0=3 a1=7ffc19ca7860 a2=0 a3=7ffc19ca784c items=0 ppid=2943 pid=3202 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:37.716000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:37.720000 audit[3202]: NETFILTER_CFG table=nat:108 family=2 entries=12 op=nft_register_rule pid=3202 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:37.720000 audit[3202]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffc19ca7860 a2=0 a3=0 items=0 ppid=2943 pid=3202 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:37.720000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:38.706494 update_engine[1581]: I20260114 00:47:38.705538 1581 update_attempter.cc:509] Updating boot flags... Jan 14 00:47:40.120000 audit[3226]: NETFILTER_CFG table=filter:109 family=2 entries=17 op=nft_register_rule pid=3226 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:40.120000 audit[3226]: SYSCALL arch=c000003e syscall=46 success=yes exit=6736 a0=3 a1=7ffe11c2da60 a2=0 a3=7ffe11c2da4c items=0 ppid=2943 pid=3226 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:40.120000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:40.127000 audit[3226]: NETFILTER_CFG table=nat:110 family=2 entries=12 op=nft_register_rule pid=3226 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:40.127000 audit[3226]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffe11c2da60 a2=0 a3=0 items=0 ppid=2943 pid=3226 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:40.127000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:40.151000 audit[3228]: NETFILTER_CFG table=filter:111 family=2 entries=18 op=nft_register_rule pid=3228 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:40.151000 audit[3228]: SYSCALL arch=c000003e syscall=46 success=yes exit=6736 a0=3 a1=7ffe22b54cc0 a2=0 a3=7ffe22b54cac items=0 ppid=2943 pid=3228 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:40.151000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:40.154000 audit[3228]: NETFILTER_CFG table=nat:112 family=2 entries=12 op=nft_register_rule pid=3228 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:40.154000 audit[3228]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffe22b54cc0 a2=0 a3=0 items=0 ppid=2943 pid=3228 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:40.154000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:41.173000 audit[3230]: NETFILTER_CFG table=filter:113 family=2 entries=19 op=nft_register_rule pid=3230 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:41.173000 audit[3230]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffdf8c49f40 a2=0 a3=7ffdf8c49f2c items=0 ppid=2943 pid=3230 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:41.173000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:41.178000 audit[3230]: NETFILTER_CFG table=nat:114 family=2 entries=12 op=nft_register_rule pid=3230 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:41.178000 audit[3230]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffdf8c49f40 a2=0 a3=0 items=0 ppid=2943 pid=3230 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:41.178000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:41.730602 systemd[1]: Created slice kubepods-besteffort-pod2fdc182d_8d82_4e17_a664_8eb8af914c6b.slice - libcontainer container kubepods-besteffort-pod2fdc182d_8d82_4e17_a664_8eb8af914c6b.slice. Jan 14 00:47:41.853418 kubelet[2802]: I0114 00:47:41.853341 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/2fdc182d-8d82-4e17-a664-8eb8af914c6b-typha-certs\") pod \"calico-typha-547ff64fb7-mfz8m\" (UID: \"2fdc182d-8d82-4e17-a664-8eb8af914c6b\") " pod="calico-system/calico-typha-547ff64fb7-mfz8m" Jan 14 00:47:41.854036 kubelet[2802]: I0114 00:47:41.853532 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkvrt\" (UniqueName: \"kubernetes.io/projected/2fdc182d-8d82-4e17-a664-8eb8af914c6b-kube-api-access-jkvrt\") pod \"calico-typha-547ff64fb7-mfz8m\" (UID: \"2fdc182d-8d82-4e17-a664-8eb8af914c6b\") " pod="calico-system/calico-typha-547ff64fb7-mfz8m" Jan 14 00:47:41.854036 kubelet[2802]: I0114 00:47:41.853594 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fdc182d-8d82-4e17-a664-8eb8af914c6b-tigera-ca-bundle\") pod \"calico-typha-547ff64fb7-mfz8m\" (UID: \"2fdc182d-8d82-4e17-a664-8eb8af914c6b\") " pod="calico-system/calico-typha-547ff64fb7-mfz8m" Jan 14 00:47:41.940036 systemd[1]: Created slice kubepods-besteffort-pod12a054c8_0aa7_4bbc_9278_b8219804f0ea.slice - libcontainer container kubepods-besteffort-pod12a054c8_0aa7_4bbc_9278_b8219804f0ea.slice. Jan 14 00:47:42.035189 kubelet[2802]: E0114 00:47:42.035048 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:42.036236 containerd[1611]: time="2026-01-14T00:47:42.036190334Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-547ff64fb7-mfz8m,Uid:2fdc182d-8d82-4e17-a664-8eb8af914c6b,Namespace:calico-system,Attempt:0,}" Jan 14 00:47:42.055672 kubelet[2802]: I0114 00:47:42.055637 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/12a054c8-0aa7-4bbc-9278-b8219804f0ea-flexvol-driver-host\") pod \"calico-node-28tb6\" (UID: \"12a054c8-0aa7-4bbc-9278-b8219804f0ea\") " pod="calico-system/calico-node-28tb6" Jan 14 00:47:42.055773 kubelet[2802]: I0114 00:47:42.055674 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12a054c8-0aa7-4bbc-9278-b8219804f0ea-tigera-ca-bundle\") pod \"calico-node-28tb6\" (UID: \"12a054c8-0aa7-4bbc-9278-b8219804f0ea\") " pod="calico-system/calico-node-28tb6" Jan 14 00:47:42.055773 kubelet[2802]: I0114 00:47:42.055693 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/12a054c8-0aa7-4bbc-9278-b8219804f0ea-xtables-lock\") pod \"calico-node-28tb6\" (UID: \"12a054c8-0aa7-4bbc-9278-b8219804f0ea\") " pod="calico-system/calico-node-28tb6" Jan 14 00:47:42.055773 kubelet[2802]: I0114 00:47:42.055706 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/12a054c8-0aa7-4bbc-9278-b8219804f0ea-cni-net-dir\") pod \"calico-node-28tb6\" (UID: \"12a054c8-0aa7-4bbc-9278-b8219804f0ea\") " pod="calico-system/calico-node-28tb6" Jan 14 00:47:42.055773 kubelet[2802]: I0114 00:47:42.055721 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/12a054c8-0aa7-4bbc-9278-b8219804f0ea-node-certs\") pod \"calico-node-28tb6\" (UID: \"12a054c8-0aa7-4bbc-9278-b8219804f0ea\") " pod="calico-system/calico-node-28tb6" Jan 14 00:47:42.055773 kubelet[2802]: I0114 00:47:42.055734 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/12a054c8-0aa7-4bbc-9278-b8219804f0ea-var-lib-calico\") pod \"calico-node-28tb6\" (UID: \"12a054c8-0aa7-4bbc-9278-b8219804f0ea\") " pod="calico-system/calico-node-28tb6" Jan 14 00:47:42.055889 kubelet[2802]: I0114 00:47:42.055747 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/12a054c8-0aa7-4bbc-9278-b8219804f0ea-var-run-calico\") pod \"calico-node-28tb6\" (UID: \"12a054c8-0aa7-4bbc-9278-b8219804f0ea\") " pod="calico-system/calico-node-28tb6" Jan 14 00:47:42.055889 kubelet[2802]: I0114 00:47:42.055763 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/12a054c8-0aa7-4bbc-9278-b8219804f0ea-cni-bin-dir\") pod \"calico-node-28tb6\" (UID: \"12a054c8-0aa7-4bbc-9278-b8219804f0ea\") " pod="calico-system/calico-node-28tb6" Jan 14 00:47:42.055889 kubelet[2802]: I0114 00:47:42.055776 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/12a054c8-0aa7-4bbc-9278-b8219804f0ea-cni-log-dir\") pod \"calico-node-28tb6\" (UID: \"12a054c8-0aa7-4bbc-9278-b8219804f0ea\") " pod="calico-system/calico-node-28tb6" Jan 14 00:47:42.055889 kubelet[2802]: I0114 00:47:42.055790 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/12a054c8-0aa7-4bbc-9278-b8219804f0ea-lib-modules\") pod \"calico-node-28tb6\" (UID: \"12a054c8-0aa7-4bbc-9278-b8219804f0ea\") " pod="calico-system/calico-node-28tb6" Jan 14 00:47:42.055889 kubelet[2802]: I0114 00:47:42.055805 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/12a054c8-0aa7-4bbc-9278-b8219804f0ea-policysync\") pod \"calico-node-28tb6\" (UID: \"12a054c8-0aa7-4bbc-9278-b8219804f0ea\") " pod="calico-system/calico-node-28tb6" Jan 14 00:47:42.055998 kubelet[2802]: I0114 00:47:42.055819 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr82x\" (UniqueName: \"kubernetes.io/projected/12a054c8-0aa7-4bbc-9278-b8219804f0ea-kube-api-access-zr82x\") pod \"calico-node-28tb6\" (UID: \"12a054c8-0aa7-4bbc-9278-b8219804f0ea\") " pod="calico-system/calico-node-28tb6" Jan 14 00:47:42.079579 containerd[1611]: time="2026-01-14T00:47:42.079420661Z" level=info msg="connecting to shim e1d9854e442bdf1d6cfdd0179e4c957b581bf46954c2f57de639ce9db9fdcfcc" address="unix:///run/containerd/s/731dca5e80e11a07c8d8f546eb19413414689c41052fec85265c8aa7c42a9ae9" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:47:42.126425 systemd[1]: Started cri-containerd-e1d9854e442bdf1d6cfdd0179e4c957b581bf46954c2f57de639ce9db9fdcfcc.scope - libcontainer container e1d9854e442bdf1d6cfdd0179e4c957b581bf46954c2f57de639ce9db9fdcfcc. Jan 14 00:47:42.162934 kubelet[2802]: E0114 00:47:42.162908 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.163204 kubelet[2802]: W0114 00:47:42.162991 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.163204 kubelet[2802]: E0114 00:47:42.163013 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.164174 kubelet[2802]: E0114 00:47:42.164140 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.164174 kubelet[2802]: W0114 00:47:42.164153 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.164375 kubelet[2802]: E0114 00:47:42.164278 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.165155 kubelet[2802]: E0114 00:47:42.164999 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.165155 kubelet[2802]: W0114 00:47:42.165022 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.165155 kubelet[2802]: E0114 00:47:42.165032 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.166328 kubelet[2802]: E0114 00:47:42.165975 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.166328 kubelet[2802]: W0114 00:47:42.165986 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.166328 kubelet[2802]: E0114 00:47:42.165996 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.177917 kubelet[2802]: E0114 00:47:42.174245 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.177917 kubelet[2802]: W0114 00:47:42.174265 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.177917 kubelet[2802]: E0114 00:47:42.174282 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.178459 kubelet[2802]: E0114 00:47:42.178430 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.178459 kubelet[2802]: W0114 00:47:42.178454 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.179670 kubelet[2802]: E0114 00:47:42.178472 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.183516 kubelet[2802]: E0114 00:47:42.183470 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:47:42.200212 kubelet[2802]: E0114 00:47:42.200147 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.200212 kubelet[2802]: W0114 00:47:42.200169 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.200212 kubelet[2802]: E0114 00:47:42.200187 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.218553 kernel: kauditd_printk_skb: 25 callbacks suppressed Jan 14 00:47:42.218613 kernel: audit: type=1334 audit(1768351662.213:546): prog-id=163 op=LOAD Jan 14 00:47:42.213000 audit: BPF prog-id=163 op=LOAD Jan 14 00:47:42.220648 kernel: audit: type=1334 audit(1768351662.214:547): prog-id=164 op=LOAD Jan 14 00:47:42.214000 audit: BPF prog-id=164 op=LOAD Jan 14 00:47:42.214000 audit[3253]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=3241 pid=3253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.230037 kubelet[2802]: E0114 00:47:42.222552 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.230037 kubelet[2802]: W0114 00:47:42.222569 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.230037 kubelet[2802]: E0114 00:47:42.223152 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.230190 kernel: audit: type=1300 audit(1768351662.214:547): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=3241 pid=3253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.214000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531643938353465343432626466316436636664643031373965346339 Jan 14 00:47:42.239361 kubelet[2802]: E0114 00:47:42.231200 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.239361 kubelet[2802]: W0114 00:47:42.231215 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.239361 kubelet[2802]: E0114 00:47:42.231234 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.239361 kubelet[2802]: E0114 00:47:42.231472 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.239361 kubelet[2802]: W0114 00:47:42.231480 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.239361 kubelet[2802]: E0114 00:47:42.231491 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.239361 kubelet[2802]: E0114 00:47:42.233226 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.239361 kubelet[2802]: W0114 00:47:42.233236 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.239361 kubelet[2802]: E0114 00:47:42.233248 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.239361 kubelet[2802]: E0114 00:47:42.235665 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.239581 kubelet[2802]: W0114 00:47:42.235674 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.239581 kubelet[2802]: E0114 00:47:42.235687 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.239581 kubelet[2802]: E0114 00:47:42.237473 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.239581 kubelet[2802]: W0114 00:47:42.237483 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.239581 kubelet[2802]: E0114 00:47:42.237493 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.239581 kubelet[2802]: E0114 00:47:42.237831 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.239581 kubelet[2802]: W0114 00:47:42.238033 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.239581 kubelet[2802]: E0114 00:47:42.238052 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.239581 kubelet[2802]: E0114 00:47:42.238396 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.239581 kubelet[2802]: W0114 00:47:42.238405 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.239789 kubelet[2802]: E0114 00:47:42.238415 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.239789 kubelet[2802]: E0114 00:47:42.239270 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.239789 kubelet[2802]: W0114 00:47:42.239305 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.239789 kubelet[2802]: E0114 00:47:42.239316 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.239789 kubelet[2802]: E0114 00:47:42.239556 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.239789 kubelet[2802]: W0114 00:47:42.239564 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.239789 kubelet[2802]: E0114 00:47:42.239573 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.240130 kernel: audit: type=1327 audit(1768351662.214:547): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531643938353465343432626466316436636664643031373965346339 Jan 14 00:47:42.214000 audit: BPF prog-id=164 op=UNLOAD Jan 14 00:47:42.242320 kubelet[2802]: E0114 00:47:42.240439 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.242320 kubelet[2802]: W0114 00:47:42.240449 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.242320 kubelet[2802]: E0114 00:47:42.240458 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.242320 kubelet[2802]: E0114 00:47:42.241384 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.242320 kubelet[2802]: W0114 00:47:42.241394 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.242320 kubelet[2802]: E0114 00:47:42.241403 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.242320 kubelet[2802]: E0114 00:47:42.241662 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.242320 kubelet[2802]: W0114 00:47:42.241670 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.242320 kubelet[2802]: E0114 00:47:42.241678 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.242320 kubelet[2802]: E0114 00:47:42.241982 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.242614 kubelet[2802]: W0114 00:47:42.241991 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.242614 kubelet[2802]: E0114 00:47:42.241999 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.251967 kernel: audit: type=1334 audit(1768351662.214:548): prog-id=164 op=UNLOAD Jan 14 00:47:42.252044 kernel: audit: type=1300 audit(1768351662.214:548): arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3241 pid=3253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.214000 audit[3253]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3241 pid=3253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.252253 containerd[1611]: time="2026-01-14T00:47:42.246780616Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-28tb6,Uid:12a054c8-0aa7-4bbc-9278-b8219804f0ea,Namespace:calico-system,Attempt:0,}" Jan 14 00:47:42.252345 kubelet[2802]: E0114 00:47:42.243343 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.252345 kubelet[2802]: W0114 00:47:42.243354 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.252345 kubelet[2802]: E0114 00:47:42.243364 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.252345 kubelet[2802]: E0114 00:47:42.243576 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.252345 kubelet[2802]: W0114 00:47:42.243584 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.252345 kubelet[2802]: E0114 00:47:42.243591 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.252345 kubelet[2802]: E0114 00:47:42.243824 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.252345 kubelet[2802]: W0114 00:47:42.243831 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.252345 kubelet[2802]: E0114 00:47:42.243840 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.252345 kubelet[2802]: E0114 00:47:42.244017 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.252988 kubelet[2802]: W0114 00:47:42.244024 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.252988 kubelet[2802]: E0114 00:47:42.244032 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.252988 kubelet[2802]: E0114 00:47:42.244240 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.252988 kubelet[2802]: W0114 00:47:42.244248 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.252988 kubelet[2802]: E0114 00:47:42.244255 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.252988 kubelet[2802]: E0114 00:47:42.244693 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.252988 kubelet[2802]: W0114 00:47:42.244702 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.252988 kubelet[2802]: E0114 00:47:42.244711 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.252988 kubelet[2802]: E0114 00:47:42.245291 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:42.263192 kernel: audit: type=1327 audit(1768351662.214:548): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531643938353465343432626466316436636664643031373965346339 Jan 14 00:47:42.214000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531643938353465343432626466316436636664643031373965346339 Jan 14 00:47:42.263280 kubelet[2802]: E0114 00:47:42.260474 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.263280 kubelet[2802]: W0114 00:47:42.260487 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.263280 kubelet[2802]: E0114 00:47:42.260501 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.263280 kubelet[2802]: I0114 00:47:42.260528 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnt5f\" (UniqueName: \"kubernetes.io/projected/f41cc627-bd07-42f1-8e88-f85dcd2c063c-kube-api-access-pnt5f\") pod \"csi-node-driver-vzqkb\" (UID: \"f41cc627-bd07-42f1-8e88-f85dcd2c063c\") " pod="calico-system/csi-node-driver-vzqkb" Jan 14 00:47:42.263280 kubelet[2802]: E0114 00:47:42.260770 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.263280 kubelet[2802]: W0114 00:47:42.260780 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.263280 kubelet[2802]: E0114 00:47:42.260790 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.263280 kubelet[2802]: I0114 00:47:42.260818 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f41cc627-bd07-42f1-8e88-f85dcd2c063c-registration-dir\") pod \"csi-node-driver-vzqkb\" (UID: \"f41cc627-bd07-42f1-8e88-f85dcd2c063c\") " pod="calico-system/csi-node-driver-vzqkb" Jan 14 00:47:42.263280 kubelet[2802]: E0114 00:47:42.261558 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.263586 kubelet[2802]: W0114 00:47:42.261570 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.263586 kubelet[2802]: E0114 00:47:42.261579 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.263586 kubelet[2802]: I0114 00:47:42.261612 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f41cc627-bd07-42f1-8e88-f85dcd2c063c-socket-dir\") pod \"csi-node-driver-vzqkb\" (UID: \"f41cc627-bd07-42f1-8e88-f85dcd2c063c\") " pod="calico-system/csi-node-driver-vzqkb" Jan 14 00:47:42.263586 kubelet[2802]: E0114 00:47:42.262433 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.263586 kubelet[2802]: W0114 00:47:42.262445 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.263586 kubelet[2802]: E0114 00:47:42.262454 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.263586 kubelet[2802]: I0114 00:47:42.263062 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f41cc627-bd07-42f1-8e88-f85dcd2c063c-kubelet-dir\") pod \"csi-node-driver-vzqkb\" (UID: \"f41cc627-bd07-42f1-8e88-f85dcd2c063c\") " pod="calico-system/csi-node-driver-vzqkb" Jan 14 00:47:42.263586 kubelet[2802]: E0114 00:47:42.263428 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.263845 kubelet[2802]: W0114 00:47:42.263437 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.263845 kubelet[2802]: E0114 00:47:42.263446 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.263900 kubelet[2802]: E0114 00:47:42.263883 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.263900 kubelet[2802]: W0114 00:47:42.263898 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.263951 kubelet[2802]: E0114 00:47:42.263908 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.264758 kubelet[2802]: E0114 00:47:42.264712 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.264758 kubelet[2802]: W0114 00:47:42.264727 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.264758 kubelet[2802]: E0114 00:47:42.264737 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.267171 kernel: audit: type=1334 audit(1768351662.214:549): prog-id=165 op=LOAD Jan 14 00:47:42.214000 audit: BPF prog-id=165 op=LOAD Jan 14 00:47:42.267292 kubelet[2802]: E0114 00:47:42.264967 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.267292 kubelet[2802]: W0114 00:47:42.264975 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.267292 kubelet[2802]: E0114 00:47:42.264983 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.267292 kubelet[2802]: I0114 00:47:42.265067 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/f41cc627-bd07-42f1-8e88-f85dcd2c063c-varrun\") pod \"csi-node-driver-vzqkb\" (UID: \"f41cc627-bd07-42f1-8e88-f85dcd2c063c\") " pod="calico-system/csi-node-driver-vzqkb" Jan 14 00:47:42.267292 kubelet[2802]: E0114 00:47:42.265321 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.267292 kubelet[2802]: W0114 00:47:42.265328 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.267292 kubelet[2802]: E0114 00:47:42.265337 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.267292 kubelet[2802]: E0114 00:47:42.265882 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.267292 kubelet[2802]: W0114 00:47:42.265891 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.267580 kubelet[2802]: E0114 00:47:42.265900 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.267580 kubelet[2802]: E0114 00:47:42.266595 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.267580 kubelet[2802]: W0114 00:47:42.266604 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.267580 kubelet[2802]: E0114 00:47:42.266613 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.267580 kubelet[2802]: E0114 00:47:42.267020 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.267580 kubelet[2802]: W0114 00:47:42.267029 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.267580 kubelet[2802]: E0114 00:47:42.267037 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.277261 kernel: audit: type=1300 audit(1768351662.214:549): arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=3241 pid=3253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.214000 audit[3253]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=3241 pid=3253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.277359 kubelet[2802]: E0114 00:47:42.268314 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.277359 kubelet[2802]: W0114 00:47:42.268325 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.277359 kubelet[2802]: E0114 00:47:42.268334 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.277359 kubelet[2802]: E0114 00:47:42.268551 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.277359 kubelet[2802]: W0114 00:47:42.268558 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.277359 kubelet[2802]: E0114 00:47:42.268566 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.277359 kubelet[2802]: E0114 00:47:42.268806 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.277359 kubelet[2802]: W0114 00:47:42.268814 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.277359 kubelet[2802]: E0114 00:47:42.268822 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.214000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531643938353465343432626466316436636664643031373965346339 Jan 14 00:47:42.215000 audit: BPF prog-id=166 op=LOAD Jan 14 00:47:42.215000 audit[3253]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=3241 pid=3253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.215000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531643938353465343432626466316436636664643031373965346339 Jan 14 00:47:42.215000 audit: BPF prog-id=166 op=UNLOAD Jan 14 00:47:42.287185 kernel: audit: type=1327 audit(1768351662.214:549): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531643938353465343432626466316436636664643031373965346339 Jan 14 00:47:42.215000 audit[3253]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3241 pid=3253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.215000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531643938353465343432626466316436636664643031373965346339 Jan 14 00:47:42.215000 audit: BPF prog-id=165 op=UNLOAD Jan 14 00:47:42.215000 audit[3253]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3241 pid=3253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.215000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531643938353465343432626466316436636664643031373965346339 Jan 14 00:47:42.215000 audit: BPF prog-id=167 op=LOAD Jan 14 00:47:42.215000 audit[3253]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=3241 pid=3253 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.215000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531643938353465343432626466316436636664643031373965346339 Jan 14 00:47:42.220000 audit[3292]: NETFILTER_CFG table=filter:115 family=2 entries=21 op=nft_register_rule pid=3292 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:42.220000 audit[3292]: SYSCALL arch=c000003e syscall=46 success=yes exit=8224 a0=3 a1=7fff49204930 a2=0 a3=7fff4920491c items=0 ppid=2943 pid=3292 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.220000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:42.250000 audit[3292]: NETFILTER_CFG table=nat:116 family=2 entries=12 op=nft_register_rule pid=3292 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:42.250000 audit[3292]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7fff49204930 a2=0 a3=0 items=0 ppid=2943 pid=3292 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.250000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:42.325350 containerd[1611]: time="2026-01-14T00:47:42.325244953Z" level=info msg="connecting to shim d81288ef4b21c802a5269ff94a28696b0ee21fba084a353c0f40fafd58f9160a" address="unix:///run/containerd/s/73f4dba3b46309f40b046e6c8440ce78be70cdcdea47bcd2729be2f201321bd9" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:47:42.368701 kubelet[2802]: E0114 00:47:42.368661 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.368701 kubelet[2802]: W0114 00:47:42.368683 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.368701 kubelet[2802]: E0114 00:47:42.368700 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.369782 kubelet[2802]: E0114 00:47:42.369758 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.370079 kubelet[2802]: W0114 00:47:42.369784 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.370079 kubelet[2802]: E0114 00:47:42.369807 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.371750 kubelet[2802]: E0114 00:47:42.371649 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.372337 kubelet[2802]: W0114 00:47:42.371809 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.372337 kubelet[2802]: E0114 00:47:42.371841 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.372924 kubelet[2802]: E0114 00:47:42.372698 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.373023 kubelet[2802]: W0114 00:47:42.373001 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.373078 kubelet[2802]: E0114 00:47:42.373067 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.373431 kubelet[2802]: E0114 00:47:42.373399 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.373481 kubelet[2802]: W0114 00:47:42.373471 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.373548 kubelet[2802]: E0114 00:47:42.373517 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.374102 kubelet[2802]: E0114 00:47:42.373837 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.374102 kubelet[2802]: W0114 00:47:42.373847 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.374301 kubelet[2802]: E0114 00:47:42.373855 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.374493 kubelet[2802]: E0114 00:47:42.374454 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.374650 kubelet[2802]: W0114 00:47:42.374568 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.374650 kubelet[2802]: E0114 00:47:42.374581 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.374970 kubelet[2802]: E0114 00:47:42.374960 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.375103 kubelet[2802]: W0114 00:47:42.375000 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.375103 kubelet[2802]: E0114 00:47:42.375009 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.375496 kubelet[2802]: E0114 00:47:42.375391 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.375496 kubelet[2802]: W0114 00:47:42.375400 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.375496 kubelet[2802]: E0114 00:47:42.375408 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.376014 kubelet[2802]: E0114 00:47:42.375902 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.376014 kubelet[2802]: W0114 00:47:42.375913 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.376014 kubelet[2802]: E0114 00:47:42.375921 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.376310 kubelet[2802]: E0114 00:47:42.376290 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.376462 kubelet[2802]: W0114 00:47:42.376349 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.376462 kubelet[2802]: E0114 00:47:42.376360 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.377147 kubelet[2802]: E0114 00:47:42.377043 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.377147 kubelet[2802]: W0114 00:47:42.377054 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.377147 kubelet[2802]: E0114 00:47:42.377063 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.377494 kubelet[2802]: E0114 00:47:42.377404 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.377494 kubelet[2802]: W0114 00:47:42.377414 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.377494 kubelet[2802]: E0114 00:47:42.377423 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.377923 kubelet[2802]: E0114 00:47:42.377912 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.378225 kubelet[2802]: W0114 00:47:42.377973 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.378225 kubelet[2802]: E0114 00:47:42.377986 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.378463 kubelet[2802]: E0114 00:47:42.378444 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.378463 kubelet[2802]: W0114 00:47:42.378461 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.378517 kubelet[2802]: E0114 00:47:42.378475 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.378694 kubelet[2802]: E0114 00:47:42.378679 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.378694 kubelet[2802]: W0114 00:47:42.378692 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.379099 kubelet[2802]: E0114 00:47:42.378702 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.380279 kubelet[2802]: E0114 00:47:42.380251 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.380279 kubelet[2802]: W0114 00:47:42.380272 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.380342 kubelet[2802]: E0114 00:47:42.380283 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.380612 kubelet[2802]: E0114 00:47:42.380589 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.380612 kubelet[2802]: W0114 00:47:42.380604 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.380612 kubelet[2802]: E0114 00:47:42.380613 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.382132 kubelet[2802]: E0114 00:47:42.381083 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.382132 kubelet[2802]: W0114 00:47:42.381097 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.382132 kubelet[2802]: E0114 00:47:42.381147 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.382358 kubelet[2802]: E0114 00:47:42.382334 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.382402 kubelet[2802]: W0114 00:47:42.382357 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.382444 kubelet[2802]: E0114 00:47:42.382414 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.383028 kubelet[2802]: E0114 00:47:42.383005 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.383028 kubelet[2802]: W0114 00:47:42.383021 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.383095 kubelet[2802]: E0114 00:47:42.383030 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.384311 kubelet[2802]: E0114 00:47:42.384284 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.384367 kubelet[2802]: W0114 00:47:42.384320 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.384367 kubelet[2802]: E0114 00:47:42.384331 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.385473 kubelet[2802]: E0114 00:47:42.384618 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.385473 kubelet[2802]: W0114 00:47:42.384722 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.385473 kubelet[2802]: E0114 00:47:42.384740 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.385473 kubelet[2802]: E0114 00:47:42.385200 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.385473 kubelet[2802]: W0114 00:47:42.385221 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.385473 kubelet[2802]: E0114 00:47:42.385230 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.385623 kubelet[2802]: E0114 00:47:42.385614 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.385660 kubelet[2802]: W0114 00:47:42.385623 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.385686 kubelet[2802]: E0114 00:47:42.385660 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.398305 kubelet[2802]: E0114 00:47:42.398277 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:42.398305 kubelet[2802]: W0114 00:47:42.398296 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:42.398579 kubelet[2802]: E0114 00:47:42.398551 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:42.410079 containerd[1611]: time="2026-01-14T00:47:42.410042043Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-547ff64fb7-mfz8m,Uid:2fdc182d-8d82-4e17-a664-8eb8af914c6b,Namespace:calico-system,Attempt:0,} returns sandbox id \"e1d9854e442bdf1d6cfdd0179e4c957b581bf46954c2f57de639ce9db9fdcfcc\"" Jan 14 00:47:42.411936 kubelet[2802]: E0114 00:47:42.411917 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:42.419354 containerd[1611]: time="2026-01-14T00:47:42.419311015Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\"" Jan 14 00:47:42.420281 systemd[1]: Started cri-containerd-d81288ef4b21c802a5269ff94a28696b0ee21fba084a353c0f40fafd58f9160a.scope - libcontainer container d81288ef4b21c802a5269ff94a28696b0ee21fba084a353c0f40fafd58f9160a. Jan 14 00:47:42.441000 audit: BPF prog-id=168 op=LOAD Jan 14 00:47:42.442000 audit: BPF prog-id=169 op=LOAD Jan 14 00:47:42.442000 audit[3348]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=3336 pid=3348 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.442000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6438313238386566346232316338303261353236396666393461323836 Jan 14 00:47:42.442000 audit: BPF prog-id=169 op=UNLOAD Jan 14 00:47:42.442000 audit[3348]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3336 pid=3348 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.442000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6438313238386566346232316338303261353236396666393461323836 Jan 14 00:47:42.442000 audit: BPF prog-id=170 op=LOAD Jan 14 00:47:42.442000 audit[3348]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3336 pid=3348 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.442000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6438313238386566346232316338303261353236396666393461323836 Jan 14 00:47:42.442000 audit: BPF prog-id=171 op=LOAD Jan 14 00:47:42.442000 audit[3348]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3336 pid=3348 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.442000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6438313238386566346232316338303261353236396666393461323836 Jan 14 00:47:42.442000 audit: BPF prog-id=171 op=UNLOAD Jan 14 00:47:42.442000 audit[3348]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3336 pid=3348 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.442000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6438313238386566346232316338303261353236396666393461323836 Jan 14 00:47:42.442000 audit: BPF prog-id=170 op=UNLOAD Jan 14 00:47:42.442000 audit[3348]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3336 pid=3348 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.442000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6438313238386566346232316338303261353236396666393461323836 Jan 14 00:47:42.442000 audit: BPF prog-id=172 op=LOAD Jan 14 00:47:42.442000 audit[3348]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=3336 pid=3348 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:42.442000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6438313238386566346232316338303261353236396666393461323836 Jan 14 00:47:42.471483 containerd[1611]: time="2026-01-14T00:47:42.471433266Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-28tb6,Uid:12a054c8-0aa7-4bbc-9278-b8219804f0ea,Namespace:calico-system,Attempt:0,} returns sandbox id \"d81288ef4b21c802a5269ff94a28696b0ee21fba084a353c0f40fafd58f9160a\"" Jan 14 00:47:42.472580 kubelet[2802]: E0114 00:47:42.472564 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:43.242836 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount410731492.mount: Deactivated successfully. Jan 14 00:47:43.499817 kubelet[2802]: E0114 00:47:43.499678 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:47:43.848370 containerd[1611]: time="2026-01-14T00:47:43.847812176Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:43.849446 containerd[1611]: time="2026-01-14T00:47:43.849075776Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.4: active requests=0, bytes read=0" Jan 14 00:47:43.850156 containerd[1611]: time="2026-01-14T00:47:43.850121561Z" level=info msg="ImageCreate event name:\"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:43.852387 containerd[1611]: time="2026-01-14T00:47:43.852357609Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:43.853895 containerd[1611]: time="2026-01-14T00:47:43.853860216Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.4\" with image id \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:6f437220b5b3c627fb4a0fc8dc323363101f3c22a8f337612c2a1ddfb73b810c\", size \"35234482\" in 1.434386008s" Jan 14 00:47:43.853895 containerd[1611]: time="2026-01-14T00:47:43.853894399Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.4\" returns image reference \"sha256:aa1490366a77160b4cc8f9af82281ab7201ffda0882871f860e1eb1c4f825958\"" Jan 14 00:47:43.859136 containerd[1611]: time="2026-01-14T00:47:43.857159580Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\"" Jan 14 00:47:43.877296 containerd[1611]: time="2026-01-14T00:47:43.877268250Z" level=info msg="CreateContainer within sandbox \"e1d9854e442bdf1d6cfdd0179e4c957b581bf46954c2f57de639ce9db9fdcfcc\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Jan 14 00:47:43.882683 containerd[1611]: time="2026-01-14T00:47:43.882649553Z" level=info msg="Container 2ec0eac1d3cf798e8887ec4aaff159253ff2f2d2d86aa836ebc2e577be70e84d: CDI devices from CRI Config.CDIDevices: []" Jan 14 00:47:43.895166 containerd[1611]: time="2026-01-14T00:47:43.895039314Z" level=info msg="CreateContainer within sandbox \"e1d9854e442bdf1d6cfdd0179e4c957b581bf46954c2f57de639ce9db9fdcfcc\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"2ec0eac1d3cf798e8887ec4aaff159253ff2f2d2d86aa836ebc2e577be70e84d\"" Jan 14 00:47:43.896474 containerd[1611]: time="2026-01-14T00:47:43.896440043Z" level=info msg="StartContainer for \"2ec0eac1d3cf798e8887ec4aaff159253ff2f2d2d86aa836ebc2e577be70e84d\"" Jan 14 00:47:43.897812 containerd[1611]: time="2026-01-14T00:47:43.897778849Z" level=info msg="connecting to shim 2ec0eac1d3cf798e8887ec4aaff159253ff2f2d2d86aa836ebc2e577be70e84d" address="unix:///run/containerd/s/731dca5e80e11a07c8d8f546eb19413414689c41052fec85265c8aa7c42a9ae9" protocol=ttrpc version=3 Jan 14 00:47:43.927250 systemd[1]: Started cri-containerd-2ec0eac1d3cf798e8887ec4aaff159253ff2f2d2d86aa836ebc2e577be70e84d.scope - libcontainer container 2ec0eac1d3cf798e8887ec4aaff159253ff2f2d2d86aa836ebc2e577be70e84d. Jan 14 00:47:43.941000 audit: BPF prog-id=173 op=LOAD Jan 14 00:47:43.942000 audit: BPF prog-id=174 op=LOAD Jan 14 00:47:43.942000 audit[3415]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=3241 pid=3415 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:43.942000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3265633065616331643363663739386538383837656334616166663135 Jan 14 00:47:43.942000 audit: BPF prog-id=174 op=UNLOAD Jan 14 00:47:43.942000 audit[3415]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3241 pid=3415 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:43.942000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3265633065616331643363663739386538383837656334616166663135 Jan 14 00:47:43.942000 audit: BPF prog-id=175 op=LOAD Jan 14 00:47:43.942000 audit[3415]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3241 pid=3415 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:43.942000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3265633065616331643363663739386538383837656334616166663135 Jan 14 00:47:43.942000 audit: BPF prog-id=176 op=LOAD Jan 14 00:47:43.942000 audit[3415]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3241 pid=3415 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:43.942000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3265633065616331643363663739386538383837656334616166663135 Jan 14 00:47:43.943000 audit: BPF prog-id=176 op=UNLOAD Jan 14 00:47:43.943000 audit[3415]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=3241 pid=3415 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:43.943000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3265633065616331643363663739386538383837656334616166663135 Jan 14 00:47:43.943000 audit: BPF prog-id=175 op=UNLOAD Jan 14 00:47:43.943000 audit[3415]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=3241 pid=3415 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:43.943000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3265633065616331643363663739386538383837656334616166663135 Jan 14 00:47:43.943000 audit: BPF prog-id=177 op=LOAD Jan 14 00:47:43.943000 audit[3415]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=3241 pid=3415 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:43.943000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3265633065616331643363663739386538383837656334616166663135 Jan 14 00:47:43.991492 containerd[1611]: time="2026-01-14T00:47:43.991429726Z" level=info msg="StartContainer for \"2ec0eac1d3cf798e8887ec4aaff159253ff2f2d2d86aa836ebc2e577be70e84d\" returns successfully" Jan 14 00:47:44.574132 containerd[1611]: time="2026-01-14T00:47:44.573773819Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:44.575149 containerd[1611]: time="2026-01-14T00:47:44.575127501Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4: active requests=0, bytes read=0" Jan 14 00:47:44.576240 containerd[1611]: time="2026-01-14T00:47:44.576206764Z" level=info msg="ImageCreate event name:\"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:44.581227 containerd[1611]: time="2026-01-14T00:47:44.580577590Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:44.581581 containerd[1611]: time="2026-01-14T00:47:44.581248995Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" with image id \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:50bdfe370b7308fa9957ed1eaccd094aa4f27f9a4f1dfcfef2f8a7696a1551e1\", size \"5941314\" in 724.056553ms" Jan 14 00:47:44.581679 containerd[1611]: time="2026-01-14T00:47:44.581662594Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.4\" returns image reference \"sha256:570719e9c34097019014ae2ad94edf4e523bc6892e77fb1c64c23e5b7f390fe5\"" Jan 14 00:47:44.585674 containerd[1611]: time="2026-01-14T00:47:44.585652514Z" level=info msg="CreateContainer within sandbox \"d81288ef4b21c802a5269ff94a28696b0ee21fba084a353c0f40fafd58f9160a\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Jan 14 00:47:44.603627 containerd[1611]: time="2026-01-14T00:47:44.603563958Z" level=info msg="Container 0d91347282bad496e6715eb4af40e1b99bea5f4fd5c9980b67d06d1979ce6971: CDI devices from CRI Config.CDIDevices: []" Jan 14 00:47:44.607883 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3689403655.mount: Deactivated successfully. Jan 14 00:47:44.610664 kubelet[2802]: E0114 00:47:44.610636 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:44.618743 containerd[1611]: time="2026-01-14T00:47:44.618695073Z" level=info msg="CreateContainer within sandbox \"d81288ef4b21c802a5269ff94a28696b0ee21fba084a353c0f40fafd58f9160a\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"0d91347282bad496e6715eb4af40e1b99bea5f4fd5c9980b67d06d1979ce6971\"" Jan 14 00:47:44.621797 containerd[1611]: time="2026-01-14T00:47:44.621285748Z" level=info msg="StartContainer for \"0d91347282bad496e6715eb4af40e1b99bea5f4fd5c9980b67d06d1979ce6971\"" Jan 14 00:47:44.623278 containerd[1611]: time="2026-01-14T00:47:44.623235730Z" level=info msg="connecting to shim 0d91347282bad496e6715eb4af40e1b99bea5f4fd5c9980b67d06d1979ce6971" address="unix:///run/containerd/s/73f4dba3b46309f40b046e6c8440ce78be70cdcdea47bcd2729be2f201321bd9" protocol=ttrpc version=3 Jan 14 00:47:44.636523 kubelet[2802]: I0114 00:47:44.636473 2802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-547ff64fb7-mfz8m" podStartSLOduration=2.196137312 podStartE2EDuration="3.636458657s" podCreationTimestamp="2026-01-14 00:47:41 +0000 UTC" firstStartedPulling="2026-01-14 00:47:42.415386812 +0000 UTC m=+21.010483226" lastFinishedPulling="2026-01-14 00:47:43.855708157 +0000 UTC m=+22.450804571" observedRunningTime="2026-01-14 00:47:44.635158839 +0000 UTC m=+23.230255253" watchObservedRunningTime="2026-01-14 00:47:44.636458657 +0000 UTC m=+23.231555071" Jan 14 00:47:44.662472 kubelet[2802]: E0114 00:47:44.662440 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.662472 kubelet[2802]: W0114 00:47:44.662465 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.662618 kubelet[2802]: E0114 00:47:44.662594 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.663529 kubelet[2802]: E0114 00:47:44.663499 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.663852 kubelet[2802]: W0114 00:47:44.663768 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.663852 kubelet[2802]: E0114 00:47:44.663794 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.666471 kubelet[2802]: E0114 00:47:44.666440 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.666471 kubelet[2802]: W0114 00:47:44.666453 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.666471 kubelet[2802]: E0114 00:47:44.666465 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.669229 kubelet[2802]: E0114 00:47:44.669213 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.669229 kubelet[2802]: W0114 00:47:44.669229 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.669317 kubelet[2802]: E0114 00:47:44.669240 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.669543 kubelet[2802]: E0114 00:47:44.669533 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.669543 kubelet[2802]: W0114 00:47:44.669543 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.669605 kubelet[2802]: E0114 00:47:44.669552 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.671350 kubelet[2802]: E0114 00:47:44.671322 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.671350 kubelet[2802]: W0114 00:47:44.671337 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.671350 kubelet[2802]: E0114 00:47:44.671349 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.671587 kubelet[2802]: E0114 00:47:44.671573 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.671587 kubelet[2802]: W0114 00:47:44.671585 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.671757 kubelet[2802]: E0114 00:47:44.671593 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.672136 kubelet[2802]: E0114 00:47:44.672055 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.672136 kubelet[2802]: W0114 00:47:44.672066 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.672136 kubelet[2802]: E0114 00:47:44.672074 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.673193 kubelet[2802]: E0114 00:47:44.673175 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.673193 kubelet[2802]: W0114 00:47:44.673188 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.673547 kubelet[2802]: E0114 00:47:44.673197 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.674151 systemd[1]: Started cri-containerd-0d91347282bad496e6715eb4af40e1b99bea5f4fd5c9980b67d06d1979ce6971.scope - libcontainer container 0d91347282bad496e6715eb4af40e1b99bea5f4fd5c9980b67d06d1979ce6971. Jan 14 00:47:44.676300 kubelet[2802]: E0114 00:47:44.676280 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.676342 kubelet[2802]: W0114 00:47:44.676294 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.676342 kubelet[2802]: E0114 00:47:44.676315 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.676605 kubelet[2802]: E0114 00:47:44.676503 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.676605 kubelet[2802]: W0114 00:47:44.676515 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.676605 kubelet[2802]: E0114 00:47:44.676523 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.677474 kubelet[2802]: E0114 00:47:44.676945 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.677474 kubelet[2802]: W0114 00:47:44.676958 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.677474 kubelet[2802]: E0114 00:47:44.676968 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.677824 kubelet[2802]: E0114 00:47:44.677791 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.677824 kubelet[2802]: W0114 00:47:44.677802 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.678274 kubelet[2802]: E0114 00:47:44.678261 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.678620 kubelet[2802]: E0114 00:47:44.678609 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.678701 kubelet[2802]: W0114 00:47:44.678691 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.678842 kubelet[2802]: E0114 00:47:44.678753 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.679373 kubelet[2802]: E0114 00:47:44.679309 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.679373 kubelet[2802]: W0114 00:47:44.679340 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.679607 kubelet[2802]: E0114 00:47:44.679350 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.693009 kubelet[2802]: E0114 00:47:44.692934 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.693009 kubelet[2802]: W0114 00:47:44.692948 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.693009 kubelet[2802]: E0114 00:47:44.692960 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.693557 kubelet[2802]: E0114 00:47:44.693530 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.693643 kubelet[2802]: W0114 00:47:44.693540 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.693643 kubelet[2802]: E0114 00:47:44.693625 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.694440 kubelet[2802]: E0114 00:47:44.694429 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.694629 kubelet[2802]: W0114 00:47:44.694473 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.694629 kubelet[2802]: E0114 00:47:44.694483 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.695152 kubelet[2802]: E0114 00:47:44.695120 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.695152 kubelet[2802]: W0114 00:47:44.695132 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.695152 kubelet[2802]: E0114 00:47:44.695141 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.695547 kubelet[2802]: E0114 00:47:44.695506 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.695547 kubelet[2802]: W0114 00:47:44.695515 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.695547 kubelet[2802]: E0114 00:47:44.695523 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.695900 kubelet[2802]: E0114 00:47:44.695889 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.695976 kubelet[2802]: W0114 00:47:44.695953 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.695976 kubelet[2802]: E0114 00:47:44.695965 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.696431 kubelet[2802]: E0114 00:47:44.696420 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.696542 kubelet[2802]: W0114 00:47:44.696503 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.696621 kubelet[2802]: E0114 00:47:44.696609 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.698251 kubelet[2802]: E0114 00:47:44.698239 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.698327 kubelet[2802]: W0114 00:47:44.698317 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.698392 kubelet[2802]: E0114 00:47:44.698364 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.698976 kubelet[2802]: E0114 00:47:44.698944 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.698976 kubelet[2802]: W0114 00:47:44.698955 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.698976 kubelet[2802]: E0114 00:47:44.698964 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.699728 kubelet[2802]: E0114 00:47:44.699705 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.699848 kubelet[2802]: W0114 00:47:44.699753 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.699848 kubelet[2802]: E0114 00:47:44.699763 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.700212 kubelet[2802]: E0114 00:47:44.700177 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.700324 kubelet[2802]: W0114 00:47:44.700288 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.700324 kubelet[2802]: E0114 00:47:44.700301 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.700788 kubelet[2802]: E0114 00:47:44.700749 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.700788 kubelet[2802]: W0114 00:47:44.700762 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.700788 kubelet[2802]: E0114 00:47:44.700774 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.701234 kubelet[2802]: E0114 00:47:44.701200 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.701234 kubelet[2802]: W0114 00:47:44.701211 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.701234 kubelet[2802]: E0114 00:47:44.701219 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.701744 kubelet[2802]: E0114 00:47:44.701714 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.701859 kubelet[2802]: W0114 00:47:44.701821 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.701859 kubelet[2802]: E0114 00:47:44.701838 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.702478 kubelet[2802]: E0114 00:47:44.702444 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.702478 kubelet[2802]: W0114 00:47:44.702456 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.702478 kubelet[2802]: E0114 00:47:44.702466 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.703321 kubelet[2802]: E0114 00:47:44.703290 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.703321 kubelet[2802]: W0114 00:47:44.703301 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.703321 kubelet[2802]: E0114 00:47:44.703310 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.703829 kubelet[2802]: E0114 00:47:44.703652 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.703829 kubelet[2802]: W0114 00:47:44.703662 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.703829 kubelet[2802]: E0114 00:47:44.703670 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.704307 kubelet[2802]: E0114 00:47:44.704296 2802 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 14 00:47:44.704391 kubelet[2802]: W0114 00:47:44.704380 2802 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 14 00:47:44.704450 kubelet[2802]: E0114 00:47:44.704430 2802 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 14 00:47:44.757000 audit: BPF prog-id=178 op=LOAD Jan 14 00:47:44.757000 audit[3453]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=3336 pid=3453 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:44.757000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3064393133343732383262616434393665363731356562346166343065 Jan 14 00:47:44.757000 audit: BPF prog-id=179 op=LOAD Jan 14 00:47:44.757000 audit[3453]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=3336 pid=3453 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:44.757000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3064393133343732383262616434393665363731356562346166343065 Jan 14 00:47:44.757000 audit: BPF prog-id=179 op=UNLOAD Jan 14 00:47:44.757000 audit[3453]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3336 pid=3453 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:44.757000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3064393133343732383262616434393665363731356562346166343065 Jan 14 00:47:44.757000 audit: BPF prog-id=178 op=UNLOAD Jan 14 00:47:44.757000 audit[3453]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3336 pid=3453 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:44.757000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3064393133343732383262616434393665363731356562346166343065 Jan 14 00:47:44.757000 audit: BPF prog-id=180 op=LOAD Jan 14 00:47:44.757000 audit[3453]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=3336 pid=3453 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:44.757000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3064393133343732383262616434393665363731356562346166343065 Jan 14 00:47:44.789649 containerd[1611]: time="2026-01-14T00:47:44.789600494Z" level=info msg="StartContainer for \"0d91347282bad496e6715eb4af40e1b99bea5f4fd5c9980b67d06d1979ce6971\" returns successfully" Jan 14 00:47:44.809483 systemd[1]: cri-containerd-0d91347282bad496e6715eb4af40e1b99bea5f4fd5c9980b67d06d1979ce6971.scope: Deactivated successfully. Jan 14 00:47:44.812244 containerd[1611]: time="2026-01-14T00:47:44.812213886Z" level=info msg="received container exit event container_id:\"0d91347282bad496e6715eb4af40e1b99bea5f4fd5c9980b67d06d1979ce6971\" id:\"0d91347282bad496e6715eb4af40e1b99bea5f4fd5c9980b67d06d1979ce6971\" pid:3484 exited_at:{seconds:1768351664 nanos:811754965}" Jan 14 00:47:44.812000 audit: BPF prog-id=180 op=UNLOAD Jan 14 00:47:44.844783 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0d91347282bad496e6715eb4af40e1b99bea5f4fd5c9980b67d06d1979ce6971-rootfs.mount: Deactivated successfully. Jan 14 00:47:45.499302 kubelet[2802]: E0114 00:47:45.499245 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:47:45.617611 kubelet[2802]: I0114 00:47:45.617431 2802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 14 00:47:45.620402 kubelet[2802]: E0114 00:47:45.617756 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:45.620402 kubelet[2802]: E0114 00:47:45.619044 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:45.622904 containerd[1611]: time="2026-01-14T00:47:45.622857853Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\"" Jan 14 00:47:47.408469 containerd[1611]: time="2026-01-14T00:47:47.408408695Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:47.409688 containerd[1611]: time="2026-01-14T00:47:47.409513060Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.4: active requests=0, bytes read=70442291" Jan 14 00:47:47.410433 containerd[1611]: time="2026-01-14T00:47:47.410230423Z" level=info msg="ImageCreate event name:\"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:47.412536 containerd[1611]: time="2026-01-14T00:47:47.412507527Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:47.413051 containerd[1611]: time="2026-01-14T00:47:47.413003067Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.4\" with image id \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:273501a9cfbd848ade2b6a8452dfafdd3adb4f9bf9aec45c398a5d19b8026627\", size \"71941459\" in 1.790107711s" Jan 14 00:47:47.413051 containerd[1611]: time="2026-01-14T00:47:47.413036619Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.4\" returns image reference \"sha256:24e1e7377c738d4080eb462a29e2c6756d383d8d25ad87b7f49165581f20c3cd\"" Jan 14 00:47:47.416740 containerd[1611]: time="2026-01-14T00:47:47.416690064Z" level=info msg="CreateContainer within sandbox \"d81288ef4b21c802a5269ff94a28696b0ee21fba084a353c0f40fafd58f9160a\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jan 14 00:47:47.426603 containerd[1611]: time="2026-01-14T00:47:47.426581637Z" level=info msg="Container 216a1e4dba246f31d04a56f52eb05deae624e38c5fbbf1ff80164982afc7d7c5: CDI devices from CRI Config.CDIDevices: []" Jan 14 00:47:47.447097 containerd[1611]: time="2026-01-14T00:47:47.447024082Z" level=info msg="CreateContainer within sandbox \"d81288ef4b21c802a5269ff94a28696b0ee21fba084a353c0f40fafd58f9160a\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"216a1e4dba246f31d04a56f52eb05deae624e38c5fbbf1ff80164982afc7d7c5\"" Jan 14 00:47:47.447973 containerd[1611]: time="2026-01-14T00:47:47.447780767Z" level=info msg="StartContainer for \"216a1e4dba246f31d04a56f52eb05deae624e38c5fbbf1ff80164982afc7d7c5\"" Jan 14 00:47:47.449543 containerd[1611]: time="2026-01-14T00:47:47.449478977Z" level=info msg="connecting to shim 216a1e4dba246f31d04a56f52eb05deae624e38c5fbbf1ff80164982afc7d7c5" address="unix:///run/containerd/s/73f4dba3b46309f40b046e6c8440ce78be70cdcdea47bcd2729be2f201321bd9" protocol=ttrpc version=3 Jan 14 00:47:47.485428 systemd[1]: Started cri-containerd-216a1e4dba246f31d04a56f52eb05deae624e38c5fbbf1ff80164982afc7d7c5.scope - libcontainer container 216a1e4dba246f31d04a56f52eb05deae624e38c5fbbf1ff80164982afc7d7c5. Jan 14 00:47:47.500066 kubelet[2802]: E0114 00:47:47.499778 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:47:47.542000 audit: BPF prog-id=181 op=LOAD Jan 14 00:47:47.547568 kernel: kauditd_printk_skb: 78 callbacks suppressed Jan 14 00:47:47.547694 kernel: audit: type=1334 audit(1768351667.542:578): prog-id=181 op=LOAD Jan 14 00:47:47.542000 audit[3551]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3336 pid=3551 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:47.556159 kernel: audit: type=1300 audit(1768351667.542:578): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=3336 pid=3551 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:47.542000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231366131653464626132343666333164303461353666353265623035 Jan 14 00:47:47.565325 kernel: audit: type=1327 audit(1768351667.542:578): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231366131653464626132343666333164303461353666353265623035 Jan 14 00:47:47.565802 kernel: audit: type=1334 audit(1768351667.543:579): prog-id=182 op=LOAD Jan 14 00:47:47.543000 audit: BPF prog-id=182 op=LOAD Jan 14 00:47:47.543000 audit[3551]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3336 pid=3551 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:47.543000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231366131653464626132343666333164303461353666353265623035 Jan 14 00:47:47.578555 kernel: audit: type=1300 audit(1768351667.543:579): arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=3336 pid=3551 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:47.578671 kernel: audit: type=1327 audit(1768351667.543:579): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231366131653464626132343666333164303461353666353265623035 Jan 14 00:47:47.587001 kernel: audit: type=1334 audit(1768351667.543:580): prog-id=182 op=UNLOAD Jan 14 00:47:47.543000 audit: BPF prog-id=182 op=UNLOAD Jan 14 00:47:47.594683 kernel: audit: type=1300 audit(1768351667.543:580): arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3336 pid=3551 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:47.543000 audit[3551]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3336 pid=3551 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:47.602775 kernel: audit: type=1327 audit(1768351667.543:580): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231366131653464626132343666333164303461353666353265623035 Jan 14 00:47:47.543000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231366131653464626132343666333164303461353666353265623035 Jan 14 00:47:47.543000 audit: BPF prog-id=181 op=UNLOAD Jan 14 00:47:47.609131 kernel: audit: type=1334 audit(1768351667.543:581): prog-id=181 op=UNLOAD Jan 14 00:47:47.543000 audit[3551]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3336 pid=3551 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:47.543000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231366131653464626132343666333164303461353666353265623035 Jan 14 00:47:47.543000 audit: BPF prog-id=183 op=LOAD Jan 14 00:47:47.543000 audit[3551]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=3336 pid=3551 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:47.543000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3231366131653464626132343666333164303461353666353265623035 Jan 14 00:47:47.612523 containerd[1611]: time="2026-01-14T00:47:47.612484466Z" level=info msg="StartContainer for \"216a1e4dba246f31d04a56f52eb05deae624e38c5fbbf1ff80164982afc7d7c5\" returns successfully" Jan 14 00:47:47.632796 kubelet[2802]: E0114 00:47:47.632738 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:48.197360 containerd[1611]: time="2026-01-14T00:47:48.197158752Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 14 00:47:48.200877 systemd[1]: cri-containerd-216a1e4dba246f31d04a56f52eb05deae624e38c5fbbf1ff80164982afc7d7c5.scope: Deactivated successfully. Jan 14 00:47:48.201639 systemd[1]: cri-containerd-216a1e4dba246f31d04a56f52eb05deae624e38c5fbbf1ff80164982afc7d7c5.scope: Consumed 584ms CPU time, 197.1M memory peak, 171.3M written to disk. Jan 14 00:47:48.203000 audit: BPF prog-id=183 op=UNLOAD Jan 14 00:47:48.204780 containerd[1611]: time="2026-01-14T00:47:48.204721318Z" level=info msg="received container exit event container_id:\"216a1e4dba246f31d04a56f52eb05deae624e38c5fbbf1ff80164982afc7d7c5\" id:\"216a1e4dba246f31d04a56f52eb05deae624e38c5fbbf1ff80164982afc7d7c5\" pid:3563 exited_at:{seconds:1768351668 nanos:204382619}" Jan 14 00:47:48.218748 kubelet[2802]: I0114 00:47:48.218633 2802 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Jan 14 00:47:48.246812 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-216a1e4dba246f31d04a56f52eb05deae624e38c5fbbf1ff80164982afc7d7c5-rootfs.mount: Deactivated successfully. Jan 14 00:47:48.309048 systemd[1]: Created slice kubepods-burstable-pode4d15318_1381_4cb4_b138_c44ed14a3a71.slice - libcontainer container kubepods-burstable-pode4d15318_1381_4cb4_b138_c44ed14a3a71.slice. Jan 14 00:47:48.320140 kubelet[2802]: I0114 00:47:48.320087 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c12ac0b5-659e-4a97-821e-20b4aabfc024-tigera-ca-bundle\") pod \"calico-kube-controllers-5f75bc94c6-2nthl\" (UID: \"c12ac0b5-659e-4a97-821e-20b4aabfc024\") " pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" Jan 14 00:47:48.320395 kubelet[2802]: I0114 00:47:48.320326 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/43d5cf61-6f70-467e-abb2-9b166aef5949-config-volume\") pod \"coredns-674b8bbfcf-hc5jt\" (UID: \"43d5cf61-6f70-467e-abb2-9b166aef5949\") " pod="kube-system/coredns-674b8bbfcf-hc5jt" Jan 14 00:47:48.320395 kubelet[2802]: I0114 00:47:48.320355 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c84p\" (UniqueName: \"kubernetes.io/projected/e4d15318-1381-4cb4-b138-c44ed14a3a71-kube-api-access-8c84p\") pod \"coredns-674b8bbfcf-v8f4p\" (UID: \"e4d15318-1381-4cb4-b138-c44ed14a3a71\") " pod="kube-system/coredns-674b8bbfcf-v8f4p" Jan 14 00:47:48.320553 kubelet[2802]: I0114 00:47:48.320487 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfqdv\" (UniqueName: \"kubernetes.io/projected/c12ac0b5-659e-4a97-821e-20b4aabfc024-kube-api-access-sfqdv\") pod \"calico-kube-controllers-5f75bc94c6-2nthl\" (UID: \"c12ac0b5-659e-4a97-821e-20b4aabfc024\") " pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" Jan 14 00:47:48.320553 kubelet[2802]: I0114 00:47:48.320509 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e4d15318-1381-4cb4-b138-c44ed14a3a71-config-volume\") pod \"coredns-674b8bbfcf-v8f4p\" (UID: \"e4d15318-1381-4cb4-b138-c44ed14a3a71\") " pod="kube-system/coredns-674b8bbfcf-v8f4p" Jan 14 00:47:48.320553 kubelet[2802]: I0114 00:47:48.320526 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6nvx\" (UniqueName: \"kubernetes.io/projected/43d5cf61-6f70-467e-abb2-9b166aef5949-kube-api-access-x6nvx\") pod \"coredns-674b8bbfcf-hc5jt\" (UID: \"43d5cf61-6f70-467e-abb2-9b166aef5949\") " pod="kube-system/coredns-674b8bbfcf-hc5jt" Jan 14 00:47:48.326691 systemd[1]: Created slice kubepods-besteffort-podc12ac0b5_659e_4a97_821e_20b4aabfc024.slice - libcontainer container kubepods-besteffort-podc12ac0b5_659e_4a97_821e_20b4aabfc024.slice. Jan 14 00:47:48.343124 systemd[1]: Created slice kubepods-besteffort-poda556fce5_e24b_4167_85b9_cd5dc0a668fe.slice - libcontainer container kubepods-besteffort-poda556fce5_e24b_4167_85b9_cd5dc0a668fe.slice. Jan 14 00:47:48.361176 systemd[1]: Created slice kubepods-besteffort-poddcaab5f4_fd68_41e8_986b_14fa1a168bce.slice - libcontainer container kubepods-besteffort-poddcaab5f4_fd68_41e8_986b_14fa1a168bce.slice. Jan 14 00:47:48.373188 systemd[1]: Created slice kubepods-burstable-pod43d5cf61_6f70_467e_abb2_9b166aef5949.slice - libcontainer container kubepods-burstable-pod43d5cf61_6f70_467e_abb2_9b166aef5949.slice. Jan 14 00:47:48.382728 systemd[1]: Created slice kubepods-besteffort-pod976c1fb0_d44d_49e6_bc13_822966e59142.slice - libcontainer container kubepods-besteffort-pod976c1fb0_d44d_49e6_bc13_822966e59142.slice. Jan 14 00:47:48.392003 systemd[1]: Created slice kubepods-besteffort-pod32668954_0a97_46fe_9f01_fa91a530c2e3.slice - libcontainer container kubepods-besteffort-pod32668954_0a97_46fe_9f01_fa91a530c2e3.slice. Jan 14 00:47:48.421533 kubelet[2802]: I0114 00:47:48.421420 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/a556fce5-e24b-4167-85b9-cd5dc0a668fe-whisker-backend-key-pair\") pod \"whisker-844d8b6846-jq4cl\" (UID: \"a556fce5-e24b-4167-85b9-cd5dc0a668fe\") " pod="calico-system/whisker-844d8b6846-jq4cl" Jan 14 00:47:48.421533 kubelet[2802]: I0114 00:47:48.421480 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmr4q\" (UniqueName: \"kubernetes.io/projected/dcaab5f4-fd68-41e8-986b-14fa1a168bce-kube-api-access-hmr4q\") pod \"calico-apiserver-5f9b965548-bgkj8\" (UID: \"dcaab5f4-fd68-41e8-986b-14fa1a168bce\") " pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" Jan 14 00:47:48.422830 kubelet[2802]: I0114 00:47:48.421719 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkk9t\" (UniqueName: \"kubernetes.io/projected/32668954-0a97-46fe-9f01-fa91a530c2e3-kube-api-access-zkk9t\") pod \"calico-apiserver-5f9b965548-xv9gd\" (UID: \"32668954-0a97-46fe-9f01-fa91a530c2e3\") " pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" Jan 14 00:47:48.422830 kubelet[2802]: I0114 00:47:48.421956 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a556fce5-e24b-4167-85b9-cd5dc0a668fe-whisker-ca-bundle\") pod \"whisker-844d8b6846-jq4cl\" (UID: \"a556fce5-e24b-4167-85b9-cd5dc0a668fe\") " pod="calico-system/whisker-844d8b6846-jq4cl" Jan 14 00:47:48.422830 kubelet[2802]: I0114 00:47:48.421978 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vwgj\" (UniqueName: \"kubernetes.io/projected/976c1fb0-d44d-49e6-bc13-822966e59142-kube-api-access-5vwgj\") pod \"goldmane-666569f655-p6wlv\" (UID: \"976c1fb0-d44d-49e6-bc13-822966e59142\") " pod="calico-system/goldmane-666569f655-p6wlv" Jan 14 00:47:48.422830 kubelet[2802]: I0114 00:47:48.421996 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/32668954-0a97-46fe-9f01-fa91a530c2e3-calico-apiserver-certs\") pod \"calico-apiserver-5f9b965548-xv9gd\" (UID: \"32668954-0a97-46fe-9f01-fa91a530c2e3\") " pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" Jan 14 00:47:48.422830 kubelet[2802]: I0114 00:47:48.422024 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl6mw\" (UniqueName: \"kubernetes.io/projected/a556fce5-e24b-4167-85b9-cd5dc0a668fe-kube-api-access-pl6mw\") pod \"whisker-844d8b6846-jq4cl\" (UID: \"a556fce5-e24b-4167-85b9-cd5dc0a668fe\") " pod="calico-system/whisker-844d8b6846-jq4cl" Jan 14 00:47:48.422971 kubelet[2802]: I0114 00:47:48.422041 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/dcaab5f4-fd68-41e8-986b-14fa1a168bce-calico-apiserver-certs\") pod \"calico-apiserver-5f9b965548-bgkj8\" (UID: \"dcaab5f4-fd68-41e8-986b-14fa1a168bce\") " pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" Jan 14 00:47:48.422971 kubelet[2802]: I0114 00:47:48.422062 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/976c1fb0-d44d-49e6-bc13-822966e59142-goldmane-ca-bundle\") pod \"goldmane-666569f655-p6wlv\" (UID: \"976c1fb0-d44d-49e6-bc13-822966e59142\") " pod="calico-system/goldmane-666569f655-p6wlv" Jan 14 00:47:48.422971 kubelet[2802]: I0114 00:47:48.422096 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/976c1fb0-d44d-49e6-bc13-822966e59142-config\") pod \"goldmane-666569f655-p6wlv\" (UID: \"976c1fb0-d44d-49e6-bc13-822966e59142\") " pod="calico-system/goldmane-666569f655-p6wlv" Jan 14 00:47:48.422971 kubelet[2802]: I0114 00:47:48.422162 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/976c1fb0-d44d-49e6-bc13-822966e59142-goldmane-key-pair\") pod \"goldmane-666569f655-p6wlv\" (UID: \"976c1fb0-d44d-49e6-bc13-822966e59142\") " pod="calico-system/goldmane-666569f655-p6wlv" Jan 14 00:47:48.618479 kubelet[2802]: E0114 00:47:48.618350 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:48.619537 containerd[1611]: time="2026-01-14T00:47:48.619376487Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-v8f4p,Uid:e4d15318-1381-4cb4-b138-c44ed14a3a71,Namespace:kube-system,Attempt:0,}" Jan 14 00:47:48.639134 containerd[1611]: time="2026-01-14T00:47:48.638808702Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5f75bc94c6-2nthl,Uid:c12ac0b5-659e-4a97-821e-20b4aabfc024,Namespace:calico-system,Attempt:0,}" Jan 14 00:47:48.641630 kubelet[2802]: E0114 00:47:48.640714 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:48.645493 containerd[1611]: time="2026-01-14T00:47:48.645397273Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\"" Jan 14 00:47:48.654631 containerd[1611]: time="2026-01-14T00:47:48.654482195Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-844d8b6846-jq4cl,Uid:a556fce5-e24b-4167-85b9-cd5dc0a668fe,Namespace:calico-system,Attempt:0,}" Jan 14 00:47:48.668295 containerd[1611]: time="2026-01-14T00:47:48.668249212Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5f9b965548-bgkj8,Uid:dcaab5f4-fd68-41e8-986b-14fa1a168bce,Namespace:calico-apiserver,Attempt:0,}" Jan 14 00:47:48.679720 kubelet[2802]: E0114 00:47:48.679565 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:48.680891 containerd[1611]: time="2026-01-14T00:47:48.680854412Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hc5jt,Uid:43d5cf61-6f70-467e-abb2-9b166aef5949,Namespace:kube-system,Attempt:0,}" Jan 14 00:47:48.690410 containerd[1611]: time="2026-01-14T00:47:48.690389250Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-p6wlv,Uid:976c1fb0-d44d-49e6-bc13-822966e59142,Namespace:calico-system,Attempt:0,}" Jan 14 00:47:48.699174 containerd[1611]: time="2026-01-14T00:47:48.699149164Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5f9b965548-xv9gd,Uid:32668954-0a97-46fe-9f01-fa91a530c2e3,Namespace:calico-apiserver,Attempt:0,}" Jan 14 00:47:48.789144 containerd[1611]: time="2026-01-14T00:47:48.789064283Z" level=error msg="Failed to destroy network for sandbox \"483d90d6fd8d6a4cd42a0ba86a7883739028d2a33c167eafa036dad8d741c5d6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.791490 containerd[1611]: time="2026-01-14T00:47:48.791468898Z" level=error msg="Failed to destroy network for sandbox \"2c448364c9c08a42ef2cb1864b628cb4ee5c4444fe3a3035da10bcc32eb57e13\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.799800 containerd[1611]: time="2026-01-14T00:47:48.799767756Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-v8f4p,Uid:e4d15318-1381-4cb4-b138-c44ed14a3a71,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"2c448364c9c08a42ef2cb1864b628cb4ee5c4444fe3a3035da10bcc32eb57e13\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.800124 containerd[1611]: time="2026-01-14T00:47:48.800077334Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5f9b965548-bgkj8,Uid:dcaab5f4-fd68-41e8-986b-14fa1a168bce,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"483d90d6fd8d6a4cd42a0ba86a7883739028d2a33c167eafa036dad8d741c5d6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.801185 kubelet[2802]: E0114 00:47:48.800614 2802 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"483d90d6fd8d6a4cd42a0ba86a7883739028d2a33c167eafa036dad8d741c5d6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.801185 kubelet[2802]: E0114 00:47:48.800669 2802 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"483d90d6fd8d6a4cd42a0ba86a7883739028d2a33c167eafa036dad8d741c5d6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" Jan 14 00:47:48.801185 kubelet[2802]: E0114 00:47:48.800691 2802 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"483d90d6fd8d6a4cd42a0ba86a7883739028d2a33c167eafa036dad8d741c5d6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" Jan 14 00:47:48.801598 kubelet[2802]: E0114 00:47:48.800754 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-5f9b965548-bgkj8_calico-apiserver(dcaab5f4-fd68-41e8-986b-14fa1a168bce)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-5f9b965548-bgkj8_calico-apiserver(dcaab5f4-fd68-41e8-986b-14fa1a168bce)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"483d90d6fd8d6a4cd42a0ba86a7883739028d2a33c167eafa036dad8d741c5d6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" podUID="dcaab5f4-fd68-41e8-986b-14fa1a168bce" Jan 14 00:47:48.801598 kubelet[2802]: E0114 00:47:48.800977 2802 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2c448364c9c08a42ef2cb1864b628cb4ee5c4444fe3a3035da10bcc32eb57e13\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.801598 kubelet[2802]: E0114 00:47:48.801002 2802 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2c448364c9c08a42ef2cb1864b628cb4ee5c4444fe3a3035da10bcc32eb57e13\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-v8f4p" Jan 14 00:47:48.801840 kubelet[2802]: E0114 00:47:48.801016 2802 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2c448364c9c08a42ef2cb1864b628cb4ee5c4444fe3a3035da10bcc32eb57e13\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-v8f4p" Jan 14 00:47:48.801840 kubelet[2802]: E0114 00:47:48.801043 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-v8f4p_kube-system(e4d15318-1381-4cb4-b138-c44ed14a3a71)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-v8f4p_kube-system(e4d15318-1381-4cb4-b138-c44ed14a3a71)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2c448364c9c08a42ef2cb1864b628cb4ee5c4444fe3a3035da10bcc32eb57e13\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-v8f4p" podUID="e4d15318-1381-4cb4-b138-c44ed14a3a71" Jan 14 00:47:48.867223 containerd[1611]: time="2026-01-14T00:47:48.867038929Z" level=error msg="Failed to destroy network for sandbox \"1e99c2bdc7340bc3d16fcf3f0b74ce20dcb63f86eb279978fbf51662e09a6e75\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.871455 containerd[1611]: time="2026-01-14T00:47:48.871290829Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5f75bc94c6-2nthl,Uid:c12ac0b5-659e-4a97-821e-20b4aabfc024,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"1e99c2bdc7340bc3d16fcf3f0b74ce20dcb63f86eb279978fbf51662e09a6e75\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.872190 kubelet[2802]: E0114 00:47:48.871876 2802 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1e99c2bdc7340bc3d16fcf3f0b74ce20dcb63f86eb279978fbf51662e09a6e75\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.872190 kubelet[2802]: E0114 00:47:48.871931 2802 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1e99c2bdc7340bc3d16fcf3f0b74ce20dcb63f86eb279978fbf51662e09a6e75\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" Jan 14 00:47:48.872190 kubelet[2802]: E0114 00:47:48.871951 2802 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1e99c2bdc7340bc3d16fcf3f0b74ce20dcb63f86eb279978fbf51662e09a6e75\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" Jan 14 00:47:48.872311 kubelet[2802]: E0114 00:47:48.872006 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5f75bc94c6-2nthl_calico-system(c12ac0b5-659e-4a97-821e-20b4aabfc024)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5f75bc94c6-2nthl_calico-system(c12ac0b5-659e-4a97-821e-20b4aabfc024)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1e99c2bdc7340bc3d16fcf3f0b74ce20dcb63f86eb279978fbf51662e09a6e75\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" podUID="c12ac0b5-659e-4a97-821e-20b4aabfc024" Jan 14 00:47:48.873454 containerd[1611]: time="2026-01-14T00:47:48.873374717Z" level=error msg="Failed to destroy network for sandbox \"dcd7afff4b6ba5d94ea12449b077a6cf6505e124500c19c77555ae6ec036cd6f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.875851 containerd[1611]: time="2026-01-14T00:47:48.875827225Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5f9b965548-xv9gd,Uid:32668954-0a97-46fe-9f01-fa91a530c2e3,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"dcd7afff4b6ba5d94ea12449b077a6cf6505e124500c19c77555ae6ec036cd6f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.876147 kubelet[2802]: E0114 00:47:48.876035 2802 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dcd7afff4b6ba5d94ea12449b077a6cf6505e124500c19c77555ae6ec036cd6f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.876305 kubelet[2802]: E0114 00:47:48.876220 2802 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dcd7afff4b6ba5d94ea12449b077a6cf6505e124500c19c77555ae6ec036cd6f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" Jan 14 00:47:48.876305 kubelet[2802]: E0114 00:47:48.876242 2802 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dcd7afff4b6ba5d94ea12449b077a6cf6505e124500c19c77555ae6ec036cd6f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" Jan 14 00:47:48.876305 kubelet[2802]: E0114 00:47:48.876277 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-5f9b965548-xv9gd_calico-apiserver(32668954-0a97-46fe-9f01-fa91a530c2e3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-5f9b965548-xv9gd_calico-apiserver(32668954-0a97-46fe-9f01-fa91a530c2e3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"dcd7afff4b6ba5d94ea12449b077a6cf6505e124500c19c77555ae6ec036cd6f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" podUID="32668954-0a97-46fe-9f01-fa91a530c2e3" Jan 14 00:47:48.877639 containerd[1611]: time="2026-01-14T00:47:48.877560382Z" level=error msg="Failed to destroy network for sandbox \"08a011fbc275a82525b12b022d1d578f27c99daea504c0461c415879eb94a114\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.878798 containerd[1611]: time="2026-01-14T00:47:48.878767250Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hc5jt,Uid:43d5cf61-6f70-467e-abb2-9b166aef5949,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"08a011fbc275a82525b12b022d1d578f27c99daea504c0461c415879eb94a114\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.879091 kubelet[2802]: E0114 00:47:48.878971 2802 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"08a011fbc275a82525b12b022d1d578f27c99daea504c0461c415879eb94a114\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.879091 kubelet[2802]: E0114 00:47:48.878997 2802 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"08a011fbc275a82525b12b022d1d578f27c99daea504c0461c415879eb94a114\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-hc5jt" Jan 14 00:47:48.879091 kubelet[2802]: E0114 00:47:48.879013 2802 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"08a011fbc275a82525b12b022d1d578f27c99daea504c0461c415879eb94a114\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-hc5jt" Jan 14 00:47:48.879309 kubelet[2802]: E0114 00:47:48.879038 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-hc5jt_kube-system(43d5cf61-6f70-467e-abb2-9b166aef5949)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-hc5jt_kube-system(43d5cf61-6f70-467e-abb2-9b166aef5949)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"08a011fbc275a82525b12b022d1d578f27c99daea504c0461c415879eb94a114\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-hc5jt" podUID="43d5cf61-6f70-467e-abb2-9b166aef5949" Jan 14 00:47:48.883773 containerd[1611]: time="2026-01-14T00:47:48.883723530Z" level=error msg="Failed to destroy network for sandbox \"ceb762318da57ac0cc845e1850bf636a34d44c0cf111cf8b18091bba8b9b68c6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.886582 containerd[1611]: time="2026-01-14T00:47:48.886528608Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-844d8b6846-jq4cl,Uid:a556fce5-e24b-4167-85b9-cd5dc0a668fe,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ceb762318da57ac0cc845e1850bf636a34d44c0cf111cf8b18091bba8b9b68c6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.887264 kubelet[2802]: E0114 00:47:48.887015 2802 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ceb762318da57ac0cc845e1850bf636a34d44c0cf111cf8b18091bba8b9b68c6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.887264 kubelet[2802]: E0114 00:47:48.887044 2802 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ceb762318da57ac0cc845e1850bf636a34d44c0cf111cf8b18091bba8b9b68c6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-844d8b6846-jq4cl" Jan 14 00:47:48.887264 kubelet[2802]: E0114 00:47:48.887060 2802 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ceb762318da57ac0cc845e1850bf636a34d44c0cf111cf8b18091bba8b9b68c6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-844d8b6846-jq4cl" Jan 14 00:47:48.887373 kubelet[2802]: E0114 00:47:48.887087 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-844d8b6846-jq4cl_calico-system(a556fce5-e24b-4167-85b9-cd5dc0a668fe)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-844d8b6846-jq4cl_calico-system(a556fce5-e24b-4167-85b9-cd5dc0a668fe)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ceb762318da57ac0cc845e1850bf636a34d44c0cf111cf8b18091bba8b9b68c6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-844d8b6846-jq4cl" podUID="a556fce5-e24b-4167-85b9-cd5dc0a668fe" Jan 14 00:47:48.902980 containerd[1611]: time="2026-01-14T00:47:48.902943274Z" level=error msg="Failed to destroy network for sandbox \"6a22677d44f1346524ff07ffdb634623e16b6923be2f5b2a78f84a6f439c671c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.904703 containerd[1611]: time="2026-01-14T00:47:48.904677171Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-p6wlv,Uid:976c1fb0-d44d-49e6-bc13-822966e59142,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"6a22677d44f1346524ff07ffdb634623e16b6923be2f5b2a78f84a6f439c671c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.904881 kubelet[2802]: E0114 00:47:48.904853 2802 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6a22677d44f1346524ff07ffdb634623e16b6923be2f5b2a78f84a6f439c671c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:48.904986 kubelet[2802]: E0114 00:47:48.904884 2802 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6a22677d44f1346524ff07ffdb634623e16b6923be2f5b2a78f84a6f439c671c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-p6wlv" Jan 14 00:47:48.904986 kubelet[2802]: E0114 00:47:48.904901 2802 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6a22677d44f1346524ff07ffdb634623e16b6923be2f5b2a78f84a6f439c671c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-666569f655-p6wlv" Jan 14 00:47:48.904986 kubelet[2802]: E0114 00:47:48.904936 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-666569f655-p6wlv_calico-system(976c1fb0-d44d-49e6-bc13-822966e59142)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-666569f655-p6wlv_calico-system(976c1fb0-d44d-49e6-bc13-822966e59142)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6a22677d44f1346524ff07ffdb634623e16b6923be2f5b2a78f84a6f439c671c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-666569f655-p6wlv" podUID="976c1fb0-d44d-49e6-bc13-822966e59142" Jan 14 00:47:49.507554 systemd[1]: Created slice kubepods-besteffort-podf41cc627_bd07_42f1_8e88_f85dcd2c063c.slice - libcontainer container kubepods-besteffort-podf41cc627_bd07_42f1_8e88_f85dcd2c063c.slice. Jan 14 00:47:49.515784 containerd[1611]: time="2026-01-14T00:47:49.515741616Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-vzqkb,Uid:f41cc627-bd07-42f1-8e88-f85dcd2c063c,Namespace:calico-system,Attempt:0,}" Jan 14 00:47:49.602158 containerd[1611]: time="2026-01-14T00:47:49.599990213Z" level=error msg="Failed to destroy network for sandbox \"4346901f48fd72b3bc577678cd8e095a976510fac7104a8257cd863b7be54158\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:49.602158 containerd[1611]: time="2026-01-14T00:47:49.601424031Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-vzqkb,Uid:f41cc627-bd07-42f1-8e88-f85dcd2c063c,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"4346901f48fd72b3bc577678cd8e095a976510fac7104a8257cd863b7be54158\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:49.607659 kubelet[2802]: E0114 00:47:49.605576 2802 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4346901f48fd72b3bc577678cd8e095a976510fac7104a8257cd863b7be54158\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 14 00:47:49.607659 kubelet[2802]: E0114 00:47:49.607289 2802 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4346901f48fd72b3bc577678cd8e095a976510fac7104a8257cd863b7be54158\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-vzqkb" Jan 14 00:47:49.607659 kubelet[2802]: E0114 00:47:49.607334 2802 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4346901f48fd72b3bc577678cd8e095a976510fac7104a8257cd863b7be54158\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-vzqkb" Jan 14 00:47:49.606180 systemd[1]: run-netns-cni\x2d0ea63391\x2d2d30\x2daccf\x2d3aba\x2d392bb6e297ae.mount: Deactivated successfully. Jan 14 00:47:49.607892 kubelet[2802]: E0114 00:47:49.607402 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-vzqkb_calico-system(f41cc627-bd07-42f1-8e88-f85dcd2c063c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-vzqkb_calico-system(f41cc627-bd07-42f1-8e88-f85dcd2c063c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4346901f48fd72b3bc577678cd8e095a976510fac7104a8257cd863b7be54158\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:47:51.966153 kubelet[2802]: I0114 00:47:51.966016 2802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 14 00:47:51.966718 kubelet[2802]: E0114 00:47:51.966444 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:52.017000 audit[3818]: NETFILTER_CFG table=filter:117 family=2 entries=21 op=nft_register_rule pid=3818 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:52.017000 audit[3818]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffc9ada43b0 a2=0 a3=7ffc9ada439c items=0 ppid=2943 pid=3818 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:52.017000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:52.024000 audit[3818]: NETFILTER_CFG table=nat:118 family=2 entries=19 op=nft_register_chain pid=3818 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:52.024000 audit[3818]: SYSCALL arch=c000003e syscall=46 success=yes exit=6276 a0=3 a1=7ffc9ada43b0 a2=0 a3=7ffc9ada439c items=0 ppid=2943 pid=3818 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:52.024000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:52.442904 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2625746832.mount: Deactivated successfully. Jan 14 00:47:52.477956 containerd[1611]: time="2026-01-14T00:47:52.477909735Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:52.478617 containerd[1611]: time="2026-01-14T00:47:52.478586307Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.4: active requests=0, bytes read=156880025" Jan 14 00:47:52.479424 containerd[1611]: time="2026-01-14T00:47:52.479367705Z" level=info msg="ImageCreate event name:\"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:52.481638 containerd[1611]: time="2026-01-14T00:47:52.481595200Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 14 00:47:52.482489 containerd[1611]: time="2026-01-14T00:47:52.482260133Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.4\" with image id \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.4\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e92cca333202c87d07bf57f38182fd68f0779f912ef55305eda1fccc9f33667c\", size \"156883537\" in 3.836827578s" Jan 14 00:47:52.482489 containerd[1611]: time="2026-01-14T00:47:52.482483613Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.4\" returns image reference \"sha256:833e8e11d9dc187377eab6f31e275114a6b0f8f0afc3bf578a2a00507e85afc9\"" Jan 14 00:47:52.507469 containerd[1611]: time="2026-01-14T00:47:52.507409850Z" level=info msg="CreateContainer within sandbox \"d81288ef4b21c802a5269ff94a28696b0ee21fba084a353c0f40fafd58f9160a\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Jan 14 00:47:52.518717 containerd[1611]: time="2026-01-14T00:47:52.516409808Z" level=info msg="Container ac16c85845a5c8e1029e8567062694f24c9d1db1cf825a767750fd77cca2d185: CDI devices from CRI Config.CDIDevices: []" Jan 14 00:47:52.527377 containerd[1611]: time="2026-01-14T00:47:52.527336888Z" level=info msg="CreateContainer within sandbox \"d81288ef4b21c802a5269ff94a28696b0ee21fba084a353c0f40fafd58f9160a\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"ac16c85845a5c8e1029e8567062694f24c9d1db1cf825a767750fd77cca2d185\"" Jan 14 00:47:52.528271 containerd[1611]: time="2026-01-14T00:47:52.528242171Z" level=info msg="StartContainer for \"ac16c85845a5c8e1029e8567062694f24c9d1db1cf825a767750fd77cca2d185\"" Jan 14 00:47:52.529797 containerd[1611]: time="2026-01-14T00:47:52.529777663Z" level=info msg="connecting to shim ac16c85845a5c8e1029e8567062694f24c9d1db1cf825a767750fd77cca2d185" address="unix:///run/containerd/s/73f4dba3b46309f40b046e6c8440ce78be70cdcdea47bcd2729be2f201321bd9" protocol=ttrpc version=3 Jan 14 00:47:52.577279 systemd[1]: Started cri-containerd-ac16c85845a5c8e1029e8567062694f24c9d1db1cf825a767750fd77cca2d185.scope - libcontainer container ac16c85845a5c8e1029e8567062694f24c9d1db1cf825a767750fd77cca2d185. Jan 14 00:47:52.643000 audit: BPF prog-id=184 op=LOAD Jan 14 00:47:52.644217 kernel: kauditd_printk_skb: 12 callbacks suppressed Jan 14 00:47:52.644273 kernel: audit: type=1334 audit(1768351672.643:586): prog-id=184 op=LOAD Jan 14 00:47:52.655137 kernel: audit: type=1300 audit(1768351672.643:586): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000186488 a2=98 a3=0 items=0 ppid=3336 pid=3821 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:52.643000 audit[3821]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000186488 a2=98 a3=0 items=0 ppid=3336 pid=3821 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:52.643000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6163313663383538343561356338653130323965383536373036323639 Jan 14 00:47:52.658770 kernel: audit: type=1327 audit(1768351672.643:586): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6163313663383538343561356338653130323965383536373036323639 Jan 14 00:47:52.667364 kernel: audit: type=1334 audit(1768351672.647:587): prog-id=185 op=LOAD Jan 14 00:47:52.647000 audit: BPF prog-id=185 op=LOAD Jan 14 00:47:52.647000 audit[3821]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000186218 a2=98 a3=0 items=0 ppid=3336 pid=3821 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:52.669180 kernel: audit: type=1300 audit(1768351672.647:587): arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000186218 a2=98 a3=0 items=0 ppid=3336 pid=3821 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:52.685344 kernel: audit: type=1327 audit(1768351672.647:587): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6163313663383538343561356338653130323965383536373036323639 Jan 14 00:47:52.647000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6163313663383538343561356338653130323965383536373036323639 Jan 14 00:47:52.647000 audit: BPF prog-id=185 op=UNLOAD Jan 14 00:47:52.699734 kernel: audit: type=1334 audit(1768351672.647:588): prog-id=185 op=UNLOAD Jan 14 00:47:52.699787 kernel: audit: type=1300 audit(1768351672.647:588): arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3336 pid=3821 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:52.647000 audit[3821]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=3336 pid=3821 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:52.699935 kubelet[2802]: E0114 00:47:52.692335 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:52.647000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6163313663383538343561356338653130323965383536373036323639 Jan 14 00:47:52.711373 kernel: audit: type=1327 audit(1768351672.647:588): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6163313663383538343561356338653130323965383536373036323639 Jan 14 00:47:52.711435 kernel: audit: type=1334 audit(1768351672.647:589): prog-id=184 op=UNLOAD Jan 14 00:47:52.647000 audit: BPF prog-id=184 op=UNLOAD Jan 14 00:47:52.647000 audit[3821]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=3336 pid=3821 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:52.647000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6163313663383538343561356338653130323965383536373036323639 Jan 14 00:47:52.647000 audit: BPF prog-id=186 op=LOAD Jan 14 00:47:52.647000 audit[3821]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001866e8 a2=98 a3=0 items=0 ppid=3336 pid=3821 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:52.647000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6163313663383538343561356338653130323965383536373036323639 Jan 14 00:47:52.717310 containerd[1611]: time="2026-01-14T00:47:52.717274176Z" level=info msg="StartContainer for \"ac16c85845a5c8e1029e8567062694f24c9d1db1cf825a767750fd77cca2d185\" returns successfully" Jan 14 00:47:52.813904 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Jan 14 00:47:52.814017 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Jan 14 00:47:52.954786 kubelet[2802]: I0114 00:47:52.954672 2802 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a556fce5-e24b-4167-85b9-cd5dc0a668fe-whisker-ca-bundle\") pod \"a556fce5-e24b-4167-85b9-cd5dc0a668fe\" (UID: \"a556fce5-e24b-4167-85b9-cd5dc0a668fe\") " Jan 14 00:47:52.956127 kubelet[2802]: I0114 00:47:52.955176 2802 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/a556fce5-e24b-4167-85b9-cd5dc0a668fe-whisker-backend-key-pair\") pod \"a556fce5-e24b-4167-85b9-cd5dc0a668fe\" (UID: \"a556fce5-e24b-4167-85b9-cd5dc0a668fe\") " Jan 14 00:47:52.956127 kubelet[2802]: I0114 00:47:52.955204 2802 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pl6mw\" (UniqueName: \"kubernetes.io/projected/a556fce5-e24b-4167-85b9-cd5dc0a668fe-kube-api-access-pl6mw\") pod \"a556fce5-e24b-4167-85b9-cd5dc0a668fe\" (UID: \"a556fce5-e24b-4167-85b9-cd5dc0a668fe\") " Jan 14 00:47:52.957139 kubelet[2802]: I0114 00:47:52.956577 2802 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a556fce5-e24b-4167-85b9-cd5dc0a668fe-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "a556fce5-e24b-4167-85b9-cd5dc0a668fe" (UID: "a556fce5-e24b-4167-85b9-cd5dc0a668fe"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Jan 14 00:47:52.962227 kubelet[2802]: I0114 00:47:52.962205 2802 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a556fce5-e24b-4167-85b9-cd5dc0a668fe-kube-api-access-pl6mw" (OuterVolumeSpecName: "kube-api-access-pl6mw") pod "a556fce5-e24b-4167-85b9-cd5dc0a668fe" (UID: "a556fce5-e24b-4167-85b9-cd5dc0a668fe"). InnerVolumeSpecName "kube-api-access-pl6mw". PluginName "kubernetes.io/projected", VolumeGIDValue "" Jan 14 00:47:52.966381 kubelet[2802]: I0114 00:47:52.966360 2802 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a556fce5-e24b-4167-85b9-cd5dc0a668fe-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "a556fce5-e24b-4167-85b9-cd5dc0a668fe" (UID: "a556fce5-e24b-4167-85b9-cd5dc0a668fe"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Jan 14 00:47:53.056212 kubelet[2802]: I0114 00:47:53.056141 2802 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/a556fce5-e24b-4167-85b9-cd5dc0a668fe-whisker-backend-key-pair\") on node \"172-236-110-177\" DevicePath \"\"" Jan 14 00:47:53.056212 kubelet[2802]: I0114 00:47:53.056176 2802 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pl6mw\" (UniqueName: \"kubernetes.io/projected/a556fce5-e24b-4167-85b9-cd5dc0a668fe-kube-api-access-pl6mw\") on node \"172-236-110-177\" DevicePath \"\"" Jan 14 00:47:53.056212 kubelet[2802]: I0114 00:47:53.056187 2802 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a556fce5-e24b-4167-85b9-cd5dc0a668fe-whisker-ca-bundle\") on node \"172-236-110-177\" DevicePath \"\"" Jan 14 00:47:53.441344 systemd[1]: var-lib-kubelet-pods-a556fce5\x2de24b\x2d4167\x2d85b9\x2dcd5dc0a668fe-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dpl6mw.mount: Deactivated successfully. Jan 14 00:47:53.441463 systemd[1]: var-lib-kubelet-pods-a556fce5\x2de24b\x2d4167\x2d85b9\x2dcd5dc0a668fe-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Jan 14 00:47:53.510365 systemd[1]: Removed slice kubepods-besteffort-poda556fce5_e24b_4167_85b9_cd5dc0a668fe.slice - libcontainer container kubepods-besteffort-poda556fce5_e24b_4167_85b9_cd5dc0a668fe.slice. Jan 14 00:47:53.703487 kubelet[2802]: E0114 00:47:53.703374 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:53.719407 kubelet[2802]: I0114 00:47:53.719032 2802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-28tb6" podStartSLOduration=2.70909689 podStartE2EDuration="12.719019435s" podCreationTimestamp="2026-01-14 00:47:41 +0000 UTC" firstStartedPulling="2026-01-14 00:47:42.473577527 +0000 UTC m=+21.068673941" lastFinishedPulling="2026-01-14 00:47:52.483500072 +0000 UTC m=+31.078596486" observedRunningTime="2026-01-14 00:47:53.718383916 +0000 UTC m=+32.313480340" watchObservedRunningTime="2026-01-14 00:47:53.719019435 +0000 UTC m=+32.314115849" Jan 14 00:47:53.793264 systemd[1]: Created slice kubepods-besteffort-poddb919709_47a0_46a4_a2ab_3f096374d888.slice - libcontainer container kubepods-besteffort-poddb919709_47a0_46a4_a2ab_3f096374d888.slice. Jan 14 00:47:53.860994 kubelet[2802]: I0114 00:47:53.860928 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/db919709-47a0-46a4-a2ab-3f096374d888-whisker-ca-bundle\") pod \"whisker-d6f8b64b4-m4ctk\" (UID: \"db919709-47a0-46a4-a2ab-3f096374d888\") " pod="calico-system/whisker-d6f8b64b4-m4ctk" Jan 14 00:47:53.860994 kubelet[2802]: I0114 00:47:53.860978 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/db919709-47a0-46a4-a2ab-3f096374d888-whisker-backend-key-pair\") pod \"whisker-d6f8b64b4-m4ctk\" (UID: \"db919709-47a0-46a4-a2ab-3f096374d888\") " pod="calico-system/whisker-d6f8b64b4-m4ctk" Jan 14 00:47:53.860994 kubelet[2802]: I0114 00:47:53.860997 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdmgl\" (UniqueName: \"kubernetes.io/projected/db919709-47a0-46a4-a2ab-3f096374d888-kube-api-access-rdmgl\") pod \"whisker-d6f8b64b4-m4ctk\" (UID: \"db919709-47a0-46a4-a2ab-3f096374d888\") " pod="calico-system/whisker-d6f8b64b4-m4ctk" Jan 14 00:47:54.099084 containerd[1611]: time="2026-01-14T00:47:54.098978644Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-d6f8b64b4-m4ctk,Uid:db919709-47a0-46a4-a2ab-3f096374d888,Namespace:calico-system,Attempt:0,}" Jan 14 00:47:54.266014 systemd-networkd[1504]: calif721a9bde91: Link UP Jan 14 00:47:54.269289 systemd-networkd[1504]: calif721a9bde91: Gained carrier Jan 14 00:47:54.294052 containerd[1611]: 2026-01-14 00:47:54.129 [INFO][3889] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 14 00:47:54.294052 containerd[1611]: 2026-01-14 00:47:54.166 [INFO][3889] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--236--110--177-k8s-whisker--d6f8b64b4--m4ctk-eth0 whisker-d6f8b64b4- calico-system db919709-47a0-46a4-a2ab-3f096374d888 954 0 2026-01-14 00:47:53 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:d6f8b64b4 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s 172-236-110-177 whisker-d6f8b64b4-m4ctk eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calif721a9bde91 [] [] }} ContainerID="337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" Namespace="calico-system" Pod="whisker-d6f8b64b4-m4ctk" WorkloadEndpoint="172--236--110--177-k8s-whisker--d6f8b64b4--m4ctk-" Jan 14 00:47:54.294052 containerd[1611]: 2026-01-14 00:47:54.167 [INFO][3889] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" Namespace="calico-system" Pod="whisker-d6f8b64b4-m4ctk" WorkloadEndpoint="172--236--110--177-k8s-whisker--d6f8b64b4--m4ctk-eth0" Jan 14 00:47:54.294052 containerd[1611]: 2026-01-14 00:47:54.205 [INFO][3900] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" HandleID="k8s-pod-network.337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" Workload="172--236--110--177-k8s-whisker--d6f8b64b4--m4ctk-eth0" Jan 14 00:47:54.294242 containerd[1611]: 2026-01-14 00:47:54.206 [INFO][3900] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" HandleID="k8s-pod-network.337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" Workload="172--236--110--177-k8s-whisker--d6f8b64b4--m4ctk-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002cb1c0), Attrs:map[string]string{"namespace":"calico-system", "node":"172-236-110-177", "pod":"whisker-d6f8b64b4-m4ctk", "timestamp":"2026-01-14 00:47:54.205739098 +0000 UTC"}, Hostname:"172-236-110-177", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 00:47:54.294242 containerd[1611]: 2026-01-14 00:47:54.206 [INFO][3900] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 00:47:54.294242 containerd[1611]: 2026-01-14 00:47:54.206 [INFO][3900] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 00:47:54.294242 containerd[1611]: 2026-01-14 00:47:54.206 [INFO][3900] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-236-110-177' Jan 14 00:47:54.294242 containerd[1611]: 2026-01-14 00:47:54.214 [INFO][3900] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" host="172-236-110-177" Jan 14 00:47:54.294242 containerd[1611]: 2026-01-14 00:47:54.219 [INFO][3900] ipam/ipam.go 394: Looking up existing affinities for host host="172-236-110-177" Jan 14 00:47:54.294242 containerd[1611]: 2026-01-14 00:47:54.225 [INFO][3900] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="172-236-110-177" Jan 14 00:47:54.294242 containerd[1611]: 2026-01-14 00:47:54.226 [INFO][3900] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:47:54.294242 containerd[1611]: 2026-01-14 00:47:54.229 [INFO][3900] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:47:54.294432 containerd[1611]: 2026-01-14 00:47:54.229 [INFO][3900] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" host="172-236-110-177" Jan 14 00:47:54.294432 containerd[1611]: 2026-01-14 00:47:54.231 [INFO][3900] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05 Jan 14 00:47:54.294432 containerd[1611]: 2026-01-14 00:47:54.236 [INFO][3900] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" host="172-236-110-177" Jan 14 00:47:54.294432 containerd[1611]: 2026-01-14 00:47:54.240 [INFO][3900] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.193/26] block=192.168.109.192/26 handle="k8s-pod-network.337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" host="172-236-110-177" Jan 14 00:47:54.294432 containerd[1611]: 2026-01-14 00:47:54.240 [INFO][3900] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.193/26] handle="k8s-pod-network.337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" host="172-236-110-177" Jan 14 00:47:54.294432 containerd[1611]: 2026-01-14 00:47:54.240 [INFO][3900] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 00:47:54.294432 containerd[1611]: 2026-01-14 00:47:54.240 [INFO][3900] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.193/26] IPv6=[] ContainerID="337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" HandleID="k8s-pod-network.337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" Workload="172--236--110--177-k8s-whisker--d6f8b64b4--m4ctk-eth0" Jan 14 00:47:54.294562 containerd[1611]: 2026-01-14 00:47:54.250 [INFO][3889] cni-plugin/k8s.go 418: Populated endpoint ContainerID="337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" Namespace="calico-system" Pod="whisker-d6f8b64b4-m4ctk" WorkloadEndpoint="172--236--110--177-k8s-whisker--d6f8b64b4--m4ctk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-whisker--d6f8b64b4--m4ctk-eth0", GenerateName:"whisker-d6f8b64b4-", Namespace:"calico-system", SelfLink:"", UID:"db919709-47a0-46a4-a2ab-3f096374d888", ResourceVersion:"954", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"d6f8b64b4", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"", Pod:"whisker-d6f8b64b4-m4ctk", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.109.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calif721a9bde91", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:47:54.294562 containerd[1611]: 2026-01-14 00:47:54.250 [INFO][3889] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.193/32] ContainerID="337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" Namespace="calico-system" Pod="whisker-d6f8b64b4-m4ctk" WorkloadEndpoint="172--236--110--177-k8s-whisker--d6f8b64b4--m4ctk-eth0" Jan 14 00:47:54.294637 containerd[1611]: 2026-01-14 00:47:54.250 [INFO][3889] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif721a9bde91 ContainerID="337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" Namespace="calico-system" Pod="whisker-d6f8b64b4-m4ctk" WorkloadEndpoint="172--236--110--177-k8s-whisker--d6f8b64b4--m4ctk-eth0" Jan 14 00:47:54.294637 containerd[1611]: 2026-01-14 00:47:54.270 [INFO][3889] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" Namespace="calico-system" Pod="whisker-d6f8b64b4-m4ctk" WorkloadEndpoint="172--236--110--177-k8s-whisker--d6f8b64b4--m4ctk-eth0" Jan 14 00:47:54.294675 containerd[1611]: 2026-01-14 00:47:54.272 [INFO][3889] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" Namespace="calico-system" Pod="whisker-d6f8b64b4-m4ctk" WorkloadEndpoint="172--236--110--177-k8s-whisker--d6f8b64b4--m4ctk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-whisker--d6f8b64b4--m4ctk-eth0", GenerateName:"whisker-d6f8b64b4-", Namespace:"calico-system", SelfLink:"", UID:"db919709-47a0-46a4-a2ab-3f096374d888", ResourceVersion:"954", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"d6f8b64b4", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05", Pod:"whisker-d6f8b64b4-m4ctk", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.109.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calif721a9bde91", MAC:"2a:ff:50:79:31:58", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:47:54.294724 containerd[1611]: 2026-01-14 00:47:54.288 [INFO][3889] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" Namespace="calico-system" Pod="whisker-d6f8b64b4-m4ctk" WorkloadEndpoint="172--236--110--177-k8s-whisker--d6f8b64b4--m4ctk-eth0" Jan 14 00:47:54.367064 containerd[1611]: time="2026-01-14T00:47:54.366912885Z" level=info msg="connecting to shim 337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05" address="unix:///run/containerd/s/6dad4f16b9963e5801e264fd7fbf4437b3eea015fc69c43d1137361fada97dfe" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:47:54.434474 systemd[1]: Started cri-containerd-337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05.scope - libcontainer container 337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05. Jan 14 00:47:54.474000 audit: BPF prog-id=187 op=LOAD Jan 14 00:47:54.475000 audit: BPF prog-id=188 op=LOAD Jan 14 00:47:54.475000 audit[4016]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=4005 pid=4016 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.475000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333373034386136366432353137396362303131353862333363376666 Jan 14 00:47:54.475000 audit: BPF prog-id=188 op=UNLOAD Jan 14 00:47:54.475000 audit[4016]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4005 pid=4016 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.475000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333373034386136366432353137396362303131353862333363376666 Jan 14 00:47:54.475000 audit: BPF prog-id=189 op=LOAD Jan 14 00:47:54.475000 audit[4016]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=4005 pid=4016 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.475000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333373034386136366432353137396362303131353862333363376666 Jan 14 00:47:54.475000 audit: BPF prog-id=190 op=LOAD Jan 14 00:47:54.475000 audit[4016]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=4005 pid=4016 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.475000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333373034386136366432353137396362303131353862333363376666 Jan 14 00:47:54.475000 audit: BPF prog-id=190 op=UNLOAD Jan 14 00:47:54.475000 audit[4016]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4005 pid=4016 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.475000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333373034386136366432353137396362303131353862333363376666 Jan 14 00:47:54.475000 audit: BPF prog-id=189 op=UNLOAD Jan 14 00:47:54.475000 audit[4016]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4005 pid=4016 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.475000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333373034386136366432353137396362303131353862333363376666 Jan 14 00:47:54.475000 audit: BPF prog-id=191 op=LOAD Jan 14 00:47:54.475000 audit[4016]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=4005 pid=4016 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.475000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3333373034386136366432353137396362303131353862333363376666 Jan 14 00:47:54.571884 containerd[1611]: time="2026-01-14T00:47:54.571831116Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-d6f8b64b4-m4ctk,Uid:db919709-47a0-46a4-a2ab-3f096374d888,Namespace:calico-system,Attempt:0,} returns sandbox id \"337048a66d25179cb01158b33c7ff233d8dfc1e84f47d4f80067b7b2ab259d05\"" Jan 14 00:47:54.576444 containerd[1611]: time="2026-01-14T00:47:54.576216219Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 14 00:47:54.680000 audit: BPF prog-id=192 op=LOAD Jan 14 00:47:54.680000 audit[4058]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffc8473a240 a2=98 a3=1fffffffffffffff items=0 ppid=3940 pid=4058 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.680000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 14 00:47:54.681000 audit: BPF prog-id=192 op=UNLOAD Jan 14 00:47:54.681000 audit[4058]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffc8473a210 a3=0 items=0 ppid=3940 pid=4058 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.681000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 14 00:47:54.681000 audit: BPF prog-id=193 op=LOAD Jan 14 00:47:54.681000 audit[4058]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffc8473a120 a2=94 a3=3 items=0 ppid=3940 pid=4058 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.681000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 14 00:47:54.681000 audit: BPF prog-id=193 op=UNLOAD Jan 14 00:47:54.681000 audit[4058]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffc8473a120 a2=94 a3=3 items=0 ppid=3940 pid=4058 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.681000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 14 00:47:54.681000 audit: BPF prog-id=194 op=LOAD Jan 14 00:47:54.681000 audit[4058]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffc8473a160 a2=94 a3=7ffc8473a340 items=0 ppid=3940 pid=4058 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.681000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 14 00:47:54.681000 audit: BPF prog-id=194 op=UNLOAD Jan 14 00:47:54.681000 audit[4058]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffc8473a160 a2=94 a3=7ffc8473a340 items=0 ppid=3940 pid=4058 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.681000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Jan 14 00:47:54.692000 audit: BPF prog-id=195 op=LOAD Jan 14 00:47:54.692000 audit[4059]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe7135b5c0 a2=98 a3=3 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.692000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.692000 audit: BPF prog-id=195 op=UNLOAD Jan 14 00:47:54.692000 audit[4059]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffe7135b590 a3=0 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.692000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.693000 audit: BPF prog-id=196 op=LOAD Jan 14 00:47:54.693000 audit[4059]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe7135b3b0 a2=94 a3=54428f items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.693000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.693000 audit: BPF prog-id=196 op=UNLOAD Jan 14 00:47:54.693000 audit[4059]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe7135b3b0 a2=94 a3=54428f items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.693000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.693000 audit: BPF prog-id=197 op=LOAD Jan 14 00:47:54.693000 audit[4059]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe7135b3e0 a2=94 a3=2 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.693000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.693000 audit: BPF prog-id=197 op=UNLOAD Jan 14 00:47:54.693000 audit[4059]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe7135b3e0 a2=0 a3=2 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.693000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.703324 containerd[1611]: time="2026-01-14T00:47:54.703170732Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:47:54.704163 kubelet[2802]: I0114 00:47:54.704068 2802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 14 00:47:54.705821 containerd[1611]: time="2026-01-14T00:47:54.705630840Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 14 00:47:54.705821 containerd[1611]: time="2026-01-14T00:47:54.705732424Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 14 00:47:54.706398 kubelet[2802]: E0114 00:47:54.706300 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:47:54.707271 kubelet[2802]: E0114 00:47:54.706546 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 14 00:47:54.707377 kubelet[2802]: E0114 00:47:54.707361 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 14 00:47:54.711235 kubelet[2802]: E0114 00:47:54.711155 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:3e9ed11e1ee14277abd17039bcf9cf15,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rdmgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-d6f8b64b4-m4ctk_calico-system(db919709-47a0-46a4-a2ab-3f096374d888): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 14 00:47:54.714763 containerd[1611]: time="2026-01-14T00:47:54.714030439Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 14 00:47:54.839368 containerd[1611]: time="2026-01-14T00:47:54.839332670Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:47:54.840506 containerd[1611]: time="2026-01-14T00:47:54.840481350Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 14 00:47:54.840687 containerd[1611]: time="2026-01-14T00:47:54.840570634Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 14 00:47:54.840885 kubelet[2802]: E0114 00:47:54.840846 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 14 00:47:54.840974 kubelet[2802]: E0114 00:47:54.840894 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 14 00:47:54.841036 kubelet[2802]: E0114 00:47:54.840996 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdmgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-d6f8b64b4-m4ctk_calico-system(db919709-47a0-46a4-a2ab-3f096374d888): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 14 00:47:54.842363 kubelet[2802]: E0114 00:47:54.842334 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-d6f8b64b4-m4ctk" podUID="db919709-47a0-46a4-a2ab-3f096374d888" Jan 14 00:47:54.956000 audit: BPF prog-id=198 op=LOAD Jan 14 00:47:54.956000 audit[4059]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7ffe7135b2a0 a2=94 a3=1 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.956000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.956000 audit: BPF prog-id=198 op=UNLOAD Jan 14 00:47:54.956000 audit[4059]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7ffe7135b2a0 a2=94 a3=1 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.956000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.965000 audit: BPF prog-id=199 op=LOAD Jan 14 00:47:54.965000 audit[4059]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe7135b290 a2=94 a3=4 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.965000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.965000 audit: BPF prog-id=199 op=UNLOAD Jan 14 00:47:54.965000 audit[4059]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffe7135b290 a2=0 a3=4 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.965000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.966000 audit: BPF prog-id=200 op=LOAD Jan 14 00:47:54.966000 audit[4059]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffe7135b0f0 a2=94 a3=5 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.966000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.966000 audit: BPF prog-id=200 op=UNLOAD Jan 14 00:47:54.966000 audit[4059]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7ffe7135b0f0 a2=0 a3=5 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.966000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.966000 audit: BPF prog-id=201 op=LOAD Jan 14 00:47:54.966000 audit[4059]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe7135b310 a2=94 a3=6 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.966000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.966000 audit: BPF prog-id=201 op=UNLOAD Jan 14 00:47:54.966000 audit[4059]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7ffe7135b310 a2=0 a3=6 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.966000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.966000 audit: BPF prog-id=202 op=LOAD Jan 14 00:47:54.966000 audit[4059]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7ffe7135aac0 a2=94 a3=88 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.966000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.966000 audit: BPF prog-id=203 op=LOAD Jan 14 00:47:54.966000 audit[4059]: SYSCALL arch=c000003e syscall=321 success=yes exit=7 a0=5 a1=7ffe7135a940 a2=94 a3=2 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.966000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.966000 audit: BPF prog-id=203 op=UNLOAD Jan 14 00:47:54.966000 audit[4059]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=7 a1=7ffe7135a970 a2=0 a3=7ffe7135aa70 items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.966000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.967000 audit: BPF prog-id=202 op=UNLOAD Jan 14 00:47:54.967000 audit[4059]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=2a63cd10 a2=0 a3=c02cbd63c8c8255b items=0 ppid=3940 pid=4059 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.967000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Jan 14 00:47:54.981000 audit: BPF prog-id=204 op=LOAD Jan 14 00:47:54.981000 audit[4083]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd73422ae0 a2=98 a3=1999999999999999 items=0 ppid=3940 pid=4083 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.981000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 14 00:47:54.981000 audit: BPF prog-id=204 op=UNLOAD Jan 14 00:47:54.981000 audit[4083]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7ffd73422ab0 a3=0 items=0 ppid=3940 pid=4083 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.981000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 14 00:47:54.981000 audit: BPF prog-id=205 op=LOAD Jan 14 00:47:54.981000 audit[4083]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd734229c0 a2=94 a3=ffff items=0 ppid=3940 pid=4083 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.981000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 14 00:47:54.981000 audit: BPF prog-id=205 op=UNLOAD Jan 14 00:47:54.981000 audit[4083]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffd734229c0 a2=94 a3=ffff items=0 ppid=3940 pid=4083 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.981000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 14 00:47:54.981000 audit: BPF prog-id=206 op=LOAD Jan 14 00:47:54.981000 audit[4083]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffd73422a00 a2=94 a3=7ffd73422be0 items=0 ppid=3940 pid=4083 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.981000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 14 00:47:54.981000 audit: BPF prog-id=206 op=UNLOAD Jan 14 00:47:54.981000 audit[4083]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7ffd73422a00 a2=94 a3=7ffd73422be0 items=0 ppid=3940 pid=4083 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:54.981000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Jan 14 00:47:55.056187 systemd-networkd[1504]: vxlan.calico: Link UP Jan 14 00:47:55.056196 systemd-networkd[1504]: vxlan.calico: Gained carrier Jan 14 00:47:55.077000 audit: BPF prog-id=207 op=LOAD Jan 14 00:47:55.077000 audit[4108]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7fff46229030 a2=98 a3=0 items=0 ppid=3940 pid=4108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.077000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 00:47:55.077000 audit: BPF prog-id=207 op=UNLOAD Jan 14 00:47:55.077000 audit[4108]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7fff46229000 a3=0 items=0 ppid=3940 pid=4108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.077000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 00:47:55.077000 audit: BPF prog-id=208 op=LOAD Jan 14 00:47:55.077000 audit[4108]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7fff46228e40 a2=94 a3=54428f items=0 ppid=3940 pid=4108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.077000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 00:47:55.078000 audit: BPF prog-id=208 op=UNLOAD Jan 14 00:47:55.078000 audit[4108]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7fff46228e40 a2=94 a3=54428f items=0 ppid=3940 pid=4108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.078000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 00:47:55.078000 audit: BPF prog-id=209 op=LOAD Jan 14 00:47:55.078000 audit[4108]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7fff46228e70 a2=94 a3=2 items=0 ppid=3940 pid=4108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.078000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 00:47:55.078000 audit: BPF prog-id=209 op=UNLOAD Jan 14 00:47:55.078000 audit[4108]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=7fff46228e70 a2=0 a3=2 items=0 ppid=3940 pid=4108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.078000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 00:47:55.078000 audit: BPF prog-id=210 op=LOAD Jan 14 00:47:55.078000 audit[4108]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fff46228c20 a2=94 a3=4 items=0 ppid=3940 pid=4108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.078000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 00:47:55.078000 audit: BPF prog-id=210 op=UNLOAD Jan 14 00:47:55.078000 audit[4108]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7fff46228c20 a2=94 a3=4 items=0 ppid=3940 pid=4108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.078000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 00:47:55.078000 audit: BPF prog-id=211 op=LOAD Jan 14 00:47:55.078000 audit[4108]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fff46228d20 a2=94 a3=7fff46228ea0 items=0 ppid=3940 pid=4108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.078000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 00:47:55.078000 audit: BPF prog-id=211 op=UNLOAD Jan 14 00:47:55.078000 audit[4108]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7fff46228d20 a2=0 a3=7fff46228ea0 items=0 ppid=3940 pid=4108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.078000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 00:47:55.079000 audit: BPF prog-id=212 op=LOAD Jan 14 00:47:55.079000 audit[4108]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fff46228450 a2=94 a3=2 items=0 ppid=3940 pid=4108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.079000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 00:47:55.079000 audit: BPF prog-id=212 op=UNLOAD Jan 14 00:47:55.079000 audit[4108]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7fff46228450 a2=0 a3=2 items=0 ppid=3940 pid=4108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.079000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 00:47:55.079000 audit: BPF prog-id=213 op=LOAD Jan 14 00:47:55.079000 audit[4108]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fff46228550 a2=94 a3=30 items=0 ppid=3940 pid=4108 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.079000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Jan 14 00:47:55.094000 audit: BPF prog-id=214 op=LOAD Jan 14 00:47:55.094000 audit[4112]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7fff7b9be080 a2=98 a3=0 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.094000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.094000 audit: BPF prog-id=214 op=UNLOAD Jan 14 00:47:55.094000 audit[4112]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=3 a1=8 a2=7fff7b9be050 a3=0 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.094000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.094000 audit: BPF prog-id=215 op=LOAD Jan 14 00:47:55.094000 audit[4112]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fff7b9bde70 a2=94 a3=54428f items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.094000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.094000 audit: BPF prog-id=215 op=UNLOAD Jan 14 00:47:55.094000 audit[4112]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7fff7b9bde70 a2=94 a3=54428f items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.094000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.094000 audit: BPF prog-id=216 op=LOAD Jan 14 00:47:55.094000 audit[4112]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fff7b9bdea0 a2=94 a3=2 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.094000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.094000 audit: BPF prog-id=216 op=UNLOAD Jan 14 00:47:55.094000 audit[4112]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7fff7b9bdea0 a2=0 a3=2 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.094000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.273000 audit: BPF prog-id=217 op=LOAD Jan 14 00:47:55.273000 audit[4112]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fff7b9bdd60 a2=94 a3=1 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.273000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.273000 audit: BPF prog-id=217 op=UNLOAD Jan 14 00:47:55.273000 audit[4112]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=4 a1=7fff7b9bdd60 a2=94 a3=1 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.273000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.282000 audit: BPF prog-id=218 op=LOAD Jan 14 00:47:55.282000 audit[4112]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7fff7b9bdd50 a2=94 a3=4 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.282000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.282000 audit: BPF prog-id=218 op=UNLOAD Jan 14 00:47:55.282000 audit[4112]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7fff7b9bdd50 a2=0 a3=4 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.282000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.283000 audit: BPF prog-id=219 op=LOAD Jan 14 00:47:55.283000 audit[4112]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fff7b9bdbb0 a2=94 a3=5 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.283000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.283000 audit: BPF prog-id=219 op=UNLOAD Jan 14 00:47:55.283000 audit[4112]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=6 a1=7fff7b9bdbb0 a2=0 a3=5 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.283000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.283000 audit: BPF prog-id=220 op=LOAD Jan 14 00:47:55.283000 audit[4112]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7fff7b9bddd0 a2=94 a3=6 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.283000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.283000 audit: BPF prog-id=220 op=UNLOAD Jan 14 00:47:55.283000 audit[4112]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=7fff7b9bddd0 a2=0 a3=6 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.283000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.283000 audit: BPF prog-id=221 op=LOAD Jan 14 00:47:55.283000 audit[4112]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=5 a1=7fff7b9bd580 a2=94 a3=88 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.283000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.283000 audit: BPF prog-id=222 op=LOAD Jan 14 00:47:55.283000 audit[4112]: SYSCALL arch=c000003e syscall=321 success=yes exit=7 a0=5 a1=7fff7b9bd400 a2=94 a3=2 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.283000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.283000 audit: BPF prog-id=222 op=UNLOAD Jan 14 00:47:55.283000 audit[4112]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=7 a1=7fff7b9bd430 a2=0 a3=7fff7b9bd530 items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.283000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.284000 audit: BPF prog-id=221 op=UNLOAD Jan 14 00:47:55.284000 audit[4112]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=5 a1=2ec05d10 a2=0 a3=12238f49401a04f items=0 ppid=3940 pid=4112 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.284000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Jan 14 00:47:55.290000 audit: BPF prog-id=213 op=UNLOAD Jan 14 00:47:55.290000 audit[3940]: SYSCALL arch=c000003e syscall=263 success=yes exit=0 a0=ffffffffffffff9c a1=c0009481c0 a2=0 a3=0 items=0 ppid=3910 pid=3940 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="calico-node" exe="/usr/bin/calico-node" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.290000 audit: PROCTITLE proctitle=63616C69636F2D6E6F6465002D66656C6978 Jan 14 00:47:55.351000 audit[4135]: NETFILTER_CFG table=raw:119 family=2 entries=21 op=nft_register_chain pid=4135 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 00:47:55.351000 audit[4135]: SYSCALL arch=c000003e syscall=46 success=yes exit=8452 a0=3 a1=7ffdfb6f7f20 a2=0 a3=7ffdfb6f7f0c items=0 ppid=3940 pid=4135 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.351000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 00:47:55.357000 audit[4137]: NETFILTER_CFG table=nat:120 family=2 entries=15 op=nft_register_chain pid=4137 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 00:47:55.357000 audit[4137]: SYSCALL arch=c000003e syscall=46 success=yes exit=5084 a0=3 a1=7ffec4d40cc0 a2=0 a3=7ffec4d40cac items=0 ppid=3940 pid=4137 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.357000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 00:47:55.361000 audit[4141]: NETFILTER_CFG table=mangle:121 family=2 entries=16 op=nft_register_chain pid=4141 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 00:47:55.361000 audit[4141]: SYSCALL arch=c000003e syscall=46 success=yes exit=6868 a0=3 a1=7fff66684fd0 a2=0 a3=7fff66684fbc items=0 ppid=3940 pid=4141 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.361000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 00:47:55.365000 audit[4140]: NETFILTER_CFG table=filter:122 family=2 entries=94 op=nft_register_chain pid=4140 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 00:47:55.365000 audit[4140]: SYSCALL arch=c000003e syscall=46 success=yes exit=53116 a0=3 a1=7fffe316b290 a2=0 a3=7fffe316b27c items=0 ppid=3940 pid=4140 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.365000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 00:47:55.502489 kubelet[2802]: I0114 00:47:55.502431 2802 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a556fce5-e24b-4167-85b9-cd5dc0a668fe" path="/var/lib/kubelet/pods/a556fce5-e24b-4167-85b9-cd5dc0a668fe/volumes" Jan 14 00:47:55.709814 kubelet[2802]: E0114 00:47:55.709325 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-d6f8b64b4-m4ctk" podUID="db919709-47a0-46a4-a2ab-3f096374d888" Jan 14 00:47:55.737000 audit[4150]: NETFILTER_CFG table=filter:123 family=2 entries=20 op=nft_register_rule pid=4150 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:55.737000 audit[4150]: SYSCALL arch=c000003e syscall=46 success=yes exit=7480 a0=3 a1=7ffde064dda0 a2=0 a3=7ffde064dd8c items=0 ppid=2943 pid=4150 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.737000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:55.739000 audit[4150]: NETFILTER_CFG table=nat:124 family=2 entries=14 op=nft_register_rule pid=4150 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:47:55.739000 audit[4150]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffde064dda0 a2=0 a3=0 items=0 ppid=2943 pid=4150 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:47:55.739000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:47:56.117364 systemd-networkd[1504]: calif721a9bde91: Gained IPv6LL Jan 14 00:47:56.437306 systemd-networkd[1504]: vxlan.calico: Gained IPv6LL Jan 14 00:48:00.500784 kubelet[2802]: E0114 00:48:00.500096 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:00.501577 containerd[1611]: time="2026-01-14T00:48:00.501387377Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hc5jt,Uid:43d5cf61-6f70-467e-abb2-9b166aef5949,Namespace:kube-system,Attempt:0,}" Jan 14 00:48:00.502478 containerd[1611]: time="2026-01-14T00:48:00.502457265Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-vzqkb,Uid:f41cc627-bd07-42f1-8e88-f85dcd2c063c,Namespace:calico-system,Attempt:0,}" Jan 14 00:48:00.690158 systemd-networkd[1504]: cali2165416c862: Link UP Jan 14 00:48:00.691823 systemd-networkd[1504]: cali2165416c862: Gained carrier Jan 14 00:48:00.747390 containerd[1611]: 2026-01-14 00:48:00.561 [INFO][4163] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--236--110--177-k8s-coredns--674b8bbfcf--hc5jt-eth0 coredns-674b8bbfcf- kube-system 43d5cf61-6f70-467e-abb2-9b166aef5949 880 0 2026-01-14 00:47:28 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s 172-236-110-177 coredns-674b8bbfcf-hc5jt eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali2165416c862 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" Namespace="kube-system" Pod="coredns-674b8bbfcf-hc5jt" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--hc5jt-" Jan 14 00:48:00.747390 containerd[1611]: 2026-01-14 00:48:00.561 [INFO][4163] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" Namespace="kube-system" Pod="coredns-674b8bbfcf-hc5jt" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--hc5jt-eth0" Jan 14 00:48:00.747390 containerd[1611]: 2026-01-14 00:48:00.601 [INFO][4188] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" HandleID="k8s-pod-network.952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" Workload="172--236--110--177-k8s-coredns--674b8bbfcf--hc5jt-eth0" Jan 14 00:48:00.749480 containerd[1611]: 2026-01-14 00:48:00.602 [INFO][4188] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" HandleID="k8s-pod-network.952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" Workload="172--236--110--177-k8s-coredns--674b8bbfcf--hc5jt-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00024fbb0), Attrs:map[string]string{"namespace":"kube-system", "node":"172-236-110-177", "pod":"coredns-674b8bbfcf-hc5jt", "timestamp":"2026-01-14 00:48:00.601958342 +0000 UTC"}, Hostname:"172-236-110-177", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 00:48:00.749480 containerd[1611]: 2026-01-14 00:48:00.602 [INFO][4188] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 00:48:00.749480 containerd[1611]: 2026-01-14 00:48:00.602 [INFO][4188] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 00:48:00.749480 containerd[1611]: 2026-01-14 00:48:00.602 [INFO][4188] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-236-110-177' Jan 14 00:48:00.749480 containerd[1611]: 2026-01-14 00:48:00.608 [INFO][4188] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" host="172-236-110-177" Jan 14 00:48:00.749480 containerd[1611]: 2026-01-14 00:48:00.614 [INFO][4188] ipam/ipam.go 394: Looking up existing affinities for host host="172-236-110-177" Jan 14 00:48:00.749480 containerd[1611]: 2026-01-14 00:48:00.618 [INFO][4188] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:00.749480 containerd[1611]: 2026-01-14 00:48:00.620 [INFO][4188] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:00.749480 containerd[1611]: 2026-01-14 00:48:00.622 [INFO][4188] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:00.749713 containerd[1611]: 2026-01-14 00:48:00.622 [INFO][4188] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" host="172-236-110-177" Jan 14 00:48:00.749713 containerd[1611]: 2026-01-14 00:48:00.623 [INFO][4188] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a Jan 14 00:48:00.749713 containerd[1611]: 2026-01-14 00:48:00.675 [INFO][4188] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" host="172-236-110-177" Jan 14 00:48:00.749713 containerd[1611]: 2026-01-14 00:48:00.679 [INFO][4188] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.194/26] block=192.168.109.192/26 handle="k8s-pod-network.952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" host="172-236-110-177" Jan 14 00:48:00.749713 containerd[1611]: 2026-01-14 00:48:00.679 [INFO][4188] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.194/26] handle="k8s-pod-network.952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" host="172-236-110-177" Jan 14 00:48:00.749713 containerd[1611]: 2026-01-14 00:48:00.679 [INFO][4188] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 00:48:00.749713 containerd[1611]: 2026-01-14 00:48:00.679 [INFO][4188] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.194/26] IPv6=[] ContainerID="952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" HandleID="k8s-pod-network.952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" Workload="172--236--110--177-k8s-coredns--674b8bbfcf--hc5jt-eth0" Jan 14 00:48:00.749857 containerd[1611]: 2026-01-14 00:48:00.686 [INFO][4163] cni-plugin/k8s.go 418: Populated endpoint ContainerID="952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" Namespace="kube-system" Pod="coredns-674b8bbfcf-hc5jt" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--hc5jt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-coredns--674b8bbfcf--hc5jt-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"43d5cf61-6f70-467e-abb2-9b166aef5949", ResourceVersion:"880", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"", Pod:"coredns-674b8bbfcf-hc5jt", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.109.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2165416c862", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:48:00.749857 containerd[1611]: 2026-01-14 00:48:00.686 [INFO][4163] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.194/32] ContainerID="952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" Namespace="kube-system" Pod="coredns-674b8bbfcf-hc5jt" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--hc5jt-eth0" Jan 14 00:48:00.749857 containerd[1611]: 2026-01-14 00:48:00.686 [INFO][4163] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2165416c862 ContainerID="952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" Namespace="kube-system" Pod="coredns-674b8bbfcf-hc5jt" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--hc5jt-eth0" Jan 14 00:48:00.749857 containerd[1611]: 2026-01-14 00:48:00.693 [INFO][4163] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" Namespace="kube-system" Pod="coredns-674b8bbfcf-hc5jt" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--hc5jt-eth0" Jan 14 00:48:00.749857 containerd[1611]: 2026-01-14 00:48:00.695 [INFO][4163] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" Namespace="kube-system" Pod="coredns-674b8bbfcf-hc5jt" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--hc5jt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-coredns--674b8bbfcf--hc5jt-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"43d5cf61-6f70-467e-abb2-9b166aef5949", ResourceVersion:"880", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a", Pod:"coredns-674b8bbfcf-hc5jt", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.109.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2165416c862", MAC:"52:6b:e9:ae:7e:e5", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:48:00.749857 containerd[1611]: 2026-01-14 00:48:00.742 [INFO][4163] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" Namespace="kube-system" Pod="coredns-674b8bbfcf-hc5jt" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--hc5jt-eth0" Jan 14 00:48:00.782244 containerd[1611]: time="2026-01-14T00:48:00.781442422Z" level=info msg="connecting to shim 952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a" address="unix:///run/containerd/s/a2a66cc0901e1242b5264456d2d319df69b5c85dc15e0700bd85ffecaa68e106" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:48:00.800931 kernel: kauditd_printk_skb: 231 callbacks suppressed Jan 14 00:48:00.801254 kernel: audit: type=1325 audit(1768351680.788:667): table=filter:125 family=2 entries=42 op=nft_register_chain pid=4223 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 00:48:00.788000 audit[4223]: NETFILTER_CFG table=filter:125 family=2 entries=42 op=nft_register_chain pid=4223 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 00:48:00.788000 audit[4223]: SYSCALL arch=c000003e syscall=46 success=yes exit=22552 a0=3 a1=7ffde147d3e0 a2=0 a3=7ffde147d3cc items=0 ppid=3940 pid=4223 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:00.818138 kernel: audit: type=1300 audit(1768351680.788:667): arch=c000003e syscall=46 success=yes exit=22552 a0=3 a1=7ffde147d3e0 a2=0 a3=7ffde147d3cc items=0 ppid=3940 pid=4223 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:00.818195 kernel: audit: type=1327 audit(1768351680.788:667): proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 00:48:00.788000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 00:48:00.832289 systemd[1]: Started cri-containerd-952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a.scope - libcontainer container 952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a. Jan 14 00:48:00.850000 audit: BPF prog-id=223 op=LOAD Jan 14 00:48:00.855465 kernel: audit: type=1334 audit(1768351680.850:668): prog-id=223 op=LOAD Jan 14 00:48:00.851000 audit: BPF prog-id=224 op=LOAD Jan 14 00:48:00.858154 systemd-networkd[1504]: cali68bf38c6acd: Link UP Jan 14 00:48:00.860049 kernel: audit: type=1334 audit(1768351680.851:669): prog-id=224 op=LOAD Jan 14 00:48:00.851000 audit[4234]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000228238 a2=98 a3=0 items=0 ppid=4222 pid=4234 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:00.861690 systemd-networkd[1504]: cali68bf38c6acd: Gained carrier Jan 14 00:48:00.869139 kernel: audit: type=1300 audit(1768351680.851:669): arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000228238 a2=98 a3=0 items=0 ppid=4222 pid=4234 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:00.851000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3935326464336238313763393635376665323262396538316365393864 Jan 14 00:48:00.854000 audit: BPF prog-id=224 op=UNLOAD Jan 14 00:48:00.905560 kernel: audit: type=1327 audit(1768351680.851:669): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3935326464336238313763393635376665323262396538316365393864 Jan 14 00:48:00.905603 kernel: audit: type=1334 audit(1768351680.854:670): prog-id=224 op=UNLOAD Jan 14 00:48:00.854000 audit[4234]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4222 pid=4234 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:00.916302 kernel: audit: type=1300 audit(1768351680.854:670): arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4222 pid=4234 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:00.854000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3935326464336238313763393635376665323262396538316365393864 Jan 14 00:48:00.857000 audit: BPF prog-id=225 op=LOAD Jan 14 00:48:00.857000 audit[4234]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000228488 a2=98 a3=0 items=0 ppid=4222 pid=4234 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:00.857000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3935326464336238313763393635376665323262396538316365393864 Jan 14 00:48:00.929144 kernel: audit: type=1327 audit(1768351680.854:670): proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3935326464336238313763393635376665323262396538316365393864 Jan 14 00:48:00.857000 audit: BPF prog-id=226 op=LOAD Jan 14 00:48:00.857000 audit[4234]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000228218 a2=98 a3=0 items=0 ppid=4222 pid=4234 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:00.857000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3935326464336238313763393635376665323262396538316365393864 Jan 14 00:48:00.857000 audit: BPF prog-id=226 op=UNLOAD Jan 14 00:48:00.857000 audit[4234]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=4222 pid=4234 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:00.857000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3935326464336238313763393635376665323262396538316365393864 Jan 14 00:48:00.858000 audit: BPF prog-id=225 op=UNLOAD Jan 14 00:48:00.858000 audit[4234]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4222 pid=4234 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:00.858000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3935326464336238313763393635376665323262396538316365393864 Jan 14 00:48:00.858000 audit: BPF prog-id=227 op=LOAD Jan 14 00:48:00.858000 audit[4234]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0002286e8 a2=98 a3=0 items=0 ppid=4222 pid=4234 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:00.858000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3935326464336238313763393635376665323262396538316365393864 Jan 14 00:48:00.935951 containerd[1611]: time="2026-01-14T00:48:00.935862171Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-hc5jt,Uid:43d5cf61-6f70-467e-abb2-9b166aef5949,Namespace:kube-system,Attempt:0,} returns sandbox id \"952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a\"" Jan 14 00:48:00.947127 kubelet[2802]: E0114 00:48:00.947026 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.561 [INFO][4165] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--236--110--177-k8s-csi--node--driver--vzqkb-eth0 csi-node-driver- calico-system f41cc627-bd07-42f1-8e88-f85dcd2c063c 770 0 2026-01-14 00:47:42 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:857b56db8f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s 172-236-110-177 csi-node-driver-vzqkb eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali68bf38c6acd [] [] }} ContainerID="00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" Namespace="calico-system" Pod="csi-node-driver-vzqkb" WorkloadEndpoint="172--236--110--177-k8s-csi--node--driver--vzqkb-" Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.561 [INFO][4165] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" Namespace="calico-system" Pod="csi-node-driver-vzqkb" WorkloadEndpoint="172--236--110--177-k8s-csi--node--driver--vzqkb-eth0" Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.605 [INFO][4189] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" HandleID="k8s-pod-network.00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" Workload="172--236--110--177-k8s-csi--node--driver--vzqkb-eth0" Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.605 [INFO][4189] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" HandleID="k8s-pod-network.00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" Workload="172--236--110--177-k8s-csi--node--driver--vzqkb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00035cd70), Attrs:map[string]string{"namespace":"calico-system", "node":"172-236-110-177", "pod":"csi-node-driver-vzqkb", "timestamp":"2026-01-14 00:48:00.605536059 +0000 UTC"}, Hostname:"172-236-110-177", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.605 [INFO][4189] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.679 [INFO][4189] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.679 [INFO][4189] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-236-110-177' Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.709 [INFO][4189] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" host="172-236-110-177" Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.752 [INFO][4189] ipam/ipam.go 394: Looking up existing affinities for host host="172-236-110-177" Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.761 [INFO][4189] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.764 [INFO][4189] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.768 [INFO][4189] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.768 [INFO][4189] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" host="172-236-110-177" Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.801 [INFO][4189] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479 Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.841 [INFO][4189] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" host="172-236-110-177" Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.847 [INFO][4189] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.195/26] block=192.168.109.192/26 handle="k8s-pod-network.00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" host="172-236-110-177" Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.847 [INFO][4189] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.195/26] handle="k8s-pod-network.00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" host="172-236-110-177" Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.847 [INFO][4189] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 00:48:00.949137 containerd[1611]: 2026-01-14 00:48:00.847 [INFO][4189] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.195/26] IPv6=[] ContainerID="00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" HandleID="k8s-pod-network.00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" Workload="172--236--110--177-k8s-csi--node--driver--vzqkb-eth0" Jan 14 00:48:00.949656 containerd[1611]: 2026-01-14 00:48:00.850 [INFO][4165] cni-plugin/k8s.go 418: Populated endpoint ContainerID="00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" Namespace="calico-system" Pod="csi-node-driver-vzqkb" WorkloadEndpoint="172--236--110--177-k8s-csi--node--driver--vzqkb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-csi--node--driver--vzqkb-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"f41cc627-bd07-42f1-8e88-f85dcd2c063c", ResourceVersion:"770", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 42, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"", Pod:"csi-node-driver-vzqkb", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.109.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali68bf38c6acd", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:48:00.949656 containerd[1611]: 2026-01-14 00:48:00.850 [INFO][4165] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.195/32] ContainerID="00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" Namespace="calico-system" Pod="csi-node-driver-vzqkb" WorkloadEndpoint="172--236--110--177-k8s-csi--node--driver--vzqkb-eth0" Jan 14 00:48:00.949656 containerd[1611]: 2026-01-14 00:48:00.850 [INFO][4165] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali68bf38c6acd ContainerID="00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" Namespace="calico-system" Pod="csi-node-driver-vzqkb" WorkloadEndpoint="172--236--110--177-k8s-csi--node--driver--vzqkb-eth0" Jan 14 00:48:00.949656 containerd[1611]: 2026-01-14 00:48:00.875 [INFO][4165] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" Namespace="calico-system" Pod="csi-node-driver-vzqkb" WorkloadEndpoint="172--236--110--177-k8s-csi--node--driver--vzqkb-eth0" Jan 14 00:48:00.949656 containerd[1611]: 2026-01-14 00:48:00.876 [INFO][4165] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" Namespace="calico-system" Pod="csi-node-driver-vzqkb" WorkloadEndpoint="172--236--110--177-k8s-csi--node--driver--vzqkb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-csi--node--driver--vzqkb-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"f41cc627-bd07-42f1-8e88-f85dcd2c063c", ResourceVersion:"770", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 42, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"857b56db8f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479", Pod:"csi-node-driver-vzqkb", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.109.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali68bf38c6acd", MAC:"1a:2a:a4:8f:b3:08", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:48:00.949656 containerd[1611]: 2026-01-14 00:48:00.935 [INFO][4165] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" Namespace="calico-system" Pod="csi-node-driver-vzqkb" WorkloadEndpoint="172--236--110--177-k8s-csi--node--driver--vzqkb-eth0" Jan 14 00:48:00.954683 containerd[1611]: time="2026-01-14T00:48:00.954649640Z" level=info msg="CreateContainer within sandbox \"952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 14 00:48:00.979000 audit[4267]: NETFILTER_CFG table=filter:126 family=2 entries=40 op=nft_register_chain pid=4267 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 00:48:00.979000 audit[4267]: SYSCALL arch=c000003e syscall=46 success=yes exit=20764 a0=3 a1=7fffade28dd0 a2=0 a3=7fffade28dbc items=0 ppid=3940 pid=4267 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:00.979000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 00:48:00.981402 containerd[1611]: time="2026-01-14T00:48:00.978873440Z" level=info msg="Container e1e6b6b726aea301f40ed086086e36f24a90919c85131e093f4d62f6b9e52ca4: CDI devices from CRI Config.CDIDevices: []" Jan 14 00:48:00.986712 containerd[1611]: time="2026-01-14T00:48:00.986671858Z" level=info msg="CreateContainer within sandbox \"952dd3b817c9657fe22b9e81ce98d81d559220950bc80f3f3209abfa0108812a\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"e1e6b6b726aea301f40ed086086e36f24a90919c85131e093f4d62f6b9e52ca4\"" Jan 14 00:48:00.990350 containerd[1611]: time="2026-01-14T00:48:00.990325408Z" level=info msg="StartContainer for \"e1e6b6b726aea301f40ed086086e36f24a90919c85131e093f4d62f6b9e52ca4\"" Jan 14 00:48:00.991732 containerd[1611]: time="2026-01-14T00:48:00.991240210Z" level=info msg="connecting to shim e1e6b6b726aea301f40ed086086e36f24a90919c85131e093f4d62f6b9e52ca4" address="unix:///run/containerd/s/a2a66cc0901e1242b5264456d2d319df69b5c85dc15e0700bd85ffecaa68e106" protocol=ttrpc version=3 Jan 14 00:48:00.999898 containerd[1611]: time="2026-01-14T00:48:00.999868536Z" level=info msg="connecting to shim 00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479" address="unix:///run/containerd/s/c8df80c4346181dd1dcc2c547036498e19156cabc27ddc4bec08b366583257b5" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:48:01.020325 systemd[1]: Started cri-containerd-e1e6b6b726aea301f40ed086086e36f24a90919c85131e093f4d62f6b9e52ca4.scope - libcontainer container e1e6b6b726aea301f40ed086086e36f24a90919c85131e093f4d62f6b9e52ca4. Jan 14 00:48:01.045248 systemd[1]: Started cri-containerd-00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479.scope - libcontainer container 00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479. Jan 14 00:48:01.048000 audit: BPF prog-id=228 op=LOAD Jan 14 00:48:01.048000 audit: BPF prog-id=229 op=LOAD Jan 14 00:48:01.048000 audit[4272]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000220238 a2=98 a3=0 items=0 ppid=4222 pid=4272 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.048000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531653662366237323661656133303166343065643038363038366533 Jan 14 00:48:01.048000 audit: BPF prog-id=229 op=UNLOAD Jan 14 00:48:01.048000 audit[4272]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4222 pid=4272 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.048000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531653662366237323661656133303166343065643038363038366533 Jan 14 00:48:01.048000 audit: BPF prog-id=230 op=LOAD Jan 14 00:48:01.048000 audit[4272]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000220488 a2=98 a3=0 items=0 ppid=4222 pid=4272 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.048000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531653662366237323661656133303166343065643038363038366533 Jan 14 00:48:01.048000 audit: BPF prog-id=231 op=LOAD Jan 14 00:48:01.048000 audit[4272]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000220218 a2=98 a3=0 items=0 ppid=4222 pid=4272 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.048000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531653662366237323661656133303166343065643038363038366533 Jan 14 00:48:01.049000 audit: BPF prog-id=231 op=UNLOAD Jan 14 00:48:01.049000 audit[4272]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4222 pid=4272 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.049000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531653662366237323661656133303166343065643038363038366533 Jan 14 00:48:01.049000 audit: BPF prog-id=230 op=UNLOAD Jan 14 00:48:01.049000 audit[4272]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4222 pid=4272 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.049000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531653662366237323661656133303166343065643038363038366533 Jan 14 00:48:01.049000 audit: BPF prog-id=232 op=LOAD Jan 14 00:48:01.049000 audit[4272]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0002206e8 a2=98 a3=0 items=0 ppid=4222 pid=4272 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.049000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6531653662366237323661656133303166343065643038363038366533 Jan 14 00:48:01.071000 audit: BPF prog-id=233 op=LOAD Jan 14 00:48:01.071000 audit: BPF prog-id=234 op=LOAD Jan 14 00:48:01.071000 audit[4300]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=4279 pid=4300 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.071000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3030643339336166366138613232393038386634623137323461666530 Jan 14 00:48:01.074000 audit: BPF prog-id=234 op=UNLOAD Jan 14 00:48:01.074000 audit[4300]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4279 pid=4300 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.074000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3030643339336166366138613232393038386634623137323461666530 Jan 14 00:48:01.074000 audit: BPF prog-id=235 op=LOAD Jan 14 00:48:01.074000 audit[4300]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=4279 pid=4300 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.074000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3030643339336166366138613232393038386634623137323461666530 Jan 14 00:48:01.074000 audit: BPF prog-id=236 op=LOAD Jan 14 00:48:01.074000 audit[4300]: SYSCALL arch=c000003e syscall=321 success=yes exit=22 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=4279 pid=4300 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.074000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3030643339336166366138613232393038386634623137323461666530 Jan 14 00:48:01.075000 audit: BPF prog-id=236 op=UNLOAD Jan 14 00:48:01.075000 audit[4300]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=16 a1=0 a2=0 a3=0 items=0 ppid=4279 pid=4300 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.075000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3030643339336166366138613232393038386634623137323461666530 Jan 14 00:48:01.075000 audit: BPF prog-id=235 op=UNLOAD Jan 14 00:48:01.075000 audit[4300]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=14 a1=0 a2=0 a3=0 items=0 ppid=4279 pid=4300 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.075000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3030643339336166366138613232393038386634623137323461666530 Jan 14 00:48:01.075000 audit: BPF prog-id=237 op=LOAD Jan 14 00:48:01.075000 audit[4300]: SYSCALL arch=c000003e syscall=321 success=yes exit=20 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=4279 pid=4300 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.075000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3030643339336166366138613232393038386634623137323461666530 Jan 14 00:48:01.079066 containerd[1611]: time="2026-01-14T00:48:01.078636461Z" level=info msg="StartContainer for \"e1e6b6b726aea301f40ed086086e36f24a90919c85131e093f4d62f6b9e52ca4\" returns successfully" Jan 14 00:48:01.098201 containerd[1611]: time="2026-01-14T00:48:01.098150092Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-vzqkb,Uid:f41cc627-bd07-42f1-8e88-f85dcd2c063c,Namespace:calico-system,Attempt:0,} returns sandbox id \"00d393af6a8a229088f4b1724afe0f7ec6633671c60920e579397a5ef1d06479\"" Jan 14 00:48:01.100873 containerd[1611]: time="2026-01-14T00:48:01.100776814Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 14 00:48:01.228939 containerd[1611]: time="2026-01-14T00:48:01.228866413Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:01.230174 containerd[1611]: time="2026-01-14T00:48:01.230104596Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 14 00:48:01.230232 containerd[1611]: time="2026-01-14T00:48:01.230135577Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:01.230550 kubelet[2802]: E0114 00:48:01.230323 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 14 00:48:01.230550 kubelet[2802]: E0114 00:48:01.230378 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 14 00:48:01.231683 kubelet[2802]: E0114 00:48:01.231606 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pnt5f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-vzqkb_calico-system(f41cc627-bd07-42f1-8e88-f85dcd2c063c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:01.233362 containerd[1611]: time="2026-01-14T00:48:01.233341528Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 14 00:48:01.360976 containerd[1611]: time="2026-01-14T00:48:01.360154435Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:01.361249 containerd[1611]: time="2026-01-14T00:48:01.361175380Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 14 00:48:01.361346 containerd[1611]: time="2026-01-14T00:48:01.361217451Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:01.361550 kubelet[2802]: E0114 00:48:01.361496 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 14 00:48:01.361550 kubelet[2802]: E0114 00:48:01.361539 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 14 00:48:01.361718 kubelet[2802]: E0114 00:48:01.361641 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pnt5f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-vzqkb_calico-system(f41cc627-bd07-42f1-8e88-f85dcd2c063c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:01.362901 kubelet[2802]: E0114 00:48:01.362808 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:48:01.514643 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3118468118.mount: Deactivated successfully. Jan 14 00:48:01.721483 kubelet[2802]: E0114 00:48:01.721412 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:48:01.723169 kubelet[2802]: E0114 00:48:01.723155 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:01.754104 kubelet[2802]: I0114 00:48:01.753959 2802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-hc5jt" podStartSLOduration=33.753912464 podStartE2EDuration="33.753912464s" podCreationTimestamp="2026-01-14 00:47:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-14 00:48:01.752201155 +0000 UTC m=+40.347297569" watchObservedRunningTime="2026-01-14 00:48:01.753912464 +0000 UTC m=+40.349008878" Jan 14 00:48:01.779000 audit[4348]: NETFILTER_CFG table=filter:127 family=2 entries=17 op=nft_register_rule pid=4348 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:48:01.779000 audit[4348]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffe4b5cc070 a2=0 a3=7ffe4b5cc05c items=0 ppid=2943 pid=4348 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.779000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:48:01.782000 audit[4348]: NETFILTER_CFG table=nat:128 family=2 entries=35 op=nft_register_chain pid=4348 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:48:01.782000 audit[4348]: SYSCALL arch=c000003e syscall=46 success=yes exit=14196 a0=3 a1=7ffe4b5cc070 a2=0 a3=7ffe4b5cc05c items=0 ppid=2943 pid=4348 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:01.782000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:48:02.389277 systemd-networkd[1504]: cali2165416c862: Gained IPv6LL Jan 14 00:48:02.389709 systemd-networkd[1504]: cali68bf38c6acd: Gained IPv6LL Jan 14 00:48:02.499724 containerd[1611]: time="2026-01-14T00:48:02.499655893Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5f9b965548-bgkj8,Uid:dcaab5f4-fd68-41e8-986b-14fa1a168bce,Namespace:calico-apiserver,Attempt:0,}" Jan 14 00:48:02.500300 containerd[1611]: time="2026-01-14T00:48:02.500274644Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5f9b965548-xv9gd,Uid:32668954-0a97-46fe-9f01-fa91a530c2e3,Namespace:calico-apiserver,Attempt:0,}" Jan 14 00:48:02.624444 systemd-networkd[1504]: cali7dd3c40bc1e: Link UP Jan 14 00:48:02.625221 systemd-networkd[1504]: cali7dd3c40bc1e: Gained carrier Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.561 [INFO][4352] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--236--110--177-k8s-calico--apiserver--5f9b965548--xv9gd-eth0 calico-apiserver-5f9b965548- calico-apiserver 32668954-0a97-46fe-9f01-fa91a530c2e3 877 0 2026-01-14 00:47:37 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:5f9b965548 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s 172-236-110-177 calico-apiserver-5f9b965548-xv9gd eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali7dd3c40bc1e [] [] }} ContainerID="977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-xv9gd" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--xv9gd-" Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.562 [INFO][4352] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-xv9gd" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--xv9gd-eth0" Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.589 [INFO][4375] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" HandleID="k8s-pod-network.977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" Workload="172--236--110--177-k8s-calico--apiserver--5f9b965548--xv9gd-eth0" Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.589 [INFO][4375] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" HandleID="k8s-pod-network.977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" Workload="172--236--110--177-k8s-calico--apiserver--5f9b965548--xv9gd-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00032d3b0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"172-236-110-177", "pod":"calico-apiserver-5f9b965548-xv9gd", "timestamp":"2026-01-14 00:48:02.589630128 +0000 UTC"}, Hostname:"172-236-110-177", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.589 [INFO][4375] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.589 [INFO][4375] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.590 [INFO][4375] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-236-110-177' Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.595 [INFO][4375] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" host="172-236-110-177" Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.599 [INFO][4375] ipam/ipam.go 394: Looking up existing affinities for host host="172-236-110-177" Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.603 [INFO][4375] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.604 [INFO][4375] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.606 [INFO][4375] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.606 [INFO][4375] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" host="172-236-110-177" Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.607 [INFO][4375] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.611 [INFO][4375] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" host="172-236-110-177" Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.616 [INFO][4375] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.196/26] block=192.168.109.192/26 handle="k8s-pod-network.977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" host="172-236-110-177" Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.616 [INFO][4375] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.196/26] handle="k8s-pod-network.977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" host="172-236-110-177" Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.617 [INFO][4375] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 00:48:02.638257 containerd[1611]: 2026-01-14 00:48:02.617 [INFO][4375] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.196/26] IPv6=[] ContainerID="977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" HandleID="k8s-pod-network.977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" Workload="172--236--110--177-k8s-calico--apiserver--5f9b965548--xv9gd-eth0" Jan 14 00:48:02.639397 containerd[1611]: 2026-01-14 00:48:02.619 [INFO][4352] cni-plugin/k8s.go 418: Populated endpoint ContainerID="977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-xv9gd" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--xv9gd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-calico--apiserver--5f9b965548--xv9gd-eth0", GenerateName:"calico-apiserver-5f9b965548-", Namespace:"calico-apiserver", SelfLink:"", UID:"32668954-0a97-46fe-9f01-fa91a530c2e3", ResourceVersion:"877", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5f9b965548", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"", Pod:"calico-apiserver-5f9b965548-xv9gd", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.109.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali7dd3c40bc1e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:48:02.639397 containerd[1611]: 2026-01-14 00:48:02.619 [INFO][4352] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.196/32] ContainerID="977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-xv9gd" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--xv9gd-eth0" Jan 14 00:48:02.639397 containerd[1611]: 2026-01-14 00:48:02.619 [INFO][4352] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali7dd3c40bc1e ContainerID="977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-xv9gd" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--xv9gd-eth0" Jan 14 00:48:02.639397 containerd[1611]: 2026-01-14 00:48:02.624 [INFO][4352] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-xv9gd" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--xv9gd-eth0" Jan 14 00:48:02.639397 containerd[1611]: 2026-01-14 00:48:02.626 [INFO][4352] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-xv9gd" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--xv9gd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-calico--apiserver--5f9b965548--xv9gd-eth0", GenerateName:"calico-apiserver-5f9b965548-", Namespace:"calico-apiserver", SelfLink:"", UID:"32668954-0a97-46fe-9f01-fa91a530c2e3", ResourceVersion:"877", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5f9b965548", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb", Pod:"calico-apiserver-5f9b965548-xv9gd", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.109.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali7dd3c40bc1e", MAC:"56:0e:2d:b9:ec:d4", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:48:02.639397 containerd[1611]: 2026-01-14 00:48:02.635 [INFO][4352] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-xv9gd" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--xv9gd-eth0" Jan 14 00:48:02.657000 audit[4396]: NETFILTER_CFG table=filter:129 family=2 entries=58 op=nft_register_chain pid=4396 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 00:48:02.657000 audit[4396]: SYSCALL arch=c000003e syscall=46 success=yes exit=30584 a0=3 a1=7ffeed349160 a2=0 a3=7ffeed34914c items=0 ppid=3940 pid=4396 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.657000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 00:48:02.669132 containerd[1611]: time="2026-01-14T00:48:02.667947492Z" level=info msg="connecting to shim 977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb" address="unix:///run/containerd/s/1a46363cbf2fb7aa24ba93356d36edd4b5d0432912a13bdd22d5d30308168652" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:48:02.698266 systemd[1]: Started cri-containerd-977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb.scope - libcontainer container 977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb. Jan 14 00:48:02.712000 audit: BPF prog-id=238 op=LOAD Jan 14 00:48:02.712000 audit: BPF prog-id=239 op=LOAD Jan 14 00:48:02.712000 audit[4417]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=4406 pid=4417 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.712000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3937376332663937623065623232323836386362633564346231313866 Jan 14 00:48:02.713000 audit: BPF prog-id=239 op=UNLOAD Jan 14 00:48:02.713000 audit[4417]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4406 pid=4417 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.713000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3937376332663937623065623232323836386362633564346231313866 Jan 14 00:48:02.713000 audit: BPF prog-id=240 op=LOAD Jan 14 00:48:02.713000 audit[4417]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=4406 pid=4417 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.713000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3937376332663937623065623232323836386362633564346231313866 Jan 14 00:48:02.713000 audit: BPF prog-id=241 op=LOAD Jan 14 00:48:02.713000 audit[4417]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=4406 pid=4417 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.713000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3937376332663937623065623232323836386362633564346231313866 Jan 14 00:48:02.713000 audit: BPF prog-id=241 op=UNLOAD Jan 14 00:48:02.713000 audit[4417]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4406 pid=4417 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.713000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3937376332663937623065623232323836386362633564346231313866 Jan 14 00:48:02.713000 audit: BPF prog-id=240 op=UNLOAD Jan 14 00:48:02.713000 audit[4417]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4406 pid=4417 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.713000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3937376332663937623065623232323836386362633564346231313866 Jan 14 00:48:02.714000 audit: BPF prog-id=242 op=LOAD Jan 14 00:48:02.714000 audit[4417]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=4406 pid=4417 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.714000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3937376332663937623065623232323836386362633564346231313866 Jan 14 00:48:02.726450 kubelet[2802]: E0114 00:48:02.726029 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:02.730281 kubelet[2802]: E0114 00:48:02.730236 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:48:02.748152 systemd-networkd[1504]: calibf1da833257: Link UP Jan 14 00:48:02.749251 systemd-networkd[1504]: calibf1da833257: Gained carrier Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.561 [INFO][4350] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--236--110--177-k8s-calico--apiserver--5f9b965548--bgkj8-eth0 calico-apiserver-5f9b965548- calico-apiserver dcaab5f4-fd68-41e8-986b-14fa1a168bce 878 0 2026-01-14 00:47:37 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:5f9b965548 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s 172-236-110-177 calico-apiserver-5f9b965548-bgkj8 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calibf1da833257 [] [] }} ContainerID="7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-bgkj8" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--bgkj8-" Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.561 [INFO][4350] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-bgkj8" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--bgkj8-eth0" Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.594 [INFO][4377] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" HandleID="k8s-pod-network.7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" Workload="172--236--110--177-k8s-calico--apiserver--5f9b965548--bgkj8-eth0" Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.594 [INFO][4377] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" HandleID="k8s-pod-network.7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" Workload="172--236--110--177-k8s-calico--apiserver--5f9b965548--bgkj8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002ad3a0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"172-236-110-177", "pod":"calico-apiserver-5f9b965548-bgkj8", "timestamp":"2026-01-14 00:48:02.594410077 +0000 UTC"}, Hostname:"172-236-110-177", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.594 [INFO][4377] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.617 [INFO][4377] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.617 [INFO][4377] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-236-110-177' Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.697 [INFO][4377] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" host="172-236-110-177" Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.703 [INFO][4377] ipam/ipam.go 394: Looking up existing affinities for host host="172-236-110-177" Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.711 [INFO][4377] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.713 [INFO][4377] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.715 [INFO][4377] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.715 [INFO][4377] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" host="172-236-110-177" Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.717 [INFO][4377] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.723 [INFO][4377] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" host="172-236-110-177" Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.736 [INFO][4377] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.197/26] block=192.168.109.192/26 handle="k8s-pod-network.7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" host="172-236-110-177" Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.736 [INFO][4377] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.197/26] handle="k8s-pod-network.7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" host="172-236-110-177" Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.736 [INFO][4377] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 00:48:02.767572 containerd[1611]: 2026-01-14 00:48:02.736 [INFO][4377] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.197/26] IPv6=[] ContainerID="7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" HandleID="k8s-pod-network.7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" Workload="172--236--110--177-k8s-calico--apiserver--5f9b965548--bgkj8-eth0" Jan 14 00:48:02.768074 containerd[1611]: 2026-01-14 00:48:02.740 [INFO][4350] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-bgkj8" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--bgkj8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-calico--apiserver--5f9b965548--bgkj8-eth0", GenerateName:"calico-apiserver-5f9b965548-", Namespace:"calico-apiserver", SelfLink:"", UID:"dcaab5f4-fd68-41e8-986b-14fa1a168bce", ResourceVersion:"878", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5f9b965548", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"", Pod:"calico-apiserver-5f9b965548-bgkj8", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.109.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calibf1da833257", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:48:02.768074 containerd[1611]: 2026-01-14 00:48:02.741 [INFO][4350] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.197/32] ContainerID="7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-bgkj8" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--bgkj8-eth0" Jan 14 00:48:02.768074 containerd[1611]: 2026-01-14 00:48:02.742 [INFO][4350] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calibf1da833257 ContainerID="7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-bgkj8" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--bgkj8-eth0" Jan 14 00:48:02.768074 containerd[1611]: 2026-01-14 00:48:02.751 [INFO][4350] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-bgkj8" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--bgkj8-eth0" Jan 14 00:48:02.768074 containerd[1611]: 2026-01-14 00:48:02.752 [INFO][4350] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-bgkj8" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--bgkj8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-calico--apiserver--5f9b965548--bgkj8-eth0", GenerateName:"calico-apiserver-5f9b965548-", Namespace:"calico-apiserver", SelfLink:"", UID:"dcaab5f4-fd68-41e8-986b-14fa1a168bce", ResourceVersion:"878", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5f9b965548", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e", Pod:"calico-apiserver-5f9b965548-bgkj8", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.109.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calibf1da833257", MAC:"82:ae:84:06:7e:5f", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:48:02.768074 containerd[1611]: 2026-01-14 00:48:02.762 [INFO][4350] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" Namespace="calico-apiserver" Pod="calico-apiserver-5f9b965548-bgkj8" WorkloadEndpoint="172--236--110--177-k8s-calico--apiserver--5f9b965548--bgkj8-eth0" Jan 14 00:48:02.791827 containerd[1611]: time="2026-01-14T00:48:02.791797888Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5f9b965548-xv9gd,Uid:32668954-0a97-46fe-9f01-fa91a530c2e3,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"977c2f97b0eb222868cbc5d4b118f7361e1923cf4a9d5492410739de38cdd1fb\"" Jan 14 00:48:02.794996 containerd[1611]: time="2026-01-14T00:48:02.794969774Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 14 00:48:02.797000 audit[4454]: NETFILTER_CFG table=filter:130 family=2 entries=49 op=nft_register_chain pid=4454 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 00:48:02.797000 audit[4454]: SYSCALL arch=c000003e syscall=46 success=yes exit=25452 a0=3 a1=7ffc654793d0 a2=0 a3=7ffc654793bc items=0 ppid=3940 pid=4454 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.797000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 00:48:02.800738 containerd[1611]: time="2026-01-14T00:48:02.800714126Z" level=info msg="connecting to shim 7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e" address="unix:///run/containerd/s/4f104618320b3c7e4763f58b8f97545e427561acca2c83801acd4efb8cc1304b" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:48:02.834277 systemd[1]: Started cri-containerd-7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e.scope - libcontainer container 7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e. Jan 14 00:48:02.851000 audit: BPF prog-id=243 op=LOAD Jan 14 00:48:02.852000 audit: BPF prog-id=244 op=LOAD Jan 14 00:48:02.852000 audit[4470]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a238 a2=98 a3=0 items=0 ppid=4459 pid=4470 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.852000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735393037363335343361626537346633656663366239613165376236 Jan 14 00:48:02.853000 audit: BPF prog-id=244 op=UNLOAD Jan 14 00:48:02.853000 audit[4470]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4459 pid=4470 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.853000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735393037363335343361626537346633656663366239613165376236 Jan 14 00:48:02.853000 audit: BPF prog-id=245 op=LOAD Jan 14 00:48:02.853000 audit[4470]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a488 a2=98 a3=0 items=0 ppid=4459 pid=4470 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.853000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735393037363335343361626537346633656663366239613165376236 Jan 14 00:48:02.853000 audit: BPF prog-id=246 op=LOAD Jan 14 00:48:02.853000 audit[4470]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c00017a218 a2=98 a3=0 items=0 ppid=4459 pid=4470 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.853000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735393037363335343361626537346633656663366239613165376236 Jan 14 00:48:02.853000 audit: BPF prog-id=246 op=UNLOAD Jan 14 00:48:02.853000 audit[4470]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4459 pid=4470 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.853000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735393037363335343361626537346633656663366239613165376236 Jan 14 00:48:02.853000 audit: BPF prog-id=245 op=UNLOAD Jan 14 00:48:02.853000 audit[4470]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4459 pid=4470 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.853000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735393037363335343361626537346633656663366239613165376236 Jan 14 00:48:02.853000 audit: BPF prog-id=247 op=LOAD Jan 14 00:48:02.853000 audit[4470]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c00017a6e8 a2=98 a3=0 items=0 ppid=4459 pid=4470 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:02.853000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3735393037363335343361626537346633656663366239613165376236 Jan 14 00:48:02.894003 containerd[1611]: time="2026-01-14T00:48:02.893963479Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5f9b965548-bgkj8,Uid:dcaab5f4-fd68-41e8-986b-14fa1a168bce,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"7590763543abe74f3efc6b9a1e7b6e4ed2812798467587ef3ed2f09b6b9d376e\"" Jan 14 00:48:02.950058 containerd[1611]: time="2026-01-14T00:48:02.949917957Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:02.950897 containerd[1611]: time="2026-01-14T00:48:02.950850979Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 14 00:48:02.951009 containerd[1611]: time="2026-01-14T00:48:02.950960012Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:02.951176 kubelet[2802]: E0114 00:48:02.951101 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:48:02.951250 kubelet[2802]: E0114 00:48:02.951186 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:48:02.951456 kubelet[2802]: E0114 00:48:02.951398 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zkk9t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5f9b965548-xv9gd_calico-apiserver(32668954-0a97-46fe-9f01-fa91a530c2e3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:02.951691 containerd[1611]: time="2026-01-14T00:48:02.951572273Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 14 00:48:02.952886 kubelet[2802]: E0114 00:48:02.952825 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" podUID="32668954-0a97-46fe-9f01-fa91a530c2e3" Jan 14 00:48:03.076891 containerd[1611]: time="2026-01-14T00:48:03.076830372Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:03.077624 containerd[1611]: time="2026-01-14T00:48:03.077599656Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 14 00:48:03.077715 containerd[1611]: time="2026-01-14T00:48:03.077658378Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:03.077839 kubelet[2802]: E0114 00:48:03.077793 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:48:03.077897 kubelet[2802]: E0114 00:48:03.077851 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:48:03.078047 kubelet[2802]: E0114 00:48:03.077988 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hmr4q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5f9b965548-bgkj8_calico-apiserver(dcaab5f4-fd68-41e8-986b-14fa1a168bce): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:03.079244 kubelet[2802]: E0114 00:48:03.079211 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" podUID="dcaab5f4-fd68-41e8-986b-14fa1a168bce" Jan 14 00:48:03.499784 containerd[1611]: time="2026-01-14T00:48:03.499722819Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-p6wlv,Uid:976c1fb0-d44d-49e6-bc13-822966e59142,Namespace:calico-system,Attempt:0,}" Jan 14 00:48:03.500718 kubelet[2802]: E0114 00:48:03.500067 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:03.501340 containerd[1611]: time="2026-01-14T00:48:03.501312170Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-v8f4p,Uid:e4d15318-1381-4cb4-b138-c44ed14a3a71,Namespace:kube-system,Attempt:0,}" Jan 14 00:48:03.626418 systemd-networkd[1504]: calid13431e53ea: Link UP Jan 14 00:48:03.627236 systemd-networkd[1504]: calid13431e53ea: Gained carrier Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.563 [INFO][4497] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--236--110--177-k8s-coredns--674b8bbfcf--v8f4p-eth0 coredns-674b8bbfcf- kube-system e4d15318-1381-4cb4-b138-c44ed14a3a71 868 0 2026-01-14 00:47:28 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s 172-236-110-177 coredns-674b8bbfcf-v8f4p eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calid13431e53ea [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" Namespace="kube-system" Pod="coredns-674b8bbfcf-v8f4p" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--v8f4p-" Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.563 [INFO][4497] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" Namespace="kube-system" Pod="coredns-674b8bbfcf-v8f4p" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--v8f4p-eth0" Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.589 [INFO][4527] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" HandleID="k8s-pod-network.c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" Workload="172--236--110--177-k8s-coredns--674b8bbfcf--v8f4p-eth0" Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.589 [INFO][4527] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" HandleID="k8s-pod-network.c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" Workload="172--236--110--177-k8s-coredns--674b8bbfcf--v8f4p-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d58f0), Attrs:map[string]string{"namespace":"kube-system", "node":"172-236-110-177", "pod":"coredns-674b8bbfcf-v8f4p", "timestamp":"2026-01-14 00:48:03.58952477 +0000 UTC"}, Hostname:"172-236-110-177", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.589 [INFO][4527] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.589 [INFO][4527] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.589 [INFO][4527] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-236-110-177' Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.595 [INFO][4527] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" host="172-236-110-177" Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.599 [INFO][4527] ipam/ipam.go 394: Looking up existing affinities for host host="172-236-110-177" Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.602 [INFO][4527] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.604 [INFO][4527] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.605 [INFO][4527] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.606 [INFO][4527] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" host="172-236-110-177" Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.607 [INFO][4527] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.611 [INFO][4527] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" host="172-236-110-177" Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.616 [INFO][4527] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.198/26] block=192.168.109.192/26 handle="k8s-pod-network.c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" host="172-236-110-177" Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.616 [INFO][4527] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.198/26] handle="k8s-pod-network.c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" host="172-236-110-177" Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.616 [INFO][4527] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 00:48:03.644156 containerd[1611]: 2026-01-14 00:48:03.617 [INFO][4527] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.198/26] IPv6=[] ContainerID="c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" HandleID="k8s-pod-network.c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" Workload="172--236--110--177-k8s-coredns--674b8bbfcf--v8f4p-eth0" Jan 14 00:48:03.646035 containerd[1611]: 2026-01-14 00:48:03.620 [INFO][4497] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" Namespace="kube-system" Pod="coredns-674b8bbfcf-v8f4p" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--v8f4p-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-coredns--674b8bbfcf--v8f4p-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"e4d15318-1381-4cb4-b138-c44ed14a3a71", ResourceVersion:"868", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"", Pod:"coredns-674b8bbfcf-v8f4p", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.109.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calid13431e53ea", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:48:03.646035 containerd[1611]: 2026-01-14 00:48:03.620 [INFO][4497] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.198/32] ContainerID="c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" Namespace="kube-system" Pod="coredns-674b8bbfcf-v8f4p" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--v8f4p-eth0" Jan 14 00:48:03.646035 containerd[1611]: 2026-01-14 00:48:03.620 [INFO][4497] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid13431e53ea ContainerID="c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" Namespace="kube-system" Pod="coredns-674b8bbfcf-v8f4p" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--v8f4p-eth0" Jan 14 00:48:03.646035 containerd[1611]: 2026-01-14 00:48:03.627 [INFO][4497] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" Namespace="kube-system" Pod="coredns-674b8bbfcf-v8f4p" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--v8f4p-eth0" Jan 14 00:48:03.646035 containerd[1611]: 2026-01-14 00:48:03.629 [INFO][4497] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" Namespace="kube-system" Pod="coredns-674b8bbfcf-v8f4p" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--v8f4p-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-coredns--674b8bbfcf--v8f4p-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"e4d15318-1381-4cb4-b138-c44ed14a3a71", ResourceVersion:"868", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb", Pod:"coredns-674b8bbfcf-v8f4p", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.109.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calid13431e53ea", MAC:"ca:22:92:f3:58:eb", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:48:03.646035 containerd[1611]: 2026-01-14 00:48:03.636 [INFO][4497] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" Namespace="kube-system" Pod="coredns-674b8bbfcf-v8f4p" WorkloadEndpoint="172--236--110--177-k8s-coredns--674b8bbfcf--v8f4p-eth0" Jan 14 00:48:03.667000 audit[4548]: NETFILTER_CFG table=filter:131 family=2 entries=48 op=nft_register_chain pid=4548 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 00:48:03.667000 audit[4548]: SYSCALL arch=c000003e syscall=46 success=yes exit=22720 a0=3 a1=7ffe94e946b0 a2=0 a3=7ffe94e9469c items=0 ppid=3940 pid=4548 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.667000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 00:48:03.675718 containerd[1611]: time="2026-01-14T00:48:03.675659062Z" level=info msg="connecting to shim c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb" address="unix:///run/containerd/s/a4608c2ca005100aa0d4ebe936e19cdfc85bfce969e92128d861cf67970cdef7" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:48:03.712255 systemd[1]: Started cri-containerd-c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb.scope - libcontainer container c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb. Jan 14 00:48:03.734838 kubelet[2802]: E0114 00:48:03.734628 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:03.737130 kubelet[2802]: E0114 00:48:03.735512 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" podUID="32668954-0a97-46fe-9f01-fa91a530c2e3" Jan 14 00:48:03.739490 kubelet[2802]: E0114 00:48:03.739382 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" podUID="dcaab5f4-fd68-41e8-986b-14fa1a168bce" Jan 14 00:48:03.744285 systemd-networkd[1504]: cali5d81dd70a67: Link UP Jan 14 00:48:03.745574 systemd-networkd[1504]: cali5d81dd70a67: Gained carrier Jan 14 00:48:03.751000 audit: BPF prog-id=248 op=LOAD Jan 14 00:48:03.751000 audit: BPF prog-id=249 op=LOAD Jan 14 00:48:03.751000 audit[4569]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=4557 pid=4569 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.751000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6336616136363763326366303734373761356161393336386162323463 Jan 14 00:48:03.754000 audit: BPF prog-id=249 op=UNLOAD Jan 14 00:48:03.754000 audit[4569]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4557 pid=4569 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.754000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6336616136363763326366303734373761356161393336386162323463 Jan 14 00:48:03.757000 audit: BPF prog-id=250 op=LOAD Jan 14 00:48:03.757000 audit[4569]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=4557 pid=4569 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.757000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6336616136363763326366303734373761356161393336386162323463 Jan 14 00:48:03.757000 audit: BPF prog-id=251 op=LOAD Jan 14 00:48:03.757000 audit[4569]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=4557 pid=4569 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.757000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6336616136363763326366303734373761356161393336386162323463 Jan 14 00:48:03.757000 audit: BPF prog-id=251 op=UNLOAD Jan 14 00:48:03.757000 audit[4569]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4557 pid=4569 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.757000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6336616136363763326366303734373761356161393336386162323463 Jan 14 00:48:03.758000 audit: BPF prog-id=250 op=UNLOAD Jan 14 00:48:03.758000 audit[4569]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4557 pid=4569 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.758000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6336616136363763326366303734373761356161393336386162323463 Jan 14 00:48:03.758000 audit: BPF prog-id=252 op=LOAD Jan 14 00:48:03.758000 audit[4569]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=4557 pid=4569 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.758000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6336616136363763326366303734373761356161393336386162323463 Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.564 [INFO][4506] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--236--110--177-k8s-goldmane--666569f655--p6wlv-eth0 goldmane-666569f655- calico-system 976c1fb0-d44d-49e6-bc13-822966e59142 879 0 2026-01-14 00:47:40 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:666569f655 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s 172-236-110-177 goldmane-666569f655-p6wlv eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali5d81dd70a67 [] [] }} ContainerID="07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" Namespace="calico-system" Pod="goldmane-666569f655-p6wlv" WorkloadEndpoint="172--236--110--177-k8s-goldmane--666569f655--p6wlv-" Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.564 [INFO][4506] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" Namespace="calico-system" Pod="goldmane-666569f655-p6wlv" WorkloadEndpoint="172--236--110--177-k8s-goldmane--666569f655--p6wlv-eth0" Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.598 [INFO][4525] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" HandleID="k8s-pod-network.07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" Workload="172--236--110--177-k8s-goldmane--666569f655--p6wlv-eth0" Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.598 [INFO][4525] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" HandleID="k8s-pod-network.07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" Workload="172--236--110--177-k8s-goldmane--666569f655--p6wlv-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d55a0), Attrs:map[string]string{"namespace":"calico-system", "node":"172-236-110-177", "pod":"goldmane-666569f655-p6wlv", "timestamp":"2026-01-14 00:48:03.59818731 +0000 UTC"}, Hostname:"172-236-110-177", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.598 [INFO][4525] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.616 [INFO][4525] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.616 [INFO][4525] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-236-110-177' Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.697 [INFO][4525] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" host="172-236-110-177" Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.702 [INFO][4525] ipam/ipam.go 394: Looking up existing affinities for host host="172-236-110-177" Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.707 [INFO][4525] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.714 [INFO][4525] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.717 [INFO][4525] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.717 [INFO][4525] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" host="172-236-110-177" Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.718 [INFO][4525] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374 Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.723 [INFO][4525] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" host="172-236-110-177" Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.728 [INFO][4525] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.199/26] block=192.168.109.192/26 handle="k8s-pod-network.07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" host="172-236-110-177" Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.729 [INFO][4525] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.199/26] handle="k8s-pod-network.07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" host="172-236-110-177" Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.729 [INFO][4525] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 00:48:03.776989 containerd[1611]: 2026-01-14 00:48:03.729 [INFO][4525] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.199/26] IPv6=[] ContainerID="07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" HandleID="k8s-pod-network.07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" Workload="172--236--110--177-k8s-goldmane--666569f655--p6wlv-eth0" Jan 14 00:48:03.777724 containerd[1611]: 2026-01-14 00:48:03.731 [INFO][4506] cni-plugin/k8s.go 418: Populated endpoint ContainerID="07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" Namespace="calico-system" Pod="goldmane-666569f655-p6wlv" WorkloadEndpoint="172--236--110--177-k8s-goldmane--666569f655--p6wlv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-goldmane--666569f655--p6wlv-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"976c1fb0-d44d-49e6-bc13-822966e59142", ResourceVersion:"879", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"", Pod:"goldmane-666569f655-p6wlv", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.109.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali5d81dd70a67", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:48:03.777724 containerd[1611]: 2026-01-14 00:48:03.731 [INFO][4506] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.199/32] ContainerID="07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" Namespace="calico-system" Pod="goldmane-666569f655-p6wlv" WorkloadEndpoint="172--236--110--177-k8s-goldmane--666569f655--p6wlv-eth0" Jan 14 00:48:03.777724 containerd[1611]: 2026-01-14 00:48:03.731 [INFO][4506] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5d81dd70a67 ContainerID="07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" Namespace="calico-system" Pod="goldmane-666569f655-p6wlv" WorkloadEndpoint="172--236--110--177-k8s-goldmane--666569f655--p6wlv-eth0" Jan 14 00:48:03.777724 containerd[1611]: 2026-01-14 00:48:03.749 [INFO][4506] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" Namespace="calico-system" Pod="goldmane-666569f655-p6wlv" WorkloadEndpoint="172--236--110--177-k8s-goldmane--666569f655--p6wlv-eth0" Jan 14 00:48:03.777724 containerd[1611]: 2026-01-14 00:48:03.750 [INFO][4506] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" Namespace="calico-system" Pod="goldmane-666569f655-p6wlv" WorkloadEndpoint="172--236--110--177-k8s-goldmane--666569f655--p6wlv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-goldmane--666569f655--p6wlv-eth0", GenerateName:"goldmane-666569f655-", Namespace:"calico-system", SelfLink:"", UID:"976c1fb0-d44d-49e6-bc13-822966e59142", ResourceVersion:"879", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"666569f655", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374", Pod:"goldmane-666569f655-p6wlv", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.109.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali5d81dd70a67", MAC:"46:bf:d7:15:7a:e4", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:48:03.777724 containerd[1611]: 2026-01-14 00:48:03.772 [INFO][4506] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" Namespace="calico-system" Pod="goldmane-666569f655-p6wlv" WorkloadEndpoint="172--236--110--177-k8s-goldmane--666569f655--p6wlv-eth0" Jan 14 00:48:03.808228 containerd[1611]: time="2026-01-14T00:48:03.808162416Z" level=info msg="connecting to shim 07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374" address="unix:///run/containerd/s/2aef53fe4712276893f8a71f01529bcdd6d65c6947c7ae661674b76fe169decb" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:48:03.826000 audit[4620]: NETFILTER_CFG table=filter:132 family=2 entries=14 op=nft_register_rule pid=4620 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:48:03.826000 audit[4620]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffc45021a20 a2=0 a3=7ffc45021a0c items=0 ppid=2943 pid=4620 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.826000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:48:03.831000 audit[4620]: NETFILTER_CFG table=nat:133 family=2 entries=20 op=nft_register_rule pid=4620 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:48:03.831000 audit[4620]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffc45021a20 a2=0 a3=7ffc45021a0c items=0 ppid=2943 pid=4620 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.831000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:48:03.849000 audit[4633]: NETFILTER_CFG table=filter:134 family=2 entries=64 op=nft_register_chain pid=4633 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 00:48:03.849000 audit[4633]: SYSCALL arch=c000003e syscall=46 success=yes exit=31120 a0=3 a1=7fff66be4460 a2=0 a3=7fff66be444c items=0 ppid=3940 pid=4633 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.849000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 00:48:03.863051 systemd[1]: Started cri-containerd-07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374.scope - libcontainer container 07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374. Jan 14 00:48:03.869270 containerd[1611]: time="2026-01-14T00:48:03.869208045Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-v8f4p,Uid:e4d15318-1381-4cb4-b138-c44ed14a3a71,Namespace:kube-system,Attempt:0,} returns sandbox id \"c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb\"" Jan 14 00:48:03.868000 audit[4640]: NETFILTER_CFG table=filter:135 family=2 entries=14 op=nft_register_rule pid=4640 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:48:03.868000 audit[4640]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffe59f51fe0 a2=0 a3=7ffe59f51fcc items=0 ppid=2943 pid=4640 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.868000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:48:03.871309 kubelet[2802]: E0114 00:48:03.871288 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:03.873000 audit[4640]: NETFILTER_CFG table=nat:136 family=2 entries=20 op=nft_register_rule pid=4640 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:48:03.873000 audit[4640]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffe59f51fe0 a2=0 a3=7ffe59f51fcc items=0 ppid=2943 pid=4640 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.873000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:48:03.876980 containerd[1611]: time="2026-01-14T00:48:03.876957566Z" level=info msg="CreateContainer within sandbox \"c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 14 00:48:03.885052 containerd[1611]: time="2026-01-14T00:48:03.885024287Z" level=info msg="Container ab1953e0dc320d321adf48fd7222c1df68d7a66e009756802a969c31ac113dd1: CDI devices from CRI Config.CDIDevices: []" Jan 14 00:48:03.888000 audit: BPF prog-id=253 op=LOAD Jan 14 00:48:03.889000 audit: BPF prog-id=254 op=LOAD Jan 14 00:48:03.889000 audit[4618]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000128238 a2=98 a3=0 items=0 ppid=4606 pid=4618 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.889000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3037666230306664316536336233313334643839613130366535323439 Jan 14 00:48:03.889000 audit: BPF prog-id=254 op=UNLOAD Jan 14 00:48:03.889000 audit[4618]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4606 pid=4618 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.889000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3037666230306664316536336233313334643839613130366535323439 Jan 14 00:48:03.889000 audit: BPF prog-id=255 op=LOAD Jan 14 00:48:03.889000 audit[4618]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000128488 a2=98 a3=0 items=0 ppid=4606 pid=4618 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.889000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3037666230306664316536336233313334643839613130366535323439 Jan 14 00:48:03.890000 audit: BPF prog-id=256 op=LOAD Jan 14 00:48:03.890000 audit[4618]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000128218 a2=98 a3=0 items=0 ppid=4606 pid=4618 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.890000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3037666230306664316536336233313334643839613130366535323439 Jan 14 00:48:03.890000 audit: BPF prog-id=256 op=UNLOAD Jan 14 00:48:03.890000 audit[4618]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4606 pid=4618 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.890000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3037666230306664316536336233313334643839613130366535323439 Jan 14 00:48:03.890000 audit: BPF prog-id=255 op=UNLOAD Jan 14 00:48:03.890000 audit[4618]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4606 pid=4618 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.890000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3037666230306664316536336233313334643839613130366535323439 Jan 14 00:48:03.890000 audit: BPF prog-id=257 op=LOAD Jan 14 00:48:03.890000 audit[4618]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001286e8 a2=98 a3=0 items=0 ppid=4606 pid=4618 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.890000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3037666230306664316536336233313334643839613130366535323439 Jan 14 00:48:03.894885 containerd[1611]: time="2026-01-14T00:48:03.894849226Z" level=info msg="CreateContainer within sandbox \"c6aa667c2cf07477a5aa9368ab24cb50fa2b0887c28107ce1d0f7a11c22852bb\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ab1953e0dc320d321adf48fd7222c1df68d7a66e009756802a969c31ac113dd1\"" Jan 14 00:48:03.896252 containerd[1611]: time="2026-01-14T00:48:03.896210310Z" level=info msg="StartContainer for \"ab1953e0dc320d321adf48fd7222c1df68d7a66e009756802a969c31ac113dd1\"" Jan 14 00:48:03.898601 containerd[1611]: time="2026-01-14T00:48:03.898385241Z" level=info msg="connecting to shim ab1953e0dc320d321adf48fd7222c1df68d7a66e009756802a969c31ac113dd1" address="unix:///run/containerd/s/a4608c2ca005100aa0d4ebe936e19cdfc85bfce969e92128d861cf67970cdef7" protocol=ttrpc version=3 Jan 14 00:48:03.920307 systemd[1]: Started cri-containerd-ab1953e0dc320d321adf48fd7222c1df68d7a66e009756802a969c31ac113dd1.scope - libcontainer container ab1953e0dc320d321adf48fd7222c1df68d7a66e009756802a969c31ac113dd1. Jan 14 00:48:03.937232 containerd[1611]: time="2026-01-14T00:48:03.937069294Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-666569f655-p6wlv,Uid:976c1fb0-d44d-49e6-bc13-822966e59142,Namespace:calico-system,Attempt:0,} returns sandbox id \"07fb00fd1e63b3134d89a106e52492a570e104e4ab319e388ab23388a7b1a374\"" Jan 14 00:48:03.937000 audit: BPF prog-id=258 op=LOAD Jan 14 00:48:03.938000 audit: BPF prog-id=259 op=LOAD Jan 14 00:48:03.938000 audit[4649]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130238 a2=98 a3=0 items=0 ppid=4557 pid=4649 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.938000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6162313935336530646333323064333231616466343866643732323263 Jan 14 00:48:03.938000 audit: BPF prog-id=259 op=UNLOAD Jan 14 00:48:03.938000 audit[4649]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4557 pid=4649 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.938000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6162313935336530646333323064333231616466343866643732323263 Jan 14 00:48:03.938000 audit: BPF prog-id=260 op=LOAD Jan 14 00:48:03.938000 audit[4649]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000130488 a2=98 a3=0 items=0 ppid=4557 pid=4649 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.938000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6162313935336530646333323064333231616466343866643732323263 Jan 14 00:48:03.938000 audit: BPF prog-id=261 op=LOAD Jan 14 00:48:03.938000 audit[4649]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000130218 a2=98 a3=0 items=0 ppid=4557 pid=4649 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.938000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6162313935336530646333323064333231616466343866643732323263 Jan 14 00:48:03.939000 audit: BPF prog-id=261 op=UNLOAD Jan 14 00:48:03.939000 audit[4649]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4557 pid=4649 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.939000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6162313935336530646333323064333231616466343866643732323263 Jan 14 00:48:03.939000 audit: BPF prog-id=260 op=UNLOAD Jan 14 00:48:03.939000 audit[4649]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4557 pid=4649 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.939000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6162313935336530646333323064333231616466343866643732323263 Jan 14 00:48:03.939000 audit: BPF prog-id=262 op=LOAD Jan 14 00:48:03.939000 audit[4649]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001306e8 a2=98 a3=0 items=0 ppid=4557 pid=4649 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:03.939000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F6162313935336530646333323064333231616466343866643732323263 Jan 14 00:48:03.941697 containerd[1611]: time="2026-01-14T00:48:03.940613930Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 14 00:48:03.957931 containerd[1611]: time="2026-01-14T00:48:03.957858949Z" level=info msg="StartContainer for \"ab1953e0dc320d321adf48fd7222c1df68d7a66e009756802a969c31ac113dd1\" returns successfully" Jan 14 00:48:04.089649 containerd[1611]: time="2026-01-14T00:48:04.089518805Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:04.090585 containerd[1611]: time="2026-01-14T00:48:04.090512977Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 14 00:48:04.090752 containerd[1611]: time="2026-01-14T00:48:04.090551498Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:04.090970 kubelet[2802]: E0114 00:48:04.090930 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 14 00:48:04.091096 kubelet[2802]: E0114 00:48:04.091073 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 14 00:48:04.091635 kubelet[2802]: E0114 00:48:04.091548 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5vwgj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-p6wlv_calico-system(976c1fb0-d44d-49e6-bc13-822966e59142): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:04.092733 kubelet[2802]: E0114 00:48:04.092695 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-p6wlv" podUID="976c1fb0-d44d-49e6-bc13-822966e59142" Jan 14 00:48:04.245777 systemd-networkd[1504]: cali7dd3c40bc1e: Gained IPv6LL Jan 14 00:48:04.499262 containerd[1611]: time="2026-01-14T00:48:04.499224378Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5f75bc94c6-2nthl,Uid:c12ac0b5-659e-4a97-821e-20b4aabfc024,Namespace:calico-system,Attempt:0,}" Jan 14 00:48:04.603318 systemd-networkd[1504]: cali569e273c9bc: Link UP Jan 14 00:48:04.604250 systemd-networkd[1504]: cali569e273c9bc: Gained carrier Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.534 [INFO][4685] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--236--110--177-k8s-calico--kube--controllers--5f75bc94c6--2nthl-eth0 calico-kube-controllers-5f75bc94c6- calico-system c12ac0b5-659e-4a97-821e-20b4aabfc024 881 0 2026-01-14 00:47:42 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:5f75bc94c6 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s 172-236-110-177 calico-kube-controllers-5f75bc94c6-2nthl eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali569e273c9bc [] [] }} ContainerID="81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" Namespace="calico-system" Pod="calico-kube-controllers-5f75bc94c6-2nthl" WorkloadEndpoint="172--236--110--177-k8s-calico--kube--controllers--5f75bc94c6--2nthl-" Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.534 [INFO][4685] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" Namespace="calico-system" Pod="calico-kube-controllers-5f75bc94c6-2nthl" WorkloadEndpoint="172--236--110--177-k8s-calico--kube--controllers--5f75bc94c6--2nthl-eth0" Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.558 [INFO][4698] ipam/ipam_plugin.go 227: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" HandleID="k8s-pod-network.81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" Workload="172--236--110--177-k8s-calico--kube--controllers--5f75bc94c6--2nthl-eth0" Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.558 [INFO][4698] ipam/ipam_plugin.go 275: Auto assigning IP ContainerID="81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" HandleID="k8s-pod-network.81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" Workload="172--236--110--177-k8s-calico--kube--controllers--5f75bc94c6--2nthl-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00024f270), Attrs:map[string]string{"namespace":"calico-system", "node":"172-236-110-177", "pod":"calico-kube-controllers-5f75bc94c6-2nthl", "timestamp":"2026-01-14 00:48:04.558209376 +0000 UTC"}, Hostname:"172-236-110-177", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.558 [INFO][4698] ipam/ipam_plugin.go 377: About to acquire host-wide IPAM lock. Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.558 [INFO][4698] ipam/ipam_plugin.go 392: Acquired host-wide IPAM lock. Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.558 [INFO][4698] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-236-110-177' Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.568 [INFO][4698] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" host="172-236-110-177" Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.574 [INFO][4698] ipam/ipam.go 394: Looking up existing affinities for host host="172-236-110-177" Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.578 [INFO][4698] ipam/ipam.go 511: Trying affinity for 192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.580 [INFO][4698] ipam/ipam.go 158: Attempting to load block cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.582 [INFO][4698] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.109.192/26 host="172-236-110-177" Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.582 [INFO][4698] ipam/ipam.go 1219: Attempting to assign 1 addresses from block block=192.168.109.192/26 handle="k8s-pod-network.81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" host="172-236-110-177" Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.587 [INFO][4698] ipam/ipam.go 1780: Creating new handle: k8s-pod-network.81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525 Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.591 [INFO][4698] ipam/ipam.go 1246: Writing block in order to claim IPs block=192.168.109.192/26 handle="k8s-pod-network.81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" host="172-236-110-177" Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.596 [INFO][4698] ipam/ipam.go 1262: Successfully claimed IPs: [192.168.109.200/26] block=192.168.109.192/26 handle="k8s-pod-network.81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" host="172-236-110-177" Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.596 [INFO][4698] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.109.200/26] handle="k8s-pod-network.81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" host="172-236-110-177" Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.596 [INFO][4698] ipam/ipam_plugin.go 398: Released host-wide IPAM lock. Jan 14 00:48:04.626323 containerd[1611]: 2026-01-14 00:48:04.596 [INFO][4698] ipam/ipam_plugin.go 299: Calico CNI IPAM assigned addresses IPv4=[192.168.109.200/26] IPv6=[] ContainerID="81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" HandleID="k8s-pod-network.81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" Workload="172--236--110--177-k8s-calico--kube--controllers--5f75bc94c6--2nthl-eth0" Jan 14 00:48:04.628688 containerd[1611]: 2026-01-14 00:48:04.598 [INFO][4685] cni-plugin/k8s.go 418: Populated endpoint ContainerID="81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" Namespace="calico-system" Pod="calico-kube-controllers-5f75bc94c6-2nthl" WorkloadEndpoint="172--236--110--177-k8s-calico--kube--controllers--5f75bc94c6--2nthl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-calico--kube--controllers--5f75bc94c6--2nthl-eth0", GenerateName:"calico-kube-controllers-5f75bc94c6-", Namespace:"calico-system", SelfLink:"", UID:"c12ac0b5-659e-4a97-821e-20b4aabfc024", ResourceVersion:"881", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 42, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5f75bc94c6", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"", Pod:"calico-kube-controllers-5f75bc94c6-2nthl", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.109.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali569e273c9bc", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:48:04.628688 containerd[1611]: 2026-01-14 00:48:04.599 [INFO][4685] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.109.200/32] ContainerID="81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" Namespace="calico-system" Pod="calico-kube-controllers-5f75bc94c6-2nthl" WorkloadEndpoint="172--236--110--177-k8s-calico--kube--controllers--5f75bc94c6--2nthl-eth0" Jan 14 00:48:04.628688 containerd[1611]: 2026-01-14 00:48:04.599 [INFO][4685] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali569e273c9bc ContainerID="81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" Namespace="calico-system" Pod="calico-kube-controllers-5f75bc94c6-2nthl" WorkloadEndpoint="172--236--110--177-k8s-calico--kube--controllers--5f75bc94c6--2nthl-eth0" Jan 14 00:48:04.628688 containerd[1611]: 2026-01-14 00:48:04.601 [INFO][4685] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" Namespace="calico-system" Pod="calico-kube-controllers-5f75bc94c6-2nthl" WorkloadEndpoint="172--236--110--177-k8s-calico--kube--controllers--5f75bc94c6--2nthl-eth0" Jan 14 00:48:04.628688 containerd[1611]: 2026-01-14 00:48:04.601 [INFO][4685] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" Namespace="calico-system" Pod="calico-kube-controllers-5f75bc94c6-2nthl" WorkloadEndpoint="172--236--110--177-k8s-calico--kube--controllers--5f75bc94c6--2nthl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--236--110--177-k8s-calico--kube--controllers--5f75bc94c6--2nthl-eth0", GenerateName:"calico-kube-controllers-5f75bc94c6-", Namespace:"calico-system", SelfLink:"", UID:"c12ac0b5-659e-4a97-821e-20b4aabfc024", ResourceVersion:"881", Generation:0, CreationTimestamp:time.Date(2026, time.January, 14, 0, 47, 42, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5f75bc94c6", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-236-110-177", ContainerID:"81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525", Pod:"calico-kube-controllers-5f75bc94c6-2nthl", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.109.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali569e273c9bc", MAC:"e2:11:e4:fb:fd:d3", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jan 14 00:48:04.628688 containerd[1611]: 2026-01-14 00:48:04.613 [INFO][4685] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" Namespace="calico-system" Pod="calico-kube-controllers-5f75bc94c6-2nthl" WorkloadEndpoint="172--236--110--177-k8s-calico--kube--controllers--5f75bc94c6--2nthl-eth0" Jan 14 00:48:04.660479 containerd[1611]: time="2026-01-14T00:48:04.660335133Z" level=info msg="connecting to shim 81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525" address="unix:///run/containerd/s/89bc02f6c8bbafc9782f387b56fdc59ae16150b97b3c6bb4a23a57afc5eedf2d" namespace=k8s.io protocol=ttrpc version=3 Jan 14 00:48:04.695296 systemd[1]: Started cri-containerd-81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525.scope - libcontainer container 81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525. Jan 14 00:48:04.697000 audit[4737]: NETFILTER_CFG table=filter:137 family=2 entries=66 op=nft_register_chain pid=4737 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Jan 14 00:48:04.697000 audit[4737]: SYSCALL arch=c000003e syscall=46 success=yes exit=29556 a0=3 a1=7ffcd75df8f0 a2=0 a3=7ffcd75df8dc items=0 ppid=3940 pid=4737 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:04.697000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Jan 14 00:48:04.719000 audit: BPF prog-id=263 op=LOAD Jan 14 00:48:04.719000 audit: BPF prog-id=264 op=LOAD Jan 14 00:48:04.719000 audit[4732]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106238 a2=98 a3=0 items=0 ppid=4721 pid=4732 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:04.719000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3831633033373733626632346662623733363938383438366666626465 Jan 14 00:48:04.720000 audit: BPF prog-id=264 op=UNLOAD Jan 14 00:48:04.720000 audit[4732]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4721 pid=4732 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:04.720000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3831633033373733626632346662623733363938383438366666626465 Jan 14 00:48:04.720000 audit: BPF prog-id=265 op=LOAD Jan 14 00:48:04.720000 audit[4732]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c000106488 a2=98 a3=0 items=0 ppid=4721 pid=4732 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:04.720000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3831633033373733626632346662623733363938383438366666626465 Jan 14 00:48:04.720000 audit: BPF prog-id=266 op=LOAD Jan 14 00:48:04.720000 audit[4732]: SYSCALL arch=c000003e syscall=321 success=yes exit=23 a0=5 a1=c000106218 a2=98 a3=0 items=0 ppid=4721 pid=4732 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:04.720000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3831633033373733626632346662623733363938383438366666626465 Jan 14 00:48:04.720000 audit: BPF prog-id=266 op=UNLOAD Jan 14 00:48:04.720000 audit[4732]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=17 a1=0 a2=0 a3=0 items=0 ppid=4721 pid=4732 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:04.720000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3831633033373733626632346662623733363938383438366666626465 Jan 14 00:48:04.720000 audit: BPF prog-id=265 op=UNLOAD Jan 14 00:48:04.720000 audit[4732]: SYSCALL arch=c000003e syscall=3 success=yes exit=0 a0=15 a1=0 a2=0 a3=0 items=0 ppid=4721 pid=4732 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:04.720000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3831633033373733626632346662623733363938383438366666626465 Jan 14 00:48:04.720000 audit: BPF prog-id=267 op=LOAD Jan 14 00:48:04.720000 audit[4732]: SYSCALL arch=c000003e syscall=321 success=yes exit=21 a0=5 a1=c0001066e8 a2=98 a3=0 items=0 ppid=4721 pid=4732 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="runc" exe="/usr/bin/runc" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:04.720000 audit: PROCTITLE proctitle=72756E63002D2D726F6F74002F72756E2F636F6E7461696E6572642F72756E632F6B38732E696F002D2D6C6F67002F72756E2F636F6E7461696E6572642F696F2E636F6E7461696E6572642E72756E74696D652E76322E7461736B2F6B38732E696F2F3831633033373733626632346662623733363938383438366666626465 Jan 14 00:48:04.743988 kubelet[2802]: E0114 00:48:04.743961 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-p6wlv" podUID="976c1fb0-d44d-49e6-bc13-822966e59142" Jan 14 00:48:04.751252 kubelet[2802]: E0114 00:48:04.750036 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:04.755375 kubelet[2802]: E0114 00:48:04.752605 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" podUID="dcaab5f4-fd68-41e8-986b-14fa1a168bce" Jan 14 00:48:04.755375 kubelet[2802]: E0114 00:48:04.752727 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" podUID="32668954-0a97-46fe-9f01-fa91a530c2e3" Jan 14 00:48:04.758601 systemd-networkd[1504]: calibf1da833257: Gained IPv6LL Jan 14 00:48:04.802978 containerd[1611]: time="2026-01-14T00:48:04.802884443Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5f75bc94c6-2nthl,Uid:c12ac0b5-659e-4a97-821e-20b4aabfc024,Namespace:calico-system,Attempt:0,} returns sandbox id \"81c03773bf24fbb736988486ffbde1b747360b437eeb93e4aa1d5f4b6f49f525\"" Jan 14 00:48:04.805468 containerd[1611]: time="2026-01-14T00:48:04.805440723Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 14 00:48:04.817587 kubelet[2802]: I0114 00:48:04.817529 2802 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-v8f4p" podStartSLOduration=36.817514393 podStartE2EDuration="36.817514393s" podCreationTimestamp="2026-01-14 00:47:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-14 00:48:04.816105629 +0000 UTC m=+43.411202073" watchObservedRunningTime="2026-01-14 00:48:04.817514393 +0000 UTC m=+43.412610817" Jan 14 00:48:04.887000 audit[4759]: NETFILTER_CFG table=filter:138 family=2 entries=14 op=nft_register_rule pid=4759 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:48:04.887000 audit[4759]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffcd7d5ac20 a2=0 a3=7ffcd7d5ac0c items=0 ppid=2943 pid=4759 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:04.887000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:48:04.891000 audit[4759]: NETFILTER_CFG table=nat:139 family=2 entries=20 op=nft_register_rule pid=4759 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:48:04.891000 audit[4759]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffcd7d5ac20 a2=0 a3=7ffcd7d5ac0c items=0 ppid=2943 pid=4759 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:04.891000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:48:04.935825 containerd[1611]: time="2026-01-14T00:48:04.935587882Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:04.936611 containerd[1611]: time="2026-01-14T00:48:04.936486570Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 14 00:48:04.936802 containerd[1611]: time="2026-01-14T00:48:04.936557333Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:04.937071 kubelet[2802]: E0114 00:48:04.937035 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 14 00:48:04.937144 kubelet[2802]: E0114 00:48:04.937077 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 14 00:48:04.937521 kubelet[2802]: E0114 00:48:04.937466 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sfqdv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-5f75bc94c6-2nthl_calico-system(c12ac0b5-659e-4a97-821e-20b4aabfc024): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:04.939297 kubelet[2802]: E0114 00:48:04.939257 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" podUID="c12ac0b5-659e-4a97-821e-20b4aabfc024" Jan 14 00:48:05.269277 systemd-networkd[1504]: cali5d81dd70a67: Gained IPv6LL Jan 14 00:48:05.653320 systemd-networkd[1504]: calid13431e53ea: Gained IPv6LL Jan 14 00:48:05.752396 kubelet[2802]: E0114 00:48:05.752333 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:05.754001 kubelet[2802]: E0114 00:48:05.753401 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" podUID="c12ac0b5-659e-4a97-821e-20b4aabfc024" Jan 14 00:48:05.754001 kubelet[2802]: E0114 00:48:05.753936 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-p6wlv" podUID="976c1fb0-d44d-49e6-bc13-822966e59142" Jan 14 00:48:05.905000 audit[4767]: NETFILTER_CFG table=filter:140 family=2 entries=14 op=nft_register_rule pid=4767 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:48:05.907247 kernel: kauditd_printk_skb: 233 callbacks suppressed Jan 14 00:48:05.907330 kernel: audit: type=1325 audit(1768351685.905:754): table=filter:140 family=2 entries=14 op=nft_register_rule pid=4767 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:48:05.911460 systemd-networkd[1504]: cali569e273c9bc: Gained IPv6LL Jan 14 00:48:05.905000 audit[4767]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffcb2335a40 a2=0 a3=7ffcb2335a2c items=0 ppid=2943 pid=4767 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:05.926603 kernel: audit: type=1300 audit(1768351685.905:754): arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffcb2335a40 a2=0 a3=7ffcb2335a2c items=0 ppid=2943 pid=4767 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:05.926719 kernel: audit: type=1327 audit(1768351685.905:754): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:48:05.905000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:48:05.926000 audit[4767]: NETFILTER_CFG table=nat:141 family=2 entries=56 op=nft_register_chain pid=4767 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:48:05.926000 audit[4767]: SYSCALL arch=c000003e syscall=46 success=yes exit=19860 a0=3 a1=7ffcb2335a40 a2=0 a3=7ffcb2335a2c items=0 ppid=2943 pid=4767 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:05.934463 kernel: audit: type=1325 audit(1768351685.926:755): table=nat:141 family=2 entries=56 op=nft_register_chain pid=4767 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:48:05.934507 kernel: audit: type=1300 audit(1768351685.926:755): arch=c000003e syscall=46 success=yes exit=19860 a0=3 a1=7ffcb2335a40 a2=0 a3=7ffcb2335a2c items=0 ppid=2943 pid=4767 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:48:05.943468 kernel: audit: type=1327 audit(1768351685.926:755): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:48:05.926000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:48:06.754572 kubelet[2802]: E0114 00:48:06.754512 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" podUID="c12ac0b5-659e-4a97-821e-20b4aabfc024" Jan 14 00:48:06.755297 kubelet[2802]: E0114 00:48:06.755174 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:07.501888 containerd[1611]: time="2026-01-14T00:48:07.501848748Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 14 00:48:07.626961 containerd[1611]: time="2026-01-14T00:48:07.626754519Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:07.627675 containerd[1611]: time="2026-01-14T00:48:07.627619384Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 14 00:48:07.627741 containerd[1611]: time="2026-01-14T00:48:07.627700786Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:07.627907 kubelet[2802]: E0114 00:48:07.627864 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 14 00:48:07.628004 kubelet[2802]: E0114 00:48:07.627916 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 14 00:48:07.628102 kubelet[2802]: E0114 00:48:07.628046 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:3e9ed11e1ee14277abd17039bcf9cf15,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rdmgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-d6f8b64b4-m4ctk_calico-system(db919709-47a0-46a4-a2ab-3f096374d888): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:07.631625 containerd[1611]: time="2026-01-14T00:48:07.631593070Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 14 00:48:07.756069 containerd[1611]: time="2026-01-14T00:48:07.755865613Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:07.757378 containerd[1611]: time="2026-01-14T00:48:07.757319335Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 14 00:48:07.757378 containerd[1611]: time="2026-01-14T00:48:07.757354636Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:07.757753 kubelet[2802]: E0114 00:48:07.757605 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 14 00:48:07.758226 kubelet[2802]: E0114 00:48:07.757763 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 14 00:48:07.758226 kubelet[2802]: E0114 00:48:07.757889 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdmgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-d6f8b64b4-m4ctk_calico-system(db919709-47a0-46a4-a2ab-3f096374d888): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:07.759310 kubelet[2802]: E0114 00:48:07.759277 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-d6f8b64b4-m4ctk" podUID="db919709-47a0-46a4-a2ab-3f096374d888" Jan 14 00:48:10.676951 kubelet[2802]: I0114 00:48:10.676553 2802 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 14 00:48:10.677455 kubelet[2802]: E0114 00:48:10.677414 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:10.761262 kubelet[2802]: E0114 00:48:10.760960 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:16.502395 containerd[1611]: time="2026-01-14T00:48:16.502312453Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 14 00:48:16.631883 containerd[1611]: time="2026-01-14T00:48:16.631763300Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:16.633141 containerd[1611]: time="2026-01-14T00:48:16.633039841Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 14 00:48:16.633314 containerd[1611]: time="2026-01-14T00:48:16.633059802Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:16.633479 kubelet[2802]: E0114 00:48:16.633421 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 14 00:48:16.633479 kubelet[2802]: E0114 00:48:16.633477 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 14 00:48:16.634996 kubelet[2802]: E0114 00:48:16.633595 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pnt5f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-vzqkb_calico-system(f41cc627-bd07-42f1-8e88-f85dcd2c063c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:16.636444 containerd[1611]: time="2026-01-14T00:48:16.636399372Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 14 00:48:16.767737 containerd[1611]: time="2026-01-14T00:48:16.767243972Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:16.769209 containerd[1611]: time="2026-01-14T00:48:16.769032606Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:16.769994 containerd[1611]: time="2026-01-14T00:48:16.769328462Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 14 00:48:16.770616 kubelet[2802]: E0114 00:48:16.770550 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 14 00:48:16.770769 kubelet[2802]: E0114 00:48:16.770735 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 14 00:48:16.771103 kubelet[2802]: E0114 00:48:16.771038 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pnt5f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-vzqkb_calico-system(f41cc627-bd07-42f1-8e88-f85dcd2c063c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:16.772215 kubelet[2802]: E0114 00:48:16.772181 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:48:18.501094 containerd[1611]: time="2026-01-14T00:48:18.501053208Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 14 00:48:18.628648 containerd[1611]: time="2026-01-14T00:48:18.628448396Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:18.629603 containerd[1611]: time="2026-01-14T00:48:18.629483421Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 14 00:48:18.629603 containerd[1611]: time="2026-01-14T00:48:18.629578613Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:18.629802 kubelet[2802]: E0114 00:48:18.629731 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:48:18.629802 kubelet[2802]: E0114 00:48:18.629784 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:48:18.631068 kubelet[2802]: E0114 00:48:18.630036 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zkk9t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5f9b965548-xv9gd_calico-apiserver(32668954-0a97-46fe-9f01-fa91a530c2e3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:18.631187 containerd[1611]: time="2026-01-14T00:48:18.630526295Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 14 00:48:18.631223 kubelet[2802]: E0114 00:48:18.631182 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" podUID="32668954-0a97-46fe-9f01-fa91a530c2e3" Jan 14 00:48:18.757220 containerd[1611]: time="2026-01-14T00:48:18.755693973Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:18.758210 containerd[1611]: time="2026-01-14T00:48:18.758098899Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 14 00:48:18.758210 containerd[1611]: time="2026-01-14T00:48:18.758185301Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:18.758543 kubelet[2802]: E0114 00:48:18.758481 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:48:18.758543 kubelet[2802]: E0114 00:48:18.758525 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:48:18.759778 kubelet[2802]: E0114 00:48:18.759104 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hmr4q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5f9b965548-bgkj8_calico-apiserver(dcaab5f4-fd68-41e8-986b-14fa1a168bce): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:18.761066 kubelet[2802]: E0114 00:48:18.761026 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" podUID="dcaab5f4-fd68-41e8-986b-14fa1a168bce" Jan 14 00:48:19.502447 containerd[1611]: time="2026-01-14T00:48:19.502354682Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 14 00:48:19.639418 containerd[1611]: time="2026-01-14T00:48:19.639339667Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:19.640690 containerd[1611]: time="2026-01-14T00:48:19.640651557Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 14 00:48:19.641042 containerd[1611]: time="2026-01-14T00:48:19.640864823Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:19.641837 kubelet[2802]: E0114 00:48:19.641801 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 14 00:48:19.643223 kubelet[2802]: E0114 00:48:19.642325 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 14 00:48:19.643356 kubelet[2802]: E0114 00:48:19.643311 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5vwgj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-p6wlv_calico-system(976c1fb0-d44d-49e6-bc13-822966e59142): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:19.644625 kubelet[2802]: E0114 00:48:19.644559 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-p6wlv" podUID="976c1fb0-d44d-49e6-bc13-822966e59142" Jan 14 00:48:20.501513 containerd[1611]: time="2026-01-14T00:48:20.501450598Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 14 00:48:20.631236 containerd[1611]: time="2026-01-14T00:48:20.631160275Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:20.633052 containerd[1611]: time="2026-01-14T00:48:20.633012067Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 14 00:48:20.633191 containerd[1611]: time="2026-01-14T00:48:20.633141110Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:20.633506 kubelet[2802]: E0114 00:48:20.633463 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 14 00:48:20.633553 kubelet[2802]: E0114 00:48:20.633510 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 14 00:48:20.633672 kubelet[2802]: E0114 00:48:20.633630 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sfqdv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-5f75bc94c6-2nthl_calico-system(c12ac0b5-659e-4a97-821e-20b4aabfc024): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:20.635129 kubelet[2802]: E0114 00:48:20.635064 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" podUID="c12ac0b5-659e-4a97-821e-20b4aabfc024" Jan 14 00:48:21.513716 kubelet[2802]: E0114 00:48:21.513624 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-d6f8b64b4-m4ctk" podUID="db919709-47a0-46a4-a2ab-3f096374d888" Jan 14 00:48:31.501390 kubelet[2802]: E0114 00:48:31.501346 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" podUID="c12ac0b5-659e-4a97-821e-20b4aabfc024" Jan 14 00:48:31.502307 kubelet[2802]: E0114 00:48:31.501658 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" podUID="32668954-0a97-46fe-9f01-fa91a530c2e3" Jan 14 00:48:31.502307 kubelet[2802]: E0114 00:48:31.502045 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:48:32.505814 containerd[1611]: time="2026-01-14T00:48:32.503564522Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 14 00:48:32.683864 containerd[1611]: time="2026-01-14T00:48:32.683790994Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:32.685100 containerd[1611]: time="2026-01-14T00:48:32.685027218Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 14 00:48:32.685209 containerd[1611]: time="2026-01-14T00:48:32.685168857Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:32.685398 kubelet[2802]: E0114 00:48:32.685346 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 14 00:48:32.686084 kubelet[2802]: E0114 00:48:32.685406 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 14 00:48:32.686084 kubelet[2802]: E0114 00:48:32.685764 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:3e9ed11e1ee14277abd17039bcf9cf15,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rdmgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-d6f8b64b4-m4ctk_calico-system(db919709-47a0-46a4-a2ab-3f096374d888): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:32.688047 containerd[1611]: time="2026-01-14T00:48:32.687896442Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 14 00:48:32.819369 containerd[1611]: time="2026-01-14T00:48:32.819212517Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:32.820135 containerd[1611]: time="2026-01-14T00:48:32.820023252Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 14 00:48:32.820135 containerd[1611]: time="2026-01-14T00:48:32.820089212Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:32.820320 kubelet[2802]: E0114 00:48:32.820280 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 14 00:48:32.820393 kubelet[2802]: E0114 00:48:32.820324 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 14 00:48:32.820485 kubelet[2802]: E0114 00:48:32.820445 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdmgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-d6f8b64b4-m4ctk_calico-system(db919709-47a0-46a4-a2ab-3f096374d888): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:32.821771 kubelet[2802]: E0114 00:48:32.821713 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-d6f8b64b4-m4ctk" podUID="db919709-47a0-46a4-a2ab-3f096374d888" Jan 14 00:48:33.501578 kubelet[2802]: E0114 00:48:33.501512 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" podUID="dcaab5f4-fd68-41e8-986b-14fa1a168bce" Jan 14 00:48:34.499852 kubelet[2802]: E0114 00:48:34.499794 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-p6wlv" podUID="976c1fb0-d44d-49e6-bc13-822966e59142" Jan 14 00:48:42.500662 containerd[1611]: time="2026-01-14T00:48:42.500502160Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 14 00:48:42.635712 containerd[1611]: time="2026-01-14T00:48:42.635664153Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:42.636464 containerd[1611]: time="2026-01-14T00:48:42.636426713Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 14 00:48:42.636543 containerd[1611]: time="2026-01-14T00:48:42.636525983Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:42.636719 kubelet[2802]: E0114 00:48:42.636677 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:48:42.637064 kubelet[2802]: E0114 00:48:42.636737 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:48:42.637064 kubelet[2802]: E0114 00:48:42.636857 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zkk9t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5f9b965548-xv9gd_calico-apiserver(32668954-0a97-46fe-9f01-fa91a530c2e3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:42.638622 kubelet[2802]: E0114 00:48:42.638585 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" podUID="32668954-0a97-46fe-9f01-fa91a530c2e3" Jan 14 00:48:43.504932 containerd[1611]: time="2026-01-14T00:48:43.504883891Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 14 00:48:43.646040 containerd[1611]: time="2026-01-14T00:48:43.645978532Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:43.646923 containerd[1611]: time="2026-01-14T00:48:43.646871542Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 14 00:48:43.647158 containerd[1611]: time="2026-01-14T00:48:43.646913272Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:43.647317 kubelet[2802]: E0114 00:48:43.647275 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 14 00:48:43.648140 kubelet[2802]: E0114 00:48:43.647328 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 14 00:48:43.648140 kubelet[2802]: E0114 00:48:43.647460 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sfqdv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-5f75bc94c6-2nthl_calico-system(c12ac0b5-659e-4a97-821e-20b4aabfc024): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:43.648751 kubelet[2802]: E0114 00:48:43.648715 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" podUID="c12ac0b5-659e-4a97-821e-20b4aabfc024" Jan 14 00:48:45.500682 kubelet[2802]: E0114 00:48:45.500564 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:45.502689 containerd[1611]: time="2026-01-14T00:48:45.501183766Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 14 00:48:45.636512 containerd[1611]: time="2026-01-14T00:48:45.636436783Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:45.637725 containerd[1611]: time="2026-01-14T00:48:45.637668055Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 14 00:48:45.637982 containerd[1611]: time="2026-01-14T00:48:45.637776645Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:45.638619 kubelet[2802]: E0114 00:48:45.638398 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:48:45.638619 kubelet[2802]: E0114 00:48:45.638482 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:48:45.639292 kubelet[2802]: E0114 00:48:45.639254 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hmr4q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5f9b965548-bgkj8_calico-apiserver(dcaab5f4-fd68-41e8-986b-14fa1a168bce): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:45.640867 kubelet[2802]: E0114 00:48:45.640827 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" podUID="dcaab5f4-fd68-41e8-986b-14fa1a168bce" Jan 14 00:48:46.502064 containerd[1611]: time="2026-01-14T00:48:46.501989612Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 14 00:48:46.504163 kubelet[2802]: E0114 00:48:46.504126 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-d6f8b64b4-m4ctk" podUID="db919709-47a0-46a4-a2ab-3f096374d888" Jan 14 00:48:46.628244 containerd[1611]: time="2026-01-14T00:48:46.628201391Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:46.629176 containerd[1611]: time="2026-01-14T00:48:46.629139442Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 14 00:48:46.629231 containerd[1611]: time="2026-01-14T00:48:46.629208512Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:46.629440 kubelet[2802]: E0114 00:48:46.629402 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 14 00:48:46.629505 kubelet[2802]: E0114 00:48:46.629453 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 14 00:48:46.629690 kubelet[2802]: E0114 00:48:46.629625 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pnt5f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-vzqkb_calico-system(f41cc627-bd07-42f1-8e88-f85dcd2c063c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:46.631983 containerd[1611]: time="2026-01-14T00:48:46.631943307Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 14 00:48:46.759688 containerd[1611]: time="2026-01-14T00:48:46.759491309Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:46.760776 containerd[1611]: time="2026-01-14T00:48:46.760729932Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 14 00:48:46.760841 containerd[1611]: time="2026-01-14T00:48:46.760804862Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:46.761030 kubelet[2802]: E0114 00:48:46.760987 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 14 00:48:46.761096 kubelet[2802]: E0114 00:48:46.761053 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 14 00:48:46.761344 kubelet[2802]: E0114 00:48:46.761231 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pnt5f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-vzqkb_calico-system(f41cc627-bd07-42f1-8e88-f85dcd2c063c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:46.762461 kubelet[2802]: E0114 00:48:46.762420 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:48:47.502988 containerd[1611]: time="2026-01-14T00:48:47.502939589Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 14 00:48:47.633610 containerd[1611]: time="2026-01-14T00:48:47.633536051Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:48:47.634927 containerd[1611]: time="2026-01-14T00:48:47.634827624Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 14 00:48:47.634927 containerd[1611]: time="2026-01-14T00:48:47.634902864Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 14 00:48:47.635287 kubelet[2802]: E0114 00:48:47.635224 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 14 00:48:47.635287 kubelet[2802]: E0114 00:48:47.635265 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 14 00:48:47.637136 kubelet[2802]: E0114 00:48:47.636292 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5vwgj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-p6wlv_calico-system(976c1fb0-d44d-49e6-bc13-822966e59142): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 14 00:48:47.638164 kubelet[2802]: E0114 00:48:47.638136 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-p6wlv" podUID="976c1fb0-d44d-49e6-bc13-822966e59142" Jan 14 00:48:49.500826 kubelet[2802]: E0114 00:48:49.500792 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:51.502139 kubelet[2802]: E0114 00:48:51.502082 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:54.500708 kubelet[2802]: E0114 00:48:54.500662 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" podUID="32668954-0a97-46fe-9f01-fa91a530c2e3" Jan 14 00:48:54.501171 kubelet[2802]: E0114 00:48:54.500942 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" podUID="c12ac0b5-659e-4a97-821e-20b4aabfc024" Jan 14 00:48:56.500048 kubelet[2802]: E0114 00:48:56.500005 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:48:59.503062 kubelet[2802]: E0114 00:48:59.502523 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-p6wlv" podUID="976c1fb0-d44d-49e6-bc13-822966e59142" Jan 14 00:48:59.503062 kubelet[2802]: E0114 00:48:59.502667 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-d6f8b64b4-m4ctk" podUID="db919709-47a0-46a4-a2ab-3f096374d888" Jan 14 00:48:59.505058 kubelet[2802]: E0114 00:48:59.505010 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:49:00.502157 kubelet[2802]: E0114 00:49:00.502054 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" podUID="dcaab5f4-fd68-41e8-986b-14fa1a168bce" Jan 14 00:49:09.501430 kubelet[2802]: E0114 00:49:09.500977 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" podUID="c12ac0b5-659e-4a97-821e-20b4aabfc024" Jan 14 00:49:09.501430 kubelet[2802]: E0114 00:49:09.501060 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" podUID="32668954-0a97-46fe-9f01-fa91a530c2e3" Jan 14 00:49:12.500930 kubelet[2802]: E0114 00:49:12.500535 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:49:12.504458 kubelet[2802]: E0114 00:49:12.504406 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:49:13.502911 kubelet[2802]: E0114 00:49:13.502210 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" podUID="dcaab5f4-fd68-41e8-986b-14fa1a168bce" Jan 14 00:49:13.505962 kubelet[2802]: E0114 00:49:13.505354 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-p6wlv" podUID="976c1fb0-d44d-49e6-bc13-822966e59142" Jan 14 00:49:14.501239 containerd[1611]: time="2026-01-14T00:49:14.501184132Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\"" Jan 14 00:49:14.634555 containerd[1611]: time="2026-01-14T00:49:14.634315602Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:49:14.635605 containerd[1611]: time="2026-01-14T00:49:14.635332622Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" Jan 14 00:49:14.635605 containerd[1611]: time="2026-01-14T00:49:14.635362372Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.4: active requests=0, bytes read=0" Jan 14 00:49:14.637240 kubelet[2802]: E0114 00:49:14.637087 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 14 00:49:14.637240 kubelet[2802]: E0114 00:49:14.637169 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker:v3.30.4" Jan 14 00:49:14.638350 kubelet[2802]: E0114 00:49:14.638293 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.4,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:3e9ed11e1ee14277abd17039bcf9cf15,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rdmgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-d6f8b64b4-m4ctk_calico-system(db919709-47a0-46a4-a2ab-3f096374d888): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found" logger="UnhandledError" Jan 14 00:49:14.641085 containerd[1611]: time="2026-01-14T00:49:14.640988717Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\"" Jan 14 00:49:14.781930 containerd[1611]: time="2026-01-14T00:49:14.781088385Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:49:14.785261 containerd[1611]: time="2026-01-14T00:49:14.785102994Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" Jan 14 00:49:14.785261 containerd[1611]: time="2026-01-14T00:49:14.785141704Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.4: active requests=0, bytes read=0" Jan 14 00:49:14.785654 kubelet[2802]: E0114 00:49:14.785588 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 14 00:49:14.785818 kubelet[2802]: E0114 00:49:14.785790 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.4" Jan 14 00:49:14.786044 kubelet[2802]: E0114 00:49:14.785999 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdmgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-d6f8b64b4-m4ctk_calico-system(db919709-47a0-46a4-a2ab-3f096374d888): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found" logger="UnhandledError" Jan 14 00:49:14.787470 kubelet[2802]: E0114 00:49:14.787407 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-d6f8b64b4-m4ctk" podUID="db919709-47a0-46a4-a2ab-3f096374d888" Jan 14 00:49:15.500828 kubelet[2802]: E0114 00:49:15.499822 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:49:16.500146 kubelet[2802]: E0114 00:49:16.499862 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:49:21.502145 kubelet[2802]: E0114 00:49:21.502028 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" podUID="c12ac0b5-659e-4a97-821e-20b4aabfc024" Jan 14 00:49:24.502147 kubelet[2802]: E0114 00:49:24.502084 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" podUID="dcaab5f4-fd68-41e8-986b-14fa1a168bce" Jan 14 00:49:24.502624 containerd[1611]: time="2026-01-14T00:49:24.502431177Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 14 00:49:24.705943 containerd[1611]: time="2026-01-14T00:49:24.705862626Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:49:24.706963 containerd[1611]: time="2026-01-14T00:49:24.706909507Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 14 00:49:24.707041 containerd[1611]: time="2026-01-14T00:49:24.706996568Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 14 00:49:24.707241 kubelet[2802]: E0114 00:49:24.707202 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:49:24.707308 kubelet[2802]: E0114 00:49:24.707253 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:49:24.707409 kubelet[2802]: E0114 00:49:24.707366 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zkk9t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5f9b965548-xv9gd_calico-apiserver(32668954-0a97-46fe-9f01-fa91a530c2e3): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 14 00:49:24.710155 kubelet[2802]: E0114 00:49:24.709367 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" podUID="32668954-0a97-46fe-9f01-fa91a530c2e3" Jan 14 00:49:25.502484 kubelet[2802]: E0114 00:49:25.502446 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-p6wlv" podUID="976c1fb0-d44d-49e6-bc13-822966e59142" Jan 14 00:49:25.503434 kubelet[2802]: E0114 00:49:25.503403 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:49:27.503888 kubelet[2802]: E0114 00:49:27.503439 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-d6f8b64b4-m4ctk" podUID="db919709-47a0-46a4-a2ab-3f096374d888" Jan 14 00:49:35.500636 kubelet[2802]: E0114 00:49:35.500595 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:49:35.504456 kubelet[2802]: E0114 00:49:35.504277 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" podUID="32668954-0a97-46fe-9f01-fa91a530c2e3" Jan 14 00:49:35.506383 containerd[1611]: time="2026-01-14T00:49:35.504892080Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\"" Jan 14 00:49:35.645370 containerd[1611]: time="2026-01-14T00:49:35.645330453Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:49:35.646616 containerd[1611]: time="2026-01-14T00:49:35.646492338Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" Jan 14 00:49:35.646616 containerd[1611]: time="2026-01-14T00:49:35.646586859Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.4: active requests=0, bytes read=0" Jan 14 00:49:35.646895 kubelet[2802]: E0114 00:49:35.646854 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:49:35.646970 kubelet[2802]: E0114 00:49:35.646907 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" image="ghcr.io/flatcar/calico/apiserver:v3.30.4" Jan 14 00:49:35.647120 kubelet[2802]: E0114 00:49:35.647041 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-apiserver,Image:ghcr.io/flatcar/calico/apiserver:v3.30.4,Command:[],Args:[--secure-port=5443 --tls-private-key-file=/calico-apiserver-certs/tls.key --tls-cert-file=/calico-apiserver-certs/tls.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:LOG_LEVEL,Value:info,ValueFrom:nil,},EnvVar{Name:MULTI_INTERFACE_MODE,Value:none,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:calico-apiserver-certs,ReadOnly:true,MountPath:/calico-apiserver-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hmr4q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 5443 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-apiserver-5f9b965548-bgkj8_calico-apiserver(dcaab5f4-fd68-41e8-986b-14fa1a168bce): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/apiserver:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found" logger="UnhandledError" Jan 14 00:49:35.648414 kubelet[2802]: E0114 00:49:35.648381 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" podUID="dcaab5f4-fd68-41e8-986b-14fa1a168bce" Jan 14 00:49:36.218077 systemd[1]: Started sshd@7-172.236.110.177:22-68.220.241.50:38788.service - OpenSSH per-connection server daemon (68.220.241.50:38788). Jan 14 00:49:36.217000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-172.236.110.177:22-68.220.241.50:38788 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:36.225221 kernel: audit: type=1130 audit(1768351776.217:756): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-172.236.110.177:22-68.220.241.50:38788 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:36.389000 audit[4941]: USER_ACCT pid=4941 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:36.391449 sshd[4941]: Accepted publickey for core from 68.220.241.50 port 38788 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:49:36.398157 kernel: audit: type=1101 audit(1768351776.389:757): pid=4941 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:36.396000 audit[4941]: CRED_ACQ pid=4941 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:36.398919 sshd-session[4941]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:49:36.406301 kernel: audit: type=1103 audit(1768351776.396:758): pid=4941 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:36.406362 kernel: audit: type=1006 audit(1768351776.396:759): pid=4941 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=9 res=1 Jan 14 00:49:36.407621 systemd-logind[1578]: New session 9 of user core. Jan 14 00:49:36.396000 audit[4941]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc6dc5f730 a2=3 a3=0 items=0 ppid=1 pid=4941 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=9 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:36.411379 kernel: audit: type=1300 audit(1768351776.396:759): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffc6dc5f730 a2=3 a3=0 items=0 ppid=1 pid=4941 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=9 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:36.396000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:49:36.420238 kernel: audit: type=1327 audit(1768351776.396:759): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:49:36.421384 systemd[1]: Started session-9.scope - Session 9 of User core. Jan 14 00:49:36.424000 audit[4941]: USER_START pid=4941 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:36.435351 kernel: audit: type=1105 audit(1768351776.424:760): pid=4941 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:36.433000 audit[4945]: CRED_ACQ pid=4945 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:36.445142 kernel: audit: type=1103 audit(1768351776.433:761): pid=4945 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:36.501560 containerd[1611]: time="2026-01-14T00:49:36.501353582Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\"" Jan 14 00:49:36.568336 sshd[4945]: Connection closed by 68.220.241.50 port 38788 Jan 14 00:49:36.569957 sshd-session[4941]: pam_unix(sshd:session): session closed for user core Jan 14 00:49:36.573000 audit[4941]: USER_END pid=4941 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:36.584153 kernel: audit: type=1106 audit(1768351776.573:762): pid=4941 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:36.584426 systemd[1]: sshd@7-172.236.110.177:22-68.220.241.50:38788.service: Deactivated successfully. Jan 14 00:49:36.592152 kernel: audit: type=1104 audit(1768351776.573:763): pid=4941 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:36.573000 audit[4941]: CRED_DISP pid=4941 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:36.586847 systemd[1]: session-9.scope: Deactivated successfully. Jan 14 00:49:36.583000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-172.236.110.177:22-68.220.241.50:38788 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:36.593156 systemd-logind[1578]: Session 9 logged out. Waiting for processes to exit. Jan 14 00:49:36.595208 systemd-logind[1578]: Removed session 9. Jan 14 00:49:36.627274 containerd[1611]: time="2026-01-14T00:49:36.627225518Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:49:36.628201 containerd[1611]: time="2026-01-14T00:49:36.628167710Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" Jan 14 00:49:36.628272 containerd[1611]: time="2026-01-14T00:49:36.628239131Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.4: active requests=0, bytes read=0" Jan 14 00:49:36.628408 kubelet[2802]: E0114 00:49:36.628356 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 14 00:49:36.628408 kubelet[2802]: E0114 00:49:36.628393 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" image="ghcr.io/flatcar/calico/kube-controllers:v3.30.4" Jan 14 00:49:36.628739 kubelet[2802]: E0114 00:49:36.628682 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-kube-controllers,Image:ghcr.io/flatcar/calico/kube-controllers:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KUBE_CONTROLLERS_CONFIG_NAME,Value:default,ValueFrom:nil,},EnvVar{Name:DATASTORE_TYPE,Value:kubernetes,ValueFrom:nil,},EnvVar{Name:ENABLED_CONTROLLERS,Value:node,loadbalancer,ValueFrom:nil,},EnvVar{Name:DISABLE_KUBE_CONTROLLERS_CONFIG_API,Value:false,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:10.96.0.1,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:443,ValueFrom:nil,},EnvVar{Name:CA_CRT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tigera-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/cert.pem,SubPath:ca-bundle.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sfqdv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -l],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:10,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/check-status -r],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*999,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod calico-kube-controllers-5f75bc94c6-2nthl_calico-system(c12ac0b5-659e-4a97-821e-20b4aabfc024): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found" logger="UnhandledError" Jan 14 00:49:36.630141 kubelet[2802]: E0114 00:49:36.629957 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" podUID="c12ac0b5-659e-4a97-821e-20b4aabfc024" Jan 14 00:49:37.505864 containerd[1611]: time="2026-01-14T00:49:37.504696095Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\"" Jan 14 00:49:37.632743 containerd[1611]: time="2026-01-14T00:49:37.632697738Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:49:37.634228 containerd[1611]: time="2026-01-14T00:49:37.634091345Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" Jan 14 00:49:37.634228 containerd[1611]: time="2026-01-14T00:49:37.634194387Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.4: active requests=0, bytes read=0" Jan 14 00:49:37.634549 kubelet[2802]: E0114 00:49:37.634480 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 14 00:49:37.635453 kubelet[2802]: E0114 00:49:37.634949 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" image="ghcr.io/flatcar/calico/csi:v3.30.4" Jan 14 00:49:37.635453 kubelet[2802]: E0114 00:49:37.635197 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:calico-csi,Image:ghcr.io/flatcar/calico/csi:v3.30.4,Command:[],Args:[--nodeid=$(KUBE_NODE_NAME) --loglevel=$(LOG_LEVEL)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:warn,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:varrun,ReadOnly:false,MountPath:/var/run,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pnt5f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-vzqkb_calico-system(f41cc627-bd07-42f1-8e88-f85dcd2c063c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/csi:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found" logger="UnhandledError" Jan 14 00:49:37.636027 containerd[1611]: time="2026-01-14T00:49:37.635860047Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\"" Jan 14 00:49:37.761774 containerd[1611]: time="2026-01-14T00:49:37.761618262Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:49:37.763010 containerd[1611]: time="2026-01-14T00:49:37.762980389Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" Jan 14 00:49:37.763357 containerd[1611]: time="2026-01-14T00:49:37.763178792Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.4: active requests=0, bytes read=0" Jan 14 00:49:37.764326 kubelet[2802]: E0114 00:49:37.763797 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 14 00:49:37.764326 kubelet[2802]: E0114 00:49:37.763837 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" image="ghcr.io/flatcar/calico/goldmane:v3.30.4" Jan 14 00:49:37.764326 kubelet[2802]: E0114 00:49:37.764033 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.4,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5vwgj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-666569f655-p6wlv_calico-system(976c1fb0-d44d-49e6-bc13-822966e59142): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found" logger="UnhandledError" Jan 14 00:49:37.765245 containerd[1611]: time="2026-01-14T00:49:37.765178177Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\"" Jan 14 00:49:37.765399 kubelet[2802]: E0114 00:49:37.765326 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-p6wlv" podUID="976c1fb0-d44d-49e6-bc13-822966e59142" Jan 14 00:49:37.896016 containerd[1611]: time="2026-01-14T00:49:37.895957994Z" level=info msg="fetch failed after status: 404 Not Found" host=ghcr.io Jan 14 00:49:37.896981 containerd[1611]: time="2026-01-14T00:49:37.896874695Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\" failed" error="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" Jan 14 00:49:37.896981 containerd[1611]: time="2026-01-14T00:49:37.896946756Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: active requests=0, bytes read=0" Jan 14 00:49:37.897508 kubelet[2802]: E0114 00:49:37.897381 2802 log.go:32] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 14 00:49:37.897651 kubelet[2802]: E0114 00:49:37.897599 2802 kuberuntime_image.go:42] "Failed to pull image" err="rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" image="ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4" Jan 14 00:49:37.898154 kubelet[2802]: E0114 00:49:37.898083 2802 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:csi-node-driver-registrar,Image:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4,Command:[],Args:[--v=5 --csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH)],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/csi.tigera.io/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pnt5f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:*false,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod csi-node-driver-vzqkb_calico-system(f41cc627-bd07-42f1-8e88-f85dcd2c063c): ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found" logger="UnhandledError" Jan 14 00:49:37.899425 kubelet[2802]: E0114 00:49:37.899385 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ErrImagePull: \"rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:49:40.501403 kubelet[2802]: E0114 00:49:40.501271 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-d6f8b64b4-m4ctk" podUID="db919709-47a0-46a4-a2ab-3f096374d888" Jan 14 00:49:41.609565 systemd[1]: Started sshd@8-172.236.110.177:22-68.220.241.50:38804.service - OpenSSH per-connection server daemon (68.220.241.50:38804). Jan 14 00:49:41.615182 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 14 00:49:41.615295 kernel: audit: type=1130 audit(1768351781.608:765): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-172.236.110.177:22-68.220.241.50:38804 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:41.608000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-172.236.110.177:22-68.220.241.50:38804 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:41.793000 audit[4984]: USER_ACCT pid=4984 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:41.796509 sshd[4984]: Accepted publickey for core from 68.220.241.50 port 38804 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:49:41.798414 sshd-session[4984]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:49:41.802156 kernel: audit: type=1101 audit(1768351781.793:766): pid=4984 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:41.795000 audit[4984]: CRED_ACQ pid=4984 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:41.808037 systemd-logind[1578]: New session 10 of user core. Jan 14 00:49:41.810665 kernel: audit: type=1103 audit(1768351781.795:767): pid=4984 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:41.810718 kernel: audit: type=1006 audit(1768351781.795:768): pid=4984 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=10 res=1 Jan 14 00:49:41.795000 audit[4984]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd3ab005c0 a2=3 a3=0 items=0 ppid=1 pid=4984 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:41.815599 kernel: audit: type=1300 audit(1768351781.795:768): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd3ab005c0 a2=3 a3=0 items=0 ppid=1 pid=4984 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:41.795000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:49:41.822817 kernel: audit: type=1327 audit(1768351781.795:768): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:49:41.823275 systemd[1]: Started session-10.scope - Session 10 of User core. Jan 14 00:49:41.826000 audit[4984]: USER_START pid=4984 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:41.836197 kernel: audit: type=1105 audit(1768351781.826:769): pid=4984 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:41.828000 audit[4988]: CRED_ACQ pid=4988 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:41.845155 kernel: audit: type=1103 audit(1768351781.828:770): pid=4988 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:41.954369 sshd[4988]: Connection closed by 68.220.241.50 port 38804 Jan 14 00:49:41.954914 sshd-session[4984]: pam_unix(sshd:session): session closed for user core Jan 14 00:49:41.957000 audit[4984]: USER_END pid=4984 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:41.964978 systemd[1]: sshd@8-172.236.110.177:22-68.220.241.50:38804.service: Deactivated successfully. Jan 14 00:49:41.968296 kernel: audit: type=1106 audit(1768351781.957:771): pid=4984 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:41.968337 kernel: audit: type=1104 audit(1768351781.957:772): pid=4984 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:41.957000 audit[4984]: CRED_DISP pid=4984 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:41.969364 systemd[1]: session-10.scope: Deactivated successfully. Jan 14 00:49:41.965000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-172.236.110.177:22-68.220.241.50:38804 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:41.975275 systemd-logind[1578]: Session 10 logged out. Waiting for processes to exit. Jan 14 00:49:41.979817 systemd-logind[1578]: Removed session 10. Jan 14 00:49:46.501620 kubelet[2802]: E0114 00:49:46.501436 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" podUID="32668954-0a97-46fe-9f01-fa91a530c2e3" Jan 14 00:49:46.504334 kubelet[2802]: E0114 00:49:46.503470 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" podUID="dcaab5f4-fd68-41e8-986b-14fa1a168bce" Jan 14 00:49:46.984000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-172.236.110.177:22-68.220.241.50:51842 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:46.985361 systemd[1]: Started sshd@9-172.236.110.177:22-68.220.241.50:51842.service - OpenSSH per-connection server daemon (68.220.241.50:51842). Jan 14 00:49:46.987775 kernel: kauditd_printk_skb: 1 callbacks suppressed Jan 14 00:49:46.987822 kernel: audit: type=1130 audit(1768351786.984:774): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-172.236.110.177:22-68.220.241.50:51842 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:47.145501 sshd[5000]: Accepted publickey for core from 68.220.241.50 port 51842 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:49:47.154132 kernel: audit: type=1101 audit(1768351787.144:775): pid=5000 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.144000 audit[5000]: USER_ACCT pid=5000 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.155593 sshd-session[5000]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:49:47.153000 audit[5000]: CRED_ACQ pid=5000 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.189661 kernel: audit: type=1103 audit(1768351787.153:776): pid=5000 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.189763 kernel: audit: type=1006 audit(1768351787.153:777): pid=5000 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=11 res=1 Jan 14 00:49:47.153000 audit[5000]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff78e83ca0 a2=3 a3=0 items=0 ppid=1 pid=5000 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=11 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:47.153000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:49:47.202519 kernel: audit: type=1300 audit(1768351787.153:777): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff78e83ca0 a2=3 a3=0 items=0 ppid=1 pid=5000 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=11 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:47.202563 kernel: audit: type=1327 audit(1768351787.153:777): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:49:47.208687 systemd-logind[1578]: New session 11 of user core. Jan 14 00:49:47.213364 systemd[1]: Started session-11.scope - Session 11 of User core. Jan 14 00:49:47.216000 audit[5000]: USER_START pid=5000 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.227154 kernel: audit: type=1105 audit(1768351787.216:778): pid=5000 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.226000 audit[5004]: CRED_ACQ pid=5004 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.240143 kernel: audit: type=1103 audit(1768351787.226:779): pid=5004 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.350887 sshd[5004]: Connection closed by 68.220.241.50 port 51842 Jan 14 00:49:47.352684 sshd-session[5000]: pam_unix(sshd:session): session closed for user core Jan 14 00:49:47.354000 audit[5000]: USER_END pid=5000 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.361242 systemd[1]: sshd@9-172.236.110.177:22-68.220.241.50:51842.service: Deactivated successfully. Jan 14 00:49:47.365316 systemd[1]: session-11.scope: Deactivated successfully. Jan 14 00:49:47.365920 systemd-logind[1578]: Session 11 logged out. Waiting for processes to exit. Jan 14 00:49:47.366326 kernel: audit: type=1106 audit(1768351787.354:780): pid=5000 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.366372 kernel: audit: type=1104 audit(1768351787.354:781): pid=5000 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.354000 audit[5000]: CRED_DISP pid=5000 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.358000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-172.236.110.177:22-68.220.241.50:51842 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:47.385695 systemd-logind[1578]: Removed session 11. Jan 14 00:49:47.387251 systemd[1]: Started sshd@10-172.236.110.177:22-68.220.241.50:51848.service - OpenSSH per-connection server daemon (68.220.241.50:51848). Jan 14 00:49:47.386000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-172.236.110.177:22-68.220.241.50:51848 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:47.541000 audit[5016]: USER_ACCT pid=5016 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.542655 sshd[5016]: Accepted publickey for core from 68.220.241.50 port 51848 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:49:47.542000 audit[5016]: CRED_ACQ pid=5016 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.542000 audit[5016]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffdeb650320 a2=3 a3=0 items=0 ppid=1 pid=5016 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=12 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:47.542000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:49:47.545370 sshd-session[5016]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:49:47.552478 systemd-logind[1578]: New session 12 of user core. Jan 14 00:49:47.557247 systemd[1]: Started session-12.scope - Session 12 of User core. Jan 14 00:49:47.561000 audit[5016]: USER_START pid=5016 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.563000 audit[5023]: CRED_ACQ pid=5023 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.724875 sshd[5023]: Connection closed by 68.220.241.50 port 51848 Jan 14 00:49:47.723953 sshd-session[5016]: pam_unix(sshd:session): session closed for user core Jan 14 00:49:47.724000 audit[5016]: USER_END pid=5016 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.724000 audit[5016]: CRED_DISP pid=5016 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.732760 systemd-logind[1578]: Session 12 logged out. Waiting for processes to exit. Jan 14 00:49:47.733885 systemd[1]: sshd@10-172.236.110.177:22-68.220.241.50:51848.service: Deactivated successfully. Jan 14 00:49:47.733000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-172.236.110.177:22-68.220.241.50:51848 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:47.736865 systemd[1]: session-12.scope: Deactivated successfully. Jan 14 00:49:47.741470 systemd-logind[1578]: Removed session 12. Jan 14 00:49:47.758646 systemd[1]: Started sshd@11-172.236.110.177:22-68.220.241.50:51862.service - OpenSSH per-connection server daemon (68.220.241.50:51862). Jan 14 00:49:47.757000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-172.236.110.177:22-68.220.241.50:51862 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:47.933000 audit[5034]: USER_ACCT pid=5034 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.936144 sshd[5034]: Accepted publickey for core from 68.220.241.50 port 51862 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:49:47.935000 audit[5034]: CRED_ACQ pid=5034 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.936000 audit[5034]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd4a1f89c0 a2=3 a3=0 items=0 ppid=1 pid=5034 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=13 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:47.936000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:49:47.939510 sshd-session[5034]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:49:47.946844 systemd-logind[1578]: New session 13 of user core. Jan 14 00:49:47.956282 systemd[1]: Started session-13.scope - Session 13 of User core. Jan 14 00:49:47.959000 audit[5034]: USER_START pid=5034 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:47.961000 audit[5038]: CRED_ACQ pid=5038 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:48.127933 sshd[5038]: Connection closed by 68.220.241.50 port 51862 Jan 14 00:49:48.127537 sshd-session[5034]: pam_unix(sshd:session): session closed for user core Jan 14 00:49:48.130000 audit[5034]: USER_END pid=5034 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:48.130000 audit[5034]: CRED_DISP pid=5034 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:48.135714 systemd[1]: sshd@11-172.236.110.177:22-68.220.241.50:51862.service: Deactivated successfully. Jan 14 00:49:48.135746 systemd-logind[1578]: Session 13 logged out. Waiting for processes to exit. Jan 14 00:49:48.136000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-172.236.110.177:22-68.220.241.50:51862 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:48.140428 systemd[1]: session-13.scope: Deactivated successfully. Jan 14 00:49:48.146488 systemd-logind[1578]: Removed session 13. Jan 14 00:49:51.507575 kubelet[2802]: E0114 00:49:51.507380 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" podUID="c12ac0b5-659e-4a97-821e-20b4aabfc024" Jan 14 00:49:51.509194 kubelet[2802]: E0114 00:49:51.507491 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-p6wlv" podUID="976c1fb0-d44d-49e6-bc13-822966e59142" Jan 14 00:49:51.510368 kubelet[2802]: E0114 00:49:51.510251 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:49:51.510926 kubelet[2802]: E0114 00:49:51.510879 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-d6f8b64b4-m4ctk" podUID="db919709-47a0-46a4-a2ab-3f096374d888" Jan 14 00:49:53.174510 kernel: kauditd_printk_skb: 23 callbacks suppressed Jan 14 00:49:53.175753 kernel: audit: type=1130 audit(1768351793.167:801): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-172.236.110.177:22-68.220.241.50:38238 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:53.167000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-172.236.110.177:22-68.220.241.50:38238 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:53.168363 systemd[1]: Started sshd@12-172.236.110.177:22-68.220.241.50:38238.service - OpenSSH per-connection server daemon (68.220.241.50:38238). Jan 14 00:49:53.348000 audit[5050]: USER_ACCT pid=5050 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.351552 sshd[5050]: Accepted publickey for core from 68.220.241.50 port 38238 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:49:53.355622 sshd-session[5050]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:49:53.358166 kernel: audit: type=1101 audit(1768351793.348:802): pid=5050 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.350000 audit[5050]: CRED_ACQ pid=5050 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.366261 systemd-logind[1578]: New session 14 of user core. Jan 14 00:49:53.369143 kernel: audit: type=1103 audit(1768351793.350:803): pid=5050 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.375239 systemd[1]: Started session-14.scope - Session 14 of User core. Jan 14 00:49:53.377725 kernel: audit: type=1006 audit(1768351793.350:804): pid=5050 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=14 res=1 Jan 14 00:49:53.350000 audit[5050]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe14dc5170 a2=3 a3=0 items=0 ppid=1 pid=5050 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:53.386129 kernel: audit: type=1300 audit(1768351793.350:804): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffe14dc5170 a2=3 a3=0 items=0 ppid=1 pid=5050 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:53.350000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:49:53.393129 kernel: audit: type=1327 audit(1768351793.350:804): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:49:53.387000 audit[5050]: USER_START pid=5050 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.404134 kernel: audit: type=1105 audit(1768351793.387:805): pid=5050 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.392000 audit[5054]: CRED_ACQ pid=5054 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.418140 kernel: audit: type=1103 audit(1768351793.392:806): pid=5054 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.522210 sshd[5054]: Connection closed by 68.220.241.50 port 38238 Jan 14 00:49:53.522727 sshd-session[5050]: pam_unix(sshd:session): session closed for user core Jan 14 00:49:53.534598 kernel: audit: type=1106 audit(1768351793.523:807): pid=5050 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.523000 audit[5050]: USER_END pid=5050 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.543561 kernel: audit: type=1104 audit(1768351793.532:808): pid=5050 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.532000 audit[5050]: CRED_DISP pid=5050 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.537184 systemd[1]: sshd@12-172.236.110.177:22-68.220.241.50:38238.service: Deactivated successfully. Jan 14 00:49:53.542502 systemd[1]: session-14.scope: Deactivated successfully. Jan 14 00:49:53.535000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-172.236.110.177:22-68.220.241.50:38238 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:53.545290 systemd-logind[1578]: Session 14 logged out. Waiting for processes to exit. Jan 14 00:49:53.556749 systemd-logind[1578]: Removed session 14. Jan 14 00:49:53.560000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-172.236.110.177:22-68.220.241.50:38244 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:53.560393 systemd[1]: Started sshd@13-172.236.110.177:22-68.220.241.50:38244.service - OpenSSH per-connection server daemon (68.220.241.50:38244). Jan 14 00:49:53.738000 audit[5066]: USER_ACCT pid=5066 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.740808 sshd[5066]: Accepted publickey for core from 68.220.241.50 port 38244 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:49:53.741000 audit[5066]: CRED_ACQ pid=5066 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.741000 audit[5066]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7fff9d53ffc0 a2=3 a3=0 items=0 ppid=1 pid=5066 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:53.741000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:49:53.744751 sshd-session[5066]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:49:53.751683 systemd-logind[1578]: New session 15 of user core. Jan 14 00:49:53.758253 systemd[1]: Started session-15.scope - Session 15 of User core. Jan 14 00:49:53.762000 audit[5066]: USER_START pid=5066 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:53.764000 audit[5070]: CRED_ACQ pid=5070 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:54.033649 sshd[5070]: Connection closed by 68.220.241.50 port 38244 Jan 14 00:49:54.035317 sshd-session[5066]: pam_unix(sshd:session): session closed for user core Jan 14 00:49:54.038000 audit[5066]: USER_END pid=5066 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:54.038000 audit[5066]: CRED_DISP pid=5066 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:54.042898 systemd[1]: sshd@13-172.236.110.177:22-68.220.241.50:38244.service: Deactivated successfully. Jan 14 00:49:54.042000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-172.236.110.177:22-68.220.241.50:38244 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:54.045778 systemd[1]: session-15.scope: Deactivated successfully. Jan 14 00:49:54.047480 systemd-logind[1578]: Session 15 logged out. Waiting for processes to exit. Jan 14 00:49:54.049783 systemd-logind[1578]: Removed session 15. Jan 14 00:49:54.069000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-172.236.110.177:22-68.220.241.50:38252 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:54.070578 systemd[1]: Started sshd@14-172.236.110.177:22-68.220.241.50:38252.service - OpenSSH per-connection server daemon (68.220.241.50:38252). Jan 14 00:49:54.242000 audit[5080]: USER_ACCT pid=5080 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:54.244334 sshd[5080]: Accepted publickey for core from 68.220.241.50 port 38252 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:49:54.244000 audit[5080]: CRED_ACQ pid=5080 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:54.244000 audit[5080]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffeff238170 a2=3 a3=0 items=0 ppid=1 pid=5080 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:54.244000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:49:54.247012 sshd-session[5080]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:49:54.253931 systemd-logind[1578]: New session 16 of user core. Jan 14 00:49:54.258269 systemd[1]: Started session-16.scope - Session 16 of User core. Jan 14 00:49:54.260000 audit[5080]: USER_START pid=5080 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:54.263000 audit[5084]: CRED_ACQ pid=5084 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.090000 audit[5094]: NETFILTER_CFG table=filter:142 family=2 entries=26 op=nft_register_rule pid=5094 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:49:55.090000 audit[5094]: SYSCALL arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7ffd1451c5f0 a2=0 a3=7ffd1451c5dc items=0 ppid=2943 pid=5094 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:55.090000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:49:55.096000 audit[5094]: NETFILTER_CFG table=nat:143 family=2 entries=20 op=nft_register_rule pid=5094 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:49:55.107291 sshd[5084]: Connection closed by 68.220.241.50 port 38252 Jan 14 00:49:55.096000 audit[5094]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffd1451c5f0 a2=0 a3=0 items=0 ppid=2943 pid=5094 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:55.096000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:49:55.107823 sshd-session[5080]: pam_unix(sshd:session): session closed for user core Jan 14 00:49:55.110000 audit[5080]: USER_END pid=5080 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.110000 audit[5080]: CRED_DISP pid=5080 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.114905 systemd[1]: sshd@14-172.236.110.177:22-68.220.241.50:38252.service: Deactivated successfully. Jan 14 00:49:55.115000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-172.236.110.177:22-68.220.241.50:38252 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:55.120440 systemd[1]: session-16.scope: Deactivated successfully. Jan 14 00:49:55.124579 systemd-logind[1578]: Session 16 logged out. Waiting for processes to exit. Jan 14 00:49:55.126212 systemd-logind[1578]: Removed session 16. Jan 14 00:49:55.126000 audit[5097]: NETFILTER_CFG table=filter:144 family=2 entries=38 op=nft_register_rule pid=5097 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:49:55.126000 audit[5097]: SYSCALL arch=c000003e syscall=46 success=yes exit=14176 a0=3 a1=7ffecd8dc5a0 a2=0 a3=7ffecd8dc58c items=0 ppid=2943 pid=5097 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:55.126000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:49:55.133000 audit[5097]: NETFILTER_CFG table=nat:145 family=2 entries=20 op=nft_register_rule pid=5097 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:49:55.133000 audit[5097]: SYSCALL arch=c000003e syscall=46 success=yes exit=5772 a0=3 a1=7ffecd8dc5a0 a2=0 a3=0 items=0 ppid=2943 pid=5097 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:55.133000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:49:55.145393 systemd[1]: Started sshd@15-172.236.110.177:22-68.220.241.50:38256.service - OpenSSH per-connection server daemon (68.220.241.50:38256). Jan 14 00:49:55.144000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-172.236.110.177:22-68.220.241.50:38256 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:55.307000 audit[5101]: USER_ACCT pid=5101 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.308800 sshd[5101]: Accepted publickey for core from 68.220.241.50 port 38256 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:49:55.308000 audit[5101]: CRED_ACQ pid=5101 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.309000 audit[5101]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffee047a9e0 a2=3 a3=0 items=0 ppid=1 pid=5101 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=17 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:55.309000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:49:55.311421 sshd-session[5101]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:49:55.318993 systemd-logind[1578]: New session 17 of user core. Jan 14 00:49:55.324755 systemd[1]: Started session-17.scope - Session 17 of User core. Jan 14 00:49:55.328000 audit[5101]: USER_START pid=5101 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.332000 audit[5105]: CRED_ACQ pid=5105 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.591155 sshd[5105]: Connection closed by 68.220.241.50 port 38256 Jan 14 00:49:55.593305 sshd-session[5101]: pam_unix(sshd:session): session closed for user core Jan 14 00:49:55.594000 audit[5101]: USER_END pid=5101 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.594000 audit[5101]: CRED_DISP pid=5101 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.600071 systemd[1]: sshd@15-172.236.110.177:22-68.220.241.50:38256.service: Deactivated successfully. Jan 14 00:49:55.602000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-172.236.110.177:22-68.220.241.50:38256 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:55.607883 systemd[1]: session-17.scope: Deactivated successfully. Jan 14 00:49:55.611252 systemd-logind[1578]: Session 17 logged out. Waiting for processes to exit. Jan 14 00:49:55.628475 systemd[1]: Started sshd@16-172.236.110.177:22-68.220.241.50:38264.service - OpenSSH per-connection server daemon (68.220.241.50:38264). Jan 14 00:49:55.627000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-172.236.110.177:22-68.220.241.50:38264 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:55.630883 systemd-logind[1578]: Removed session 17. Jan 14 00:49:55.792000 audit[5115]: USER_ACCT pid=5115 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.793969 sshd[5115]: Accepted publickey for core from 68.220.241.50 port 38264 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:49:55.793000 audit[5115]: CRED_ACQ pid=5115 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.793000 audit[5115]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffefd94a3b0 a2=3 a3=0 items=0 ppid=1 pid=5115 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=18 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:49:55.793000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:49:55.796383 sshd-session[5115]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:49:55.802358 systemd-logind[1578]: New session 18 of user core. Jan 14 00:49:55.806256 systemd[1]: Started session-18.scope - Session 18 of User core. Jan 14 00:49:55.809000 audit[5115]: USER_START pid=5115 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.811000 audit[5119]: CRED_ACQ pid=5119 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.925466 sshd[5119]: Connection closed by 68.220.241.50 port 38264 Jan 14 00:49:55.926343 sshd-session[5115]: pam_unix(sshd:session): session closed for user core Jan 14 00:49:55.927000 audit[5115]: USER_END pid=5115 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.927000 audit[5115]: CRED_DISP pid=5115 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:49:55.931334 systemd[1]: sshd@16-172.236.110.177:22-68.220.241.50:38264.service: Deactivated successfully. Jan 14 00:49:55.930000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-172.236.110.177:22-68.220.241.50:38264 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:49:55.936973 systemd[1]: session-18.scope: Deactivated successfully. Jan 14 00:49:55.938683 systemd-logind[1578]: Session 18 logged out. Waiting for processes to exit. Jan 14 00:49:55.942782 systemd-logind[1578]: Removed session 18. Jan 14 00:49:58.504328 kubelet[2802]: E0114 00:49:58.500447 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-bgkj8" podUID="dcaab5f4-fd68-41e8-986b-14fa1a168bce" Jan 14 00:49:58.506444 kubelet[2802]: E0114 00:49:58.504413 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-apiserver\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/apiserver:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/apiserver:v3.30.4: not found\"" pod="calico-apiserver/calico-apiserver-5f9b965548-xv9gd" podUID="32668954-0a97-46fe-9f01-fa91a530c2e3" Jan 14 00:50:00.074000 audit[5133]: NETFILTER_CFG table=filter:146 family=2 entries=26 op=nft_register_rule pid=5133 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:50:00.078007 kernel: kauditd_printk_skb: 57 callbacks suppressed Jan 14 00:50:00.078081 kernel: audit: type=1325 audit(1768351800.074:850): table=filter:146 family=2 entries=26 op=nft_register_rule pid=5133 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:50:00.074000 audit[5133]: SYSCALL arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffc426bcf10 a2=0 a3=7ffc426bcefc items=0 ppid=2943 pid=5133 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:50:00.083922 kernel: audit: type=1300 audit(1768351800.074:850): arch=c000003e syscall=46 success=yes exit=5248 a0=3 a1=7ffc426bcf10 a2=0 a3=7ffc426bcefc items=0 ppid=2943 pid=5133 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:50:00.074000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:50:00.091243 kernel: audit: type=1327 audit(1768351800.074:850): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:50:00.086000 audit[5133]: NETFILTER_CFG table=nat:147 family=2 entries=104 op=nft_register_chain pid=5133 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:50:00.095378 kernel: audit: type=1325 audit(1768351800.086:851): table=nat:147 family=2 entries=104 op=nft_register_chain pid=5133 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Jan 14 00:50:00.086000 audit[5133]: SYSCALL arch=c000003e syscall=46 success=yes exit=48684 a0=3 a1=7ffc426bcf10 a2=0 a3=7ffc426bcefc items=0 ppid=2943 pid=5133 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:50:00.100225 kernel: audit: type=1300 audit(1768351800.086:851): arch=c000003e syscall=46 success=yes exit=48684 a0=3 a1=7ffc426bcf10 a2=0 a3=7ffc426bcefc items=0 ppid=2943 pid=5133 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:50:00.086000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:50:00.107843 kernel: audit: type=1327 audit(1768351800.086:851): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Jan 14 00:50:00.970336 systemd[1]: Started sshd@17-172.236.110.177:22-68.220.241.50:38266.service - OpenSSH per-connection server daemon (68.220.241.50:38266). Jan 14 00:50:00.969000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-172.236.110.177:22-68.220.241.50:38266 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:50:00.978145 kernel: audit: type=1130 audit(1768351800.969:852): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-172.236.110.177:22-68.220.241.50:38266 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:50:01.162000 audit[5135]: USER_ACCT pid=5135 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:01.164166 sshd[5135]: Accepted publickey for core from 68.220.241.50 port 38266 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:50:01.171149 kernel: audit: type=1101 audit(1768351801.162:853): pid=5135 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:01.172964 sshd-session[5135]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:50:01.170000 audit[5135]: CRED_ACQ pid=5135 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:01.184139 kernel: audit: type=1103 audit(1768351801.170:854): pid=5135 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:01.192590 kernel: audit: type=1006 audit(1768351801.170:855): pid=5135 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=19 res=1 Jan 14 00:50:01.192227 systemd-logind[1578]: New session 19 of user core. Jan 14 00:50:01.170000 audit[5135]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffd92133070 a2=3 a3=0 items=0 ppid=1 pid=5135 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=19 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:50:01.170000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:50:01.195676 systemd[1]: Started session-19.scope - Session 19 of User core. Jan 14 00:50:01.199000 audit[5135]: USER_START pid=5135 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:01.201000 audit[5139]: CRED_ACQ pid=5139 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:01.327104 sshd[5139]: Connection closed by 68.220.241.50 port 38266 Jan 14 00:50:01.329795 sshd-session[5135]: pam_unix(sshd:session): session closed for user core Jan 14 00:50:01.330000 audit[5135]: USER_END pid=5135 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:01.330000 audit[5135]: CRED_DISP pid=5135 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:01.334265 systemd[1]: sshd@17-172.236.110.177:22-68.220.241.50:38266.service: Deactivated successfully. Jan 14 00:50:01.334000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-172.236.110.177:22-68.220.241.50:38266 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:50:01.334850 systemd-logind[1578]: Session 19 logged out. Waiting for processes to exit. Jan 14 00:50:01.337916 systemd[1]: session-19.scope: Deactivated successfully. Jan 14 00:50:01.342802 systemd-logind[1578]: Removed session 19. Jan 14 00:50:03.502674 kubelet[2802]: E0114 00:50:03.502624 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-kube-controllers\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/kube-controllers:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/kube-controllers:v3.30.4: not found\"" pod="calico-system/calico-kube-controllers-5f75bc94c6-2nthl" podUID="c12ac0b5-659e-4a97-821e-20b4aabfc024" Jan 14 00:50:03.503365 kubelet[2802]: E0114 00:50:03.503322 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker:v3.30.4: not found\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/whisker-backend:v3.30.4: not found\"]" pod="calico-system/whisker-d6f8b64b4-m4ctk" podUID="db919709-47a0-46a4-a2ab-3f096374d888" Jan 14 00:50:04.499842 kubelet[2802]: E0114 00:50:04.499492 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18" Jan 14 00:50:04.502216 kubelet[2802]: E0114 00:50:04.501960 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/goldmane:v3.30.4: not found\"" pod="calico-system/goldmane-666569f655-p6wlv" podUID="976c1fb0-d44d-49e6-bc13-822966e59142" Jan 14 00:50:05.504076 kubelet[2802]: E0114 00:50:05.504008 2802 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"calico-csi\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/csi:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/csi:v3.30.4: not found\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": ErrImagePull: rpc error: code = NotFound desc = failed to pull and unpack image \\\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4\\\": failed to resolve image: ghcr.io/flatcar/calico/node-driver-registrar:v3.30.4: not found\"]" pod="calico-system/csi-node-driver-vzqkb" podUID="f41cc627-bd07-42f1-8e88-f85dcd2c063c" Jan 14 00:50:06.365234 kernel: kauditd_printk_skb: 7 callbacks suppressed Jan 14 00:50:06.365429 kernel: audit: type=1130 audit(1768351806.357:861): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-172.236.110.177:22-68.220.241.50:33452 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:50:06.357000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-172.236.110.177:22-68.220.241.50:33452 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:50:06.357346 systemd[1]: Started sshd@18-172.236.110.177:22-68.220.241.50:33452.service - OpenSSH per-connection server daemon (68.220.241.50:33452). Jan 14 00:50:06.515000 audit[5151]: USER_ACCT pid=5151 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:06.524133 kernel: audit: type=1101 audit(1768351806.515:862): pid=5151 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_time,pam_unix,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:06.524325 sshd[5151]: Accepted publickey for core from 68.220.241.50 port 33452 ssh2: RSA SHA256:Q5X4hp1XZh1OSKmHGw8mk8d/XuB/6jjevFNn+X5lSfs Jan 14 00:50:06.527047 sshd-session[5151]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 14 00:50:06.525000 audit[5151]: CRED_ACQ pid=5151 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:06.535185 kernel: audit: type=1103 audit(1768351806.525:863): pid=5151 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:06.541401 kernel: audit: type=1006 audit(1768351806.525:864): pid=5151 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=20 res=1 Jan 14 00:50:06.545385 systemd-logind[1578]: New session 20 of user core. Jan 14 00:50:06.525000 audit[5151]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffce5eb4660 a2=3 a3=0 items=0 ppid=1 pid=5151 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=20 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:50:06.554168 kernel: audit: type=1300 audit(1768351806.525:864): arch=c000003e syscall=1 success=yes exit=3 a0=8 a1=7ffce5eb4660 a2=3 a3=0 items=0 ppid=1 pid=5151 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=20 comm="sshd-session" exe="/usr/lib64/misc/sshd-session" subj=system_u:system_r:kernel_t:s0 key=(null) Jan 14 00:50:06.555846 systemd[1]: Started session-20.scope - Session 20 of User core. Jan 14 00:50:06.525000 audit: PROCTITLE proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:50:06.562138 kernel: audit: type=1327 audit(1768351806.525:864): proctitle=737368642D73657373696F6E3A20636F7265205B707269765D Jan 14 00:50:06.563000 audit[5151]: USER_START pid=5151 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:06.565000 audit[5155]: CRED_ACQ pid=5155 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:06.576398 kernel: audit: type=1105 audit(1768351806.563:865): pid=5151 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:06.576460 kernel: audit: type=1103 audit(1768351806.565:866): pid=5155 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:06.691798 sshd[5155]: Connection closed by 68.220.241.50 port 33452 Jan 14 00:50:06.692340 sshd-session[5151]: pam_unix(sshd:session): session closed for user core Jan 14 00:50:06.695000 audit[5151]: USER_END pid=5151 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:06.707160 kernel: audit: type=1106 audit(1768351806.695:867): pid=5151 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_namespace,pam_keyinit,pam_limits,pam_env,pam_umask,pam_unix,pam_systemd,pam_lastlog,pam_mail acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:06.708492 systemd-logind[1578]: Session 20 logged out. Waiting for processes to exit. Jan 14 00:50:06.695000 audit[5151]: CRED_DISP pid=5151 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:06.716225 systemd[1]: sshd@18-172.236.110.177:22-68.220.241.50:33452.service: Deactivated successfully. Jan 14 00:50:06.720182 kernel: audit: type=1104 audit(1768351806.695:868): pid=5151 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock acct="core" exe="/usr/lib64/misc/sshd-session" hostname=68.220.241.50 addr=68.220.241.50 terminal=ssh res=success' Jan 14 00:50:06.720509 systemd[1]: session-20.scope: Deactivated successfully. Jan 14 00:50:06.717000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-172.236.110.177:22-68.220.241.50:33452 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jan 14 00:50:06.725263 systemd-logind[1578]: Removed session 20. Jan 14 00:50:08.499537 kubelet[2802]: E0114 00:50:08.499490 2802 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.20 172.232.0.15 172.232.0.18"