May 13 00:14:36.748987 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] May 13 00:14:36.749007 kernel: Linux version 5.15.181-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP PREEMPT Mon May 12 23:22:00 -00 2025 May 13 00:14:36.749015 kernel: efi: EFI v2.70 by EDK II May 13 00:14:36.749021 kernel: efi: SMBIOS 3.0=0xd9260000 ACPI 2.0=0xd9240000 MEMATTR=0xda32b018 RNG=0xd9220018 MEMRESERVE=0xd9521c18 May 13 00:14:36.749026 kernel: random: crng init done May 13 00:14:36.749031 kernel: ACPI: Early table checksum verification disabled May 13 00:14:36.749037 kernel: ACPI: RSDP 0x00000000D9240000 000024 (v02 BOCHS ) May 13 00:14:36.749044 kernel: ACPI: XSDT 0x00000000D9230000 000064 (v01 BOCHS BXPC 00000001 01000013) May 13 00:14:36.749050 kernel: ACPI: FACP 0x00000000D91E0000 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) May 13 00:14:36.749055 kernel: ACPI: DSDT 0x00000000D91F0000 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) May 13 00:14:36.749060 kernel: ACPI: APIC 0x00000000D91D0000 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) May 13 00:14:36.749065 kernel: ACPI: PPTT 0x00000000D91C0000 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) May 13 00:14:36.749071 kernel: ACPI: GTDT 0x00000000D91B0000 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) May 13 00:14:36.749076 kernel: ACPI: MCFG 0x00000000D91A0000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) May 13 00:14:36.749084 kernel: ACPI: SPCR 0x00000000D9190000 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) May 13 00:14:36.749090 kernel: ACPI: DBG2 0x00000000D9180000 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) May 13 00:14:36.749095 kernel: ACPI: IORT 0x00000000D9170000 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) May 13 00:14:36.749101 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 May 13 00:14:36.749107 kernel: NUMA: Failed to initialise from firmware May 13 00:14:36.749112 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] May 13 00:14:36.749118 kernel: NUMA: NODE_DATA [mem 0xdcb0a900-0xdcb0ffff] May 13 00:14:36.749124 kernel: Zone ranges: May 13 00:14:36.749129 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] May 13 00:14:36.749136 kernel: DMA32 empty May 13 00:14:36.749141 kernel: Normal empty May 13 00:14:36.749147 kernel: Movable zone start for each node May 13 00:14:36.749152 kernel: Early memory node ranges May 13 00:14:36.749158 kernel: node 0: [mem 0x0000000040000000-0x00000000d924ffff] May 13 00:14:36.749164 kernel: node 0: [mem 0x00000000d9250000-0x00000000d951ffff] May 13 00:14:36.749169 kernel: node 0: [mem 0x00000000d9520000-0x00000000dc7fffff] May 13 00:14:36.749175 kernel: node 0: [mem 0x00000000dc800000-0x00000000dc88ffff] May 13 00:14:36.749180 kernel: node 0: [mem 0x00000000dc890000-0x00000000dc89ffff] May 13 00:14:36.749186 kernel: node 0: [mem 0x00000000dc8a0000-0x00000000dc9bffff] May 13 00:14:36.749191 kernel: node 0: [mem 0x00000000dc9c0000-0x00000000dcffffff] May 13 00:14:36.749197 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] May 13 00:14:36.749204 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges May 13 00:14:36.749210 kernel: psci: probing for conduit method from ACPI. May 13 00:14:36.749215 kernel: psci: PSCIv1.1 detected in firmware. May 13 00:14:36.749221 kernel: psci: Using standard PSCI v0.2 function IDs May 13 00:14:36.749227 kernel: psci: Trusted OS migration not required May 13 00:14:36.749235 kernel: psci: SMC Calling Convention v1.1 May 13 00:14:36.749241 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) May 13 00:14:36.749249 kernel: ACPI: SRAT not present May 13 00:14:36.749255 kernel: percpu: Embedded 30 pages/cpu s83032 r8192 d31656 u122880 May 13 00:14:36.749261 kernel: pcpu-alloc: s83032 r8192 d31656 u122880 alloc=30*4096 May 13 00:14:36.749267 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 May 13 00:14:36.749273 kernel: Detected PIPT I-cache on CPU0 May 13 00:14:36.749279 kernel: CPU features: detected: GIC system register CPU interface May 13 00:14:36.749285 kernel: CPU features: detected: Hardware dirty bit management May 13 00:14:36.749291 kernel: CPU features: detected: Spectre-v4 May 13 00:14:36.749297 kernel: CPU features: detected: Spectre-BHB May 13 00:14:36.749304 kernel: CPU features: kernel page table isolation forced ON by KASLR May 13 00:14:36.749310 kernel: CPU features: detected: Kernel page table isolation (KPTI) May 13 00:14:36.749316 kernel: CPU features: detected: ARM erratum 1418040 May 13 00:14:36.749322 kernel: CPU features: detected: SSBS not fully self-synchronizing May 13 00:14:36.749328 kernel: Built 1 zonelists, mobility grouping on. Total pages: 633024 May 13 00:14:36.749334 kernel: Policy zone: DMA May 13 00:14:36.749352 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=ae60136413c5686d5b1e9c38408a367f831e354d706496e9f743f02289aad53d May 13 00:14:36.749358 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 13 00:14:36.749364 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) May 13 00:14:36.749370 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 13 00:14:36.749377 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 13 00:14:36.749392 kernel: Memory: 2457336K/2572288K available (9792K kernel code, 2094K rwdata, 7584K rodata, 36480K init, 777K bss, 114952K reserved, 0K cma-reserved) May 13 00:14:36.749398 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 May 13 00:14:36.749404 kernel: trace event string verifier disabled May 13 00:14:36.749410 kernel: rcu: Preemptible hierarchical RCU implementation. May 13 00:14:36.749417 kernel: rcu: RCU event tracing is enabled. May 13 00:14:36.749423 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. May 13 00:14:36.749429 kernel: Trampoline variant of Tasks RCU enabled. May 13 00:14:36.749435 kernel: Tracing variant of Tasks RCU enabled. May 13 00:14:36.749441 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 13 00:14:36.749448 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 May 13 00:14:36.749454 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 May 13 00:14:36.749461 kernel: GICv3: 256 SPIs implemented May 13 00:14:36.749467 kernel: GICv3: 0 Extended SPIs implemented May 13 00:14:36.749473 kernel: GICv3: Distributor has no Range Selector support May 13 00:14:36.749479 kernel: Root IRQ handler: gic_handle_irq May 13 00:14:36.749485 kernel: GICv3: 16 PPIs implemented May 13 00:14:36.749491 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 May 13 00:14:36.749497 kernel: ACPI: SRAT not present May 13 00:14:36.749502 kernel: ITS [mem 0x08080000-0x0809ffff] May 13 00:14:36.749509 kernel: ITS@0x0000000008080000: allocated 8192 Devices @400b0000 (indirect, esz 8, psz 64K, shr 1) May 13 00:14:36.749515 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @400c0000 (flat, esz 8, psz 64K, shr 1) May 13 00:14:36.749521 kernel: GICv3: using LPI property table @0x00000000400d0000 May 13 00:14:36.749527 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000000400e0000 May 13 00:14:36.749534 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 May 13 00:14:36.749541 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). May 13 00:14:36.749547 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns May 13 00:14:36.749553 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns May 13 00:14:36.749559 kernel: arm-pv: using stolen time PV May 13 00:14:36.749565 kernel: Console: colour dummy device 80x25 May 13 00:14:36.749572 kernel: ACPI: Core revision 20210730 May 13 00:14:36.749578 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) May 13 00:14:36.749584 kernel: pid_max: default: 32768 minimum: 301 May 13 00:14:36.749590 kernel: LSM: Security Framework initializing May 13 00:14:36.749598 kernel: SELinux: Initializing. May 13 00:14:36.749604 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 13 00:14:36.749610 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 13 00:14:36.749617 kernel: ACPI PPTT: PPTT table found, but unable to locate core 3 (3) May 13 00:14:36.749623 kernel: rcu: Hierarchical SRCU implementation. May 13 00:14:36.749629 kernel: Platform MSI: ITS@0x8080000 domain created May 13 00:14:36.749635 kernel: PCI/MSI: ITS@0x8080000 domain created May 13 00:14:36.749642 kernel: Remapping and enabling EFI services. May 13 00:14:36.749648 kernel: smp: Bringing up secondary CPUs ... May 13 00:14:36.749655 kernel: Detected PIPT I-cache on CPU1 May 13 00:14:36.749661 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 May 13 00:14:36.749668 kernel: GICv3: CPU1: using allocated LPI pending table @0x00000000400f0000 May 13 00:14:36.749674 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 May 13 00:14:36.749680 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] May 13 00:14:36.749686 kernel: Detected PIPT I-cache on CPU2 May 13 00:14:36.749693 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 May 13 00:14:36.749699 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040100000 May 13 00:14:36.749705 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 May 13 00:14:36.749711 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] May 13 00:14:36.749719 kernel: Detected PIPT I-cache on CPU3 May 13 00:14:36.749726 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 May 13 00:14:36.749732 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040110000 May 13 00:14:36.749738 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 May 13 00:14:36.749748 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] May 13 00:14:36.749756 kernel: smp: Brought up 1 node, 4 CPUs May 13 00:14:36.749763 kernel: SMP: Total of 4 processors activated. May 13 00:14:36.749769 kernel: CPU features: detected: 32-bit EL0 Support May 13 00:14:36.749776 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence May 13 00:14:36.749782 kernel: CPU features: detected: Common not Private translations May 13 00:14:36.749789 kernel: CPU features: detected: CRC32 instructions May 13 00:14:36.749795 kernel: CPU features: detected: RCpc load-acquire (LDAPR) May 13 00:14:36.749803 kernel: CPU features: detected: LSE atomic instructions May 13 00:14:36.749810 kernel: CPU features: detected: Privileged Access Never May 13 00:14:36.749816 kernel: CPU features: detected: RAS Extension Support May 13 00:14:36.749823 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) May 13 00:14:36.749830 kernel: CPU: All CPU(s) started at EL1 May 13 00:14:36.749838 kernel: alternatives: patching kernel code May 13 00:14:36.749844 kernel: devtmpfs: initialized May 13 00:14:36.749851 kernel: KASLR enabled May 13 00:14:36.749857 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 13 00:14:36.749864 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) May 13 00:14:36.749870 kernel: pinctrl core: initialized pinctrl subsystem May 13 00:14:36.749877 kernel: SMBIOS 3.0.0 present. May 13 00:14:36.749884 kernel: DMI: QEMU KVM Virtual Machine, BIOS 0.0.0 02/06/2015 May 13 00:14:36.749890 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 13 00:14:36.749898 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations May 13 00:14:36.749904 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations May 13 00:14:36.749911 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations May 13 00:14:36.749918 kernel: audit: initializing netlink subsys (disabled) May 13 00:14:36.749924 kernel: audit: type=2000 audit(0.030:1): state=initialized audit_enabled=0 res=1 May 13 00:14:36.749931 kernel: thermal_sys: Registered thermal governor 'step_wise' May 13 00:14:36.749940 kernel: cpuidle: using governor menu May 13 00:14:36.749947 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. May 13 00:14:36.749953 kernel: ASID allocator initialised with 32768 entries May 13 00:14:36.749961 kernel: ACPI: bus type PCI registered May 13 00:14:36.749968 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 13 00:14:36.749974 kernel: Serial: AMBA PL011 UART driver May 13 00:14:36.749981 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages May 13 00:14:36.749987 kernel: HugeTLB registered 32.0 MiB page size, pre-allocated 0 pages May 13 00:14:36.749994 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages May 13 00:14:36.750000 kernel: HugeTLB registered 64.0 KiB page size, pre-allocated 0 pages May 13 00:14:36.750007 kernel: cryptd: max_cpu_qlen set to 1000 May 13 00:14:36.750013 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) May 13 00:14:36.750021 kernel: ACPI: Added _OSI(Module Device) May 13 00:14:36.750028 kernel: ACPI: Added _OSI(Processor Device) May 13 00:14:36.750034 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 13 00:14:36.750041 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 13 00:14:36.750047 kernel: ACPI: Added _OSI(Linux-Dell-Video) May 13 00:14:36.750054 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) May 13 00:14:36.750060 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) May 13 00:14:36.750067 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded May 13 00:14:36.750073 kernel: ACPI: Interpreter enabled May 13 00:14:36.750081 kernel: ACPI: Using GIC for interrupt routing May 13 00:14:36.750087 kernel: ACPI: MCFG table detected, 1 entries May 13 00:14:36.750094 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA May 13 00:14:36.750100 kernel: printk: console [ttyAMA0] enabled May 13 00:14:36.750107 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) May 13 00:14:36.750229 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] May 13 00:14:36.750292 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] May 13 00:14:36.750364 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] May 13 00:14:36.750431 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 May 13 00:14:36.750489 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] May 13 00:14:36.750498 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] May 13 00:14:36.750504 kernel: PCI host bridge to bus 0000:00 May 13 00:14:36.750581 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] May 13 00:14:36.750639 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] May 13 00:14:36.750691 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] May 13 00:14:36.750794 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] May 13 00:14:36.750872 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 May 13 00:14:36.750940 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 May 13 00:14:36.751002 kernel: pci 0000:00:01.0: reg 0x10: [io 0x0000-0x001f] May 13 00:14:36.751061 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x10000000-0x10000fff] May 13 00:14:36.751120 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] May 13 00:14:36.751181 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] May 13 00:14:36.751240 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x10000000-0x10000fff] May 13 00:14:36.751299 kernel: pci 0000:00:01.0: BAR 0: assigned [io 0x1000-0x101f] May 13 00:14:36.751372 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] May 13 00:14:36.751433 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] May 13 00:14:36.751489 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] May 13 00:14:36.751498 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 May 13 00:14:36.751505 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 May 13 00:14:36.751514 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 May 13 00:14:36.751521 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 May 13 00:14:36.751528 kernel: iommu: Default domain type: Translated May 13 00:14:36.751535 kernel: iommu: DMA domain TLB invalidation policy: strict mode May 13 00:14:36.751541 kernel: vgaarb: loaded May 13 00:14:36.751548 kernel: pps_core: LinuxPPS API ver. 1 registered May 13 00:14:36.751554 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti May 13 00:14:36.751561 kernel: PTP clock support registered May 13 00:14:36.751567 kernel: Registered efivars operations May 13 00:14:36.751575 kernel: clocksource: Switched to clocksource arch_sys_counter May 13 00:14:36.751582 kernel: VFS: Disk quotas dquot_6.6.0 May 13 00:14:36.751589 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 13 00:14:36.751595 kernel: pnp: PnP ACPI init May 13 00:14:36.751660 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved May 13 00:14:36.751669 kernel: pnp: PnP ACPI: found 1 devices May 13 00:14:36.751676 kernel: NET: Registered PF_INET protocol family May 13 00:14:36.751683 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) May 13 00:14:36.751691 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) May 13 00:14:36.751698 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 13 00:14:36.751705 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) May 13 00:14:36.751712 kernel: TCP bind hash table entries: 32768 (order: 7, 524288 bytes, linear) May 13 00:14:36.751718 kernel: TCP: Hash tables configured (established 32768 bind 32768) May 13 00:14:36.751725 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) May 13 00:14:36.751732 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) May 13 00:14:36.751738 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 13 00:14:36.751745 kernel: PCI: CLS 0 bytes, default 64 May 13 00:14:36.751754 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available May 13 00:14:36.751760 kernel: kvm [1]: HYP mode not available May 13 00:14:36.751767 kernel: Initialise system trusted keyrings May 13 00:14:36.751773 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 May 13 00:14:36.751780 kernel: Key type asymmetric registered May 13 00:14:36.751787 kernel: Asymmetric key parser 'x509' registered May 13 00:14:36.751793 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) May 13 00:14:36.751800 kernel: io scheduler mq-deadline registered May 13 00:14:36.751806 kernel: io scheduler kyber registered May 13 00:14:36.751814 kernel: io scheduler bfq registered May 13 00:14:36.751821 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 May 13 00:14:36.751827 kernel: ACPI: button: Power Button [PWRB] May 13 00:14:36.751834 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 May 13 00:14:36.751892 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) May 13 00:14:36.751901 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 13 00:14:36.751908 kernel: thunder_xcv, ver 1.0 May 13 00:14:36.751914 kernel: thunder_bgx, ver 1.0 May 13 00:14:36.751921 kernel: nicpf, ver 1.0 May 13 00:14:36.751929 kernel: nicvf, ver 1.0 May 13 00:14:36.751995 kernel: rtc-efi rtc-efi.0: registered as rtc0 May 13 00:14:36.752050 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-05-13T00:14:36 UTC (1747095276) May 13 00:14:36.752059 kernel: hid: raw HID events driver (C) Jiri Kosina May 13 00:14:36.752066 kernel: NET: Registered PF_INET6 protocol family May 13 00:14:36.752073 kernel: Segment Routing with IPv6 May 13 00:14:36.752079 kernel: In-situ OAM (IOAM) with IPv6 May 13 00:14:36.752086 kernel: NET: Registered PF_PACKET protocol family May 13 00:14:36.752094 kernel: Key type dns_resolver registered May 13 00:14:36.752101 kernel: registered taskstats version 1 May 13 00:14:36.752108 kernel: Loading compiled-in X.509 certificates May 13 00:14:36.752116 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.181-flatcar: d291b704d59536a3c0ba96fd6f5a99459de8de99' May 13 00:14:36.752123 kernel: Key type .fscrypt registered May 13 00:14:36.752129 kernel: Key type fscrypt-provisioning registered May 13 00:14:36.752136 kernel: ima: No TPM chip found, activating TPM-bypass! May 13 00:14:36.752143 kernel: ima: Allocated hash algorithm: sha1 May 13 00:14:36.752150 kernel: ima: No architecture policies found May 13 00:14:36.752158 kernel: clk: Disabling unused clocks May 13 00:14:36.752165 kernel: Freeing unused kernel memory: 36480K May 13 00:14:36.752171 kernel: Run /init as init process May 13 00:14:36.752178 kernel: with arguments: May 13 00:14:36.752184 kernel: /init May 13 00:14:36.752191 kernel: with environment: May 13 00:14:36.752197 kernel: HOME=/ May 13 00:14:36.752204 kernel: TERM=linux May 13 00:14:36.752211 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 13 00:14:36.752220 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) May 13 00:14:36.752229 systemd[1]: Detected virtualization kvm. May 13 00:14:36.752236 systemd[1]: Detected architecture arm64. May 13 00:14:36.752243 systemd[1]: Running in initrd. May 13 00:14:36.752250 systemd[1]: No hostname configured, using default hostname. May 13 00:14:36.752257 systemd[1]: Hostname set to . May 13 00:14:36.752264 systemd[1]: Initializing machine ID from VM UUID. May 13 00:14:36.752273 systemd[1]: Queued start job for default target initrd.target. May 13 00:14:36.752280 systemd[1]: Started systemd-ask-password-console.path. May 13 00:14:36.752287 systemd[1]: Reached target cryptsetup.target. May 13 00:14:36.752294 systemd[1]: Reached target paths.target. May 13 00:14:36.752301 systemd[1]: Reached target slices.target. May 13 00:14:36.752308 systemd[1]: Reached target swap.target. May 13 00:14:36.752315 systemd[1]: Reached target timers.target. May 13 00:14:36.752323 systemd[1]: Listening on iscsid.socket. May 13 00:14:36.752332 systemd[1]: Listening on iscsiuio.socket. May 13 00:14:36.752374 systemd[1]: Listening on systemd-journald-audit.socket. May 13 00:14:36.752386 systemd[1]: Listening on systemd-journald-dev-log.socket. May 13 00:14:36.752394 systemd[1]: Listening on systemd-journald.socket. May 13 00:14:36.752401 systemd[1]: Listening on systemd-networkd.socket. May 13 00:14:36.752408 systemd[1]: Listening on systemd-udevd-control.socket. May 13 00:14:36.752415 systemd[1]: Listening on systemd-udevd-kernel.socket. May 13 00:14:36.752422 systemd[1]: Reached target sockets.target. May 13 00:14:36.752431 systemd[1]: Starting kmod-static-nodes.service... May 13 00:14:36.752439 systemd[1]: Finished network-cleanup.service. May 13 00:14:36.752446 systemd[1]: Starting systemd-fsck-usr.service... May 13 00:14:36.752453 systemd[1]: Starting systemd-journald.service... May 13 00:14:36.752460 systemd[1]: Starting systemd-modules-load.service... May 13 00:14:36.752467 systemd[1]: Starting systemd-resolved.service... May 13 00:14:36.752475 systemd[1]: Starting systemd-vconsole-setup.service... May 13 00:14:36.752482 systemd[1]: Finished kmod-static-nodes.service. May 13 00:14:36.752489 systemd[1]: Finished systemd-fsck-usr.service. May 13 00:14:36.752497 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... May 13 00:14:36.752505 systemd[1]: Finished systemd-vconsole-setup.service. May 13 00:14:36.752512 systemd[1]: Starting dracut-cmdline-ask.service... May 13 00:14:36.752519 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. May 13 00:14:36.752526 kernel: audit: type=1130 audit(1747095276.750:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:36.752537 systemd-journald[289]: Journal started May 13 00:14:36.752581 systemd-journald[289]: Runtime Journal (/run/log/journal/7878dbb6338145bcab4321800b19da4e) is 6.0M, max 48.7M, 42.6M free. May 13 00:14:36.750000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:36.745969 systemd-modules-load[290]: Inserted module 'overlay' May 13 00:14:36.754709 systemd[1]: Started systemd-journald.service. May 13 00:14:36.755000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:36.759373 kernel: audit: type=1130 audit(1747095276.755:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:36.769359 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 13 00:14:36.770849 systemd-resolved[291]: Positive Trust Anchors: May 13 00:14:36.773693 kernel: Bridge firewalling registered May 13 00:14:36.770860 systemd-resolved[291]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 13 00:14:36.770887 systemd-resolved[291]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test May 13 00:14:36.772756 systemd-modules-load[290]: Inserted module 'br_netfilter' May 13 00:14:36.787207 kernel: audit: type=1130 audit(1747095276.775:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:36.787230 kernel: audit: type=1130 audit(1747095276.783:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:36.775000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:36.783000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:36.773906 systemd[1]: Finished dracut-cmdline-ask.service. May 13 00:14:36.775194 systemd-resolved[291]: Defaulting to hostname 'linux'. May 13 00:14:36.775986 systemd[1]: Started systemd-resolved.service. May 13 00:14:36.784177 systemd[1]: Reached target nss-lookup.target. May 13 00:14:36.792394 kernel: SCSI subsystem initialized May 13 00:14:36.788862 systemd[1]: Starting dracut-cmdline.service... May 13 00:14:36.798095 dracut-cmdline[308]: dracut-dracut-053 May 13 00:14:36.801606 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 13 00:14:36.801632 kernel: device-mapper: uevent: version 1.0.3 May 13 00:14:36.801645 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com May 13 00:14:36.802072 systemd-modules-load[290]: Inserted module 'dm_multipath' May 13 00:14:36.802880 systemd[1]: Finished systemd-modules-load.service. May 13 00:14:36.803000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:36.807173 dracut-cmdline[308]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=ae60136413c5686d5b1e9c38408a367f831e354d706496e9f743f02289aad53d May 13 00:14:36.812292 kernel: audit: type=1130 audit(1747095276.803:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:36.804656 systemd[1]: Starting systemd-sysctl.service... May 13 00:14:36.813114 systemd[1]: Finished systemd-sysctl.service. May 13 00:14:36.813000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:36.817365 kernel: audit: type=1130 audit(1747095276.813:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:36.864359 kernel: Loading iSCSI transport class v2.0-870. May 13 00:14:36.876364 kernel: iscsi: registered transport (tcp) May 13 00:14:36.893605 kernel: iscsi: registered transport (qla4xxx) May 13 00:14:36.893633 kernel: QLogic iSCSI HBA Driver May 13 00:14:36.928661 systemd[1]: Finished dracut-cmdline.service. May 13 00:14:36.929000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:36.930392 systemd[1]: Starting dracut-pre-udev.service... May 13 00:14:36.934313 kernel: audit: type=1130 audit(1747095276.929:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:36.976371 kernel: raid6: neonx8 gen() 13718 MB/s May 13 00:14:36.993362 kernel: raid6: neonx8 xor() 10791 MB/s May 13 00:14:37.010357 kernel: raid6: neonx4 gen() 13529 MB/s May 13 00:14:37.027359 kernel: raid6: neonx4 xor() 11101 MB/s May 13 00:14:37.044363 kernel: raid6: neonx2 gen() 13044 MB/s May 13 00:14:37.061357 kernel: raid6: neonx2 xor() 10434 MB/s May 13 00:14:37.078375 kernel: raid6: neonx1 gen() 10583 MB/s May 13 00:14:37.095357 kernel: raid6: neonx1 xor() 8767 MB/s May 13 00:14:37.112356 kernel: raid6: int64x8 gen() 6247 MB/s May 13 00:14:37.129356 kernel: raid6: int64x8 xor() 3537 MB/s May 13 00:14:37.146373 kernel: raid6: int64x4 gen() 7195 MB/s May 13 00:14:37.163357 kernel: raid6: int64x4 xor() 3852 MB/s May 13 00:14:37.180362 kernel: raid6: int64x2 gen() 6127 MB/s May 13 00:14:37.197366 kernel: raid6: int64x2 xor() 3316 MB/s May 13 00:14:37.214361 kernel: raid6: int64x1 gen() 5036 MB/s May 13 00:14:37.231507 kernel: raid6: int64x1 xor() 2638 MB/s May 13 00:14:37.231550 kernel: raid6: using algorithm neonx8 gen() 13718 MB/s May 13 00:14:37.231560 kernel: raid6: .... xor() 10791 MB/s, rmw enabled May 13 00:14:37.232604 kernel: raid6: using neon recovery algorithm May 13 00:14:37.243362 kernel: xor: measuring software checksum speed May 13 00:14:37.243388 kernel: 8regs : 17217 MB/sec May 13 00:14:37.244682 kernel: 32regs : 18086 MB/sec May 13 00:14:37.244694 kernel: arm64_neon : 27654 MB/sec May 13 00:14:37.244703 kernel: xor: using function: arm64_neon (27654 MB/sec) May 13 00:14:37.301368 kernel: Btrfs loaded, crc32c=crc32c-generic, zoned=no, fsverity=no May 13 00:14:37.315000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:37.314658 systemd[1]: Finished dracut-pre-udev.service. May 13 00:14:37.320321 kernel: audit: type=1130 audit(1747095277.315:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:37.320359 kernel: audit: type=1334 audit(1747095277.316:10): prog-id=7 op=LOAD May 13 00:14:37.316000 audit: BPF prog-id=7 op=LOAD May 13 00:14:37.320000 audit: BPF prog-id=8 op=LOAD May 13 00:14:37.321229 systemd[1]: Starting systemd-udevd.service... May 13 00:14:37.342446 systemd-udevd[491]: Using default interface naming scheme 'v252'. May 13 00:14:37.348000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:37.347898 systemd[1]: Started systemd-udevd.service. May 13 00:14:37.349945 systemd[1]: Starting dracut-pre-trigger.service... May 13 00:14:37.368404 dracut-pre-trigger[497]: rd.md=0: removing MD RAID activation May 13 00:14:37.398034 systemd[1]: Finished dracut-pre-trigger.service. May 13 00:14:37.399843 systemd[1]: Starting systemd-udev-trigger.service... May 13 00:14:37.398000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:37.434000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:37.434159 systemd[1]: Finished systemd-udev-trigger.service. May 13 00:14:37.470413 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) May 13 00:14:37.475857 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. May 13 00:14:37.475881 kernel: GPT:9289727 != 19775487 May 13 00:14:37.475890 kernel: GPT:Alternate GPT header not at the end of the disk. May 13 00:14:37.475899 kernel: GPT:9289727 != 19775487 May 13 00:14:37.475908 kernel: GPT: Use GNU Parted to correct GPT errors. May 13 00:14:37.475917 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 13 00:14:37.495223 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. May 13 00:14:37.498657 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (547) May 13 00:14:37.500082 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. May 13 00:14:37.507604 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. May 13 00:14:37.511163 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. May 13 00:14:37.514599 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. May 13 00:14:37.516453 systemd[1]: Starting disk-uuid.service... May 13 00:14:37.522686 disk-uuid[564]: Primary Header is updated. May 13 00:14:37.522686 disk-uuid[564]: Secondary Entries is updated. May 13 00:14:37.522686 disk-uuid[564]: Secondary Header is updated. May 13 00:14:37.525829 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 13 00:14:38.544681 disk-uuid[565]: The operation has completed successfully. May 13 00:14:38.545827 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 13 00:14:38.565972 systemd[1]: disk-uuid.service: Deactivated successfully. May 13 00:14:38.566000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:38.566000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:38.566068 systemd[1]: Finished disk-uuid.service. May 13 00:14:38.570268 systemd[1]: Starting verity-setup.service... May 13 00:14:38.585358 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" May 13 00:14:38.607905 systemd[1]: Found device dev-mapper-usr.device. May 13 00:14:38.610332 systemd[1]: Mounting sysusr-usr.mount... May 13 00:14:38.612460 systemd[1]: Finished verity-setup.service. May 13 00:14:38.613000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:38.659366 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. May 13 00:14:38.659659 systemd[1]: Mounted sysusr-usr.mount. May 13 00:14:38.660542 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. May 13 00:14:38.661355 systemd[1]: Starting ignition-setup.service... May 13 00:14:38.663671 systemd[1]: Starting parse-ip-for-networkd.service... May 13 00:14:38.671922 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm May 13 00:14:38.671971 kernel: BTRFS info (device vda6): using free space tree May 13 00:14:38.671981 kernel: BTRFS info (device vda6): has skinny extents May 13 00:14:38.681827 systemd[1]: mnt-oem.mount: Deactivated successfully. May 13 00:14:38.687837 systemd[1]: Finished ignition-setup.service. May 13 00:14:38.688000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:38.689500 systemd[1]: Starting ignition-fetch-offline.service... May 13 00:14:38.764367 systemd[1]: Finished parse-ip-for-networkd.service. May 13 00:14:38.764000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:38.766000 audit: BPF prog-id=9 op=LOAD May 13 00:14:38.767149 systemd[1]: Starting systemd-networkd.service... May 13 00:14:38.789459 ignition[644]: Ignition 2.14.0 May 13 00:14:38.789469 ignition[644]: Stage: fetch-offline May 13 00:14:38.789509 ignition[644]: no configs at "/usr/lib/ignition/base.d" May 13 00:14:38.789518 ignition[644]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 13 00:14:38.789651 ignition[644]: parsed url from cmdline: "" May 13 00:14:38.789654 ignition[644]: no config URL provided May 13 00:14:38.789658 ignition[644]: reading system config file "/usr/lib/ignition/user.ign" May 13 00:14:38.789665 ignition[644]: no config at "/usr/lib/ignition/user.ign" May 13 00:14:38.789683 ignition[644]: op(1): [started] loading QEMU firmware config module May 13 00:14:38.789688 ignition[644]: op(1): executing: "modprobe" "qemu_fw_cfg" May 13 00:14:38.799000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:38.797792 systemd-networkd[739]: lo: Link UP May 13 00:14:38.795536 ignition[644]: op(1): [finished] loading QEMU firmware config module May 13 00:14:38.797796 systemd-networkd[739]: lo: Gained carrier May 13 00:14:38.798146 systemd-networkd[739]: Enumeration completed May 13 00:14:38.798258 systemd[1]: Started systemd-networkd.service. May 13 00:14:38.798324 systemd-networkd[739]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 13 00:14:38.799499 systemd[1]: Reached target network.target. May 13 00:14:38.799506 systemd-networkd[739]: eth0: Link UP May 13 00:14:38.799510 systemd-networkd[739]: eth0: Gained carrier May 13 00:14:38.801245 systemd[1]: Starting iscsiuio.service... May 13 00:14:38.811252 systemd[1]: Started iscsiuio.service. May 13 00:14:38.811000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:38.813077 systemd[1]: Starting iscsid.service... May 13 00:14:38.817482 iscsid[745]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi May 13 00:14:38.817482 iscsid[745]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. May 13 00:14:38.817482 iscsid[745]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. May 13 00:14:38.817482 iscsid[745]: If using hardware iscsi like qla4xxx this message can be ignored. May 13 00:14:38.817482 iscsid[745]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi May 13 00:14:38.817482 iscsid[745]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf May 13 00:14:38.825000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:38.825011 systemd[1]: Started iscsid.service. May 13 00:14:38.826433 systemd-networkd[739]: eth0: DHCPv4 address 10.0.0.8/16, gateway 10.0.0.1 acquired from 10.0.0.1 May 13 00:14:38.826744 systemd[1]: Starting dracut-initqueue.service... May 13 00:14:38.838000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:38.837861 systemd[1]: Finished dracut-initqueue.service. May 13 00:14:38.838894 systemd[1]: Reached target remote-fs-pre.target. May 13 00:14:38.839827 systemd[1]: Reached target remote-cryptsetup.target. May 13 00:14:38.840769 systemd[1]: Reached target remote-fs.target. May 13 00:14:38.843123 systemd[1]: Starting dracut-pre-mount.service... May 13 00:14:38.851455 systemd[1]: Finished dracut-pre-mount.service. May 13 00:14:38.851000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:38.864120 ignition[644]: parsing config with SHA512: b53b2b6f84beb4335ce8ab144ca04908da93c2c5cd2ccf22db92c3adae9df8b54c3ca8831bc3ada407bd0ae36e68895b601adbf6673acc40eae7dd3a4fc66973 May 13 00:14:38.875049 unknown[644]: fetched base config from "system" May 13 00:14:38.875060 unknown[644]: fetched user config from "qemu" May 13 00:14:38.875528 ignition[644]: fetch-offline: fetch-offline passed May 13 00:14:38.877000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:38.876918 systemd[1]: Finished ignition-fetch-offline.service. May 13 00:14:38.875577 ignition[644]: Ignition finished successfully May 13 00:14:38.877956 systemd[1]: ignition-fetch.service was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). May 13 00:14:38.878754 systemd[1]: Starting ignition-kargs.service... May 13 00:14:38.887768 ignition[760]: Ignition 2.14.0 May 13 00:14:38.887778 ignition[760]: Stage: kargs May 13 00:14:38.887870 ignition[760]: no configs at "/usr/lib/ignition/base.d" May 13 00:14:38.887879 ignition[760]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 13 00:14:38.888828 ignition[760]: kargs: kargs passed May 13 00:14:38.892000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:38.890963 systemd[1]: Finished ignition-kargs.service. May 13 00:14:38.888869 ignition[760]: Ignition finished successfully May 13 00:14:38.893454 systemd[1]: Starting ignition-disks.service... May 13 00:14:38.900069 ignition[766]: Ignition 2.14.0 May 13 00:14:38.900078 ignition[766]: Stage: disks May 13 00:14:38.900173 ignition[766]: no configs at "/usr/lib/ignition/base.d" May 13 00:14:38.900183 ignition[766]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 13 00:14:38.902000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:38.902076 systemd[1]: Finished ignition-disks.service. May 13 00:14:38.901174 ignition[766]: disks: disks passed May 13 00:14:38.903143 systemd[1]: Reached target initrd-root-device.target. May 13 00:14:38.901216 ignition[766]: Ignition finished successfully May 13 00:14:38.904807 systemd[1]: Reached target local-fs-pre.target. May 13 00:14:38.906712 systemd[1]: Reached target local-fs.target. May 13 00:14:38.909398 systemd[1]: Reached target sysinit.target. May 13 00:14:38.910769 systemd[1]: Reached target basic.target. May 13 00:14:38.912950 systemd[1]: Starting systemd-fsck-root.service... May 13 00:14:38.924210 systemd-fsck[774]: ROOT: clean, 619/553520 files, 56022/553472 blocks May 13 00:14:38.929710 systemd[1]: Finished systemd-fsck-root.service. May 13 00:14:38.930000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:38.931614 systemd[1]: Mounting sysroot.mount... May 13 00:14:38.939363 kernel: EXT4-fs (vda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. May 13 00:14:38.939456 systemd[1]: Mounted sysroot.mount. May 13 00:14:38.940192 systemd[1]: Reached target initrd-root-fs.target. May 13 00:14:38.942744 systemd[1]: Mounting sysroot-usr.mount... May 13 00:14:38.943623 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. May 13 00:14:38.943666 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 13 00:14:38.943690 systemd[1]: Reached target ignition-diskful.target. May 13 00:14:38.945626 systemd[1]: Mounted sysroot-usr.mount. May 13 00:14:38.947404 systemd[1]: Starting initrd-setup-root.service... May 13 00:14:38.951673 initrd-setup-root[784]: cut: /sysroot/etc/passwd: No such file or directory May 13 00:14:38.958445 initrd-setup-root[792]: cut: /sysroot/etc/group: No such file or directory May 13 00:14:38.967411 initrd-setup-root[800]: cut: /sysroot/etc/shadow: No such file or directory May 13 00:14:38.974261 initrd-setup-root[808]: cut: /sysroot/etc/gshadow: No such file or directory May 13 00:14:39.009043 systemd[1]: Finished initrd-setup-root.service. May 13 00:14:39.009000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:39.012525 systemd[1]: Starting ignition-mount.service... May 13 00:14:39.013965 systemd[1]: Starting sysroot-boot.service... May 13 00:14:39.021208 bash[825]: umount: /sysroot/usr/share/oem: not mounted. May 13 00:14:39.030516 ignition[827]: INFO : Ignition 2.14.0 May 13 00:14:39.030516 ignition[827]: INFO : Stage: mount May 13 00:14:39.032704 ignition[827]: INFO : no configs at "/usr/lib/ignition/base.d" May 13 00:14:39.032704 ignition[827]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 13 00:14:39.032704 ignition[827]: INFO : mount: mount passed May 13 00:14:39.032704 ignition[827]: INFO : Ignition finished successfully May 13 00:14:39.035000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:39.034327 systemd[1]: Finished ignition-mount.service. May 13 00:14:39.040876 systemd[1]: Finished sysroot-boot.service. May 13 00:14:39.041000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:39.570212 systemd-resolved[291]: Detected conflict on linux IN A 10.0.0.8 May 13 00:14:39.570239 systemd-resolved[291]: Hostname conflict, changing published hostname from 'linux' to 'linux6'. May 13 00:14:39.620143 systemd[1]: Mounting sysroot-usr-share-oem.mount... May 13 00:14:39.628248 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (835) May 13 00:14:39.628303 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm May 13 00:14:39.629061 kernel: BTRFS info (device vda6): using free space tree May 13 00:14:39.629073 kernel: BTRFS info (device vda6): has skinny extents May 13 00:14:39.633592 systemd[1]: Mounted sysroot-usr-share-oem.mount. May 13 00:14:39.635362 systemd[1]: Starting ignition-files.service... May 13 00:14:39.650635 ignition[855]: INFO : Ignition 2.14.0 May 13 00:14:39.650635 ignition[855]: INFO : Stage: files May 13 00:14:39.652312 ignition[855]: INFO : no configs at "/usr/lib/ignition/base.d" May 13 00:14:39.652312 ignition[855]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 13 00:14:39.652312 ignition[855]: DEBUG : files: compiled without relabeling support, skipping May 13 00:14:39.656977 ignition[855]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 13 00:14:39.656977 ignition[855]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 13 00:14:39.662857 ignition[855]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 13 00:14:39.664319 ignition[855]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 13 00:14:39.664319 ignition[855]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 13 00:14:39.663924 unknown[855]: wrote ssh authorized keys file for user: core May 13 00:14:39.668402 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" May 13 00:14:39.668402 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" May 13 00:14:39.668402 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" May 13 00:14:39.668402 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 May 13 00:14:39.714638 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK May 13 00:14:39.949011 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" May 13 00:14:39.949011 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" May 13 00:14:39.952956 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" May 13 00:14:39.952956 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" May 13 00:14:39.952956 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" May 13 00:14:39.952956 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 13 00:14:39.952956 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 13 00:14:39.952956 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 13 00:14:39.952956 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 13 00:14:39.952956 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" May 13 00:14:39.952956 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 13 00:14:39.952956 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" May 13 00:14:39.952956 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" May 13 00:14:39.952956 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" May 13 00:14:39.952956 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-arm64.raw: attempt #1 May 13 00:14:40.095525 systemd-networkd[739]: eth0: Gained IPv6LL May 13 00:14:40.293407 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK May 13 00:14:40.789923 ignition[855]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" May 13 00:14:40.789923 ignition[855]: INFO : files: op(c): [started] processing unit "containerd.service" May 13 00:14:40.794249 ignition[855]: INFO : files: op(c): op(d): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" May 13 00:14:40.794249 ignition[855]: INFO : files: op(c): op(d): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" May 13 00:14:40.794249 ignition[855]: INFO : files: op(c): [finished] processing unit "containerd.service" May 13 00:14:40.794249 ignition[855]: INFO : files: op(e): [started] processing unit "prepare-helm.service" May 13 00:14:40.794249 ignition[855]: INFO : files: op(e): op(f): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 13 00:14:40.794249 ignition[855]: INFO : files: op(e): op(f): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 13 00:14:40.794249 ignition[855]: INFO : files: op(e): [finished] processing unit "prepare-helm.service" May 13 00:14:40.794249 ignition[855]: INFO : files: op(10): [started] processing unit "coreos-metadata.service" May 13 00:14:40.794249 ignition[855]: INFO : files: op(10): op(11): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" May 13 00:14:40.794249 ignition[855]: INFO : files: op(10): op(11): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" May 13 00:14:40.794249 ignition[855]: INFO : files: op(10): [finished] processing unit "coreos-metadata.service" May 13 00:14:40.794249 ignition[855]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" May 13 00:14:40.794249 ignition[855]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" May 13 00:14:40.794249 ignition[855]: INFO : files: op(13): [started] setting preset to disabled for "coreos-metadata.service" May 13 00:14:40.794249 ignition[855]: INFO : files: op(13): op(14): [started] removing enablement symlink(s) for "coreos-metadata.service" May 13 00:14:40.846378 ignition[855]: INFO : files: op(13): op(14): [finished] removing enablement symlink(s) for "coreos-metadata.service" May 13 00:14:40.848509 ignition[855]: INFO : files: op(13): [finished] setting preset to disabled for "coreos-metadata.service" May 13 00:14:40.848509 ignition[855]: INFO : files: createResultFile: createFiles: op(15): [started] writing file "/sysroot/etc/.ignition-result.json" May 13 00:14:40.848509 ignition[855]: INFO : files: createResultFile: createFiles: op(15): [finished] writing file "/sysroot/etc/.ignition-result.json" May 13 00:14:40.848509 ignition[855]: INFO : files: files passed May 13 00:14:40.848509 ignition[855]: INFO : Ignition finished successfully May 13 00:14:40.850000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.848945 systemd[1]: Finished ignition-files.service. May 13 00:14:40.851998 systemd[1]: Starting initrd-setup-root-after-ignition.service... May 13 00:14:40.859000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.859000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.853589 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). May 13 00:14:40.862529 initrd-setup-root-after-ignition[880]: grep: /sysroot/usr/share/oem/oem-release: No such file or directory May 13 00:14:40.862000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.854434 systemd[1]: Starting ignition-quench.service... May 13 00:14:40.866161 initrd-setup-root-after-ignition[882]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 13 00:14:40.857928 systemd[1]: ignition-quench.service: Deactivated successfully. May 13 00:14:40.858029 systemd[1]: Finished ignition-quench.service. May 13 00:14:40.860702 systemd[1]: Finished initrd-setup-root-after-ignition.service. May 13 00:14:40.863503 systemd[1]: Reached target ignition-complete.target. May 13 00:14:40.866210 systemd[1]: Starting initrd-parse-etc.service... May 13 00:14:40.880297 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 13 00:14:40.880424 systemd[1]: Finished initrd-parse-etc.service. May 13 00:14:40.881000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.881000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.882259 systemd[1]: Reached target initrd-fs.target. May 13 00:14:40.883508 systemd[1]: Reached target initrd.target. May 13 00:14:40.885003 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. May 13 00:14:40.885944 systemd[1]: Starting dracut-pre-pivot.service... May 13 00:14:40.897309 systemd[1]: Finished dracut-pre-pivot.service. May 13 00:14:40.898000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.899962 systemd[1]: Starting initrd-cleanup.service... May 13 00:14:40.908540 systemd[1]: Stopped target nss-lookup.target. May 13 00:14:40.909493 systemd[1]: Stopped target remote-cryptsetup.target. May 13 00:14:40.910977 systemd[1]: Stopped target timers.target. May 13 00:14:40.912323 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 13 00:14:40.913000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.912492 systemd[1]: Stopped dracut-pre-pivot.service. May 13 00:14:40.913814 systemd[1]: Stopped target initrd.target. May 13 00:14:40.915198 systemd[1]: Stopped target basic.target. May 13 00:14:40.916558 systemd[1]: Stopped target ignition-complete.target. May 13 00:14:40.917921 systemd[1]: Stopped target ignition-diskful.target. May 13 00:14:40.919248 systemd[1]: Stopped target initrd-root-device.target. May 13 00:14:40.920815 systemd[1]: Stopped target remote-fs.target. May 13 00:14:40.922285 systemd[1]: Stopped target remote-fs-pre.target. May 13 00:14:40.923892 systemd[1]: Stopped target sysinit.target. May 13 00:14:40.925177 systemd[1]: Stopped target local-fs.target. May 13 00:14:40.926636 systemd[1]: Stopped target local-fs-pre.target. May 13 00:14:40.927968 systemd[1]: Stopped target swap.target. May 13 00:14:40.930000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.929220 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 13 00:14:40.929360 systemd[1]: Stopped dracut-pre-mount.service. May 13 00:14:40.933000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.930793 systemd[1]: Stopped target cryptsetup.target. May 13 00:14:40.934000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.932030 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 13 00:14:40.932138 systemd[1]: Stopped dracut-initqueue.service. May 13 00:14:40.933646 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 13 00:14:40.933747 systemd[1]: Stopped ignition-fetch-offline.service. May 13 00:14:40.935112 systemd[1]: Stopped target paths.target. May 13 00:14:40.936324 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 13 00:14:40.941377 systemd[1]: Stopped systemd-ask-password-console.path. May 13 00:14:40.942294 systemd[1]: Stopped target slices.target. May 13 00:14:40.943672 systemd[1]: Stopped target sockets.target. May 13 00:14:40.944931 systemd[1]: iscsid.socket: Deactivated successfully. May 13 00:14:40.945007 systemd[1]: Closed iscsid.socket. May 13 00:14:40.947000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.946139 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 13 00:14:40.948000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.946239 systemd[1]: Stopped initrd-setup-root-after-ignition.service. May 13 00:14:40.947752 systemd[1]: ignition-files.service: Deactivated successfully. May 13 00:14:40.947844 systemd[1]: Stopped ignition-files.service. May 13 00:14:40.949879 systemd[1]: Stopping ignition-mount.service... May 13 00:14:40.955000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.951384 systemd[1]: Stopping iscsiuio.service... May 13 00:14:40.956000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.952945 systemd[1]: Stopping sysroot-boot.service... May 13 00:14:40.953722 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 13 00:14:40.953860 systemd[1]: Stopped systemd-udev-trigger.service. May 13 00:14:40.959000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.955252 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 13 00:14:40.955378 systemd[1]: Stopped dracut-pre-trigger.service. May 13 00:14:40.958209 systemd[1]: iscsiuio.service: Deactivated successfully. May 13 00:14:40.958375 systemd[1]: Stopped iscsiuio.service. May 13 00:14:40.960270 systemd[1]: iscsiuio.socket: Deactivated successfully. May 13 00:14:40.967318 ignition[895]: INFO : Ignition 2.14.0 May 13 00:14:40.967318 ignition[895]: INFO : Stage: umount May 13 00:14:40.967318 ignition[895]: INFO : no configs at "/usr/lib/ignition/base.d" May 13 00:14:40.967318 ignition[895]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 13 00:14:40.967318 ignition[895]: INFO : umount: umount passed May 13 00:14:40.967318 ignition[895]: INFO : Ignition finished successfully May 13 00:14:40.968000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.968000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.969000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.973000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.975000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.960332 systemd[1]: Closed iscsiuio.socket. May 13 00:14:40.976000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.967319 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 13 00:14:40.967814 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 13 00:14:40.967898 systemd[1]: Finished initrd-cleanup.service. May 13 00:14:40.969237 systemd[1]: ignition-mount.service: Deactivated successfully. May 13 00:14:40.969311 systemd[1]: Stopped ignition-mount.service. May 13 00:14:40.971072 systemd[1]: Stopped target network.target. May 13 00:14:40.971888 systemd[1]: ignition-disks.service: Deactivated successfully. May 13 00:14:40.971941 systemd[1]: Stopped ignition-disks.service. May 13 00:14:40.974162 systemd[1]: ignition-kargs.service: Deactivated successfully. May 13 00:14:40.988000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.974203 systemd[1]: Stopped ignition-kargs.service. May 13 00:14:40.990000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.975526 systemd[1]: ignition-setup.service: Deactivated successfully. May 13 00:14:40.975565 systemd[1]: Stopped ignition-setup.service. May 13 00:14:40.977240 systemd[1]: Stopping systemd-networkd.service... May 13 00:14:40.978703 systemd[1]: Stopping systemd-resolved.service... May 13 00:14:40.995000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.996000 audit: BPF prog-id=6 op=UNLOAD May 13 00:14:40.985406 systemd-networkd[739]: eth0: DHCPv6 lease lost May 13 00:14:40.996000 audit: BPF prog-id=9 op=UNLOAD May 13 00:14:40.996000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.987872 systemd[1]: systemd-networkd.service: Deactivated successfully. May 13 00:14:41.003930 kernel: kauditd_printk_skb: 50 callbacks suppressed May 13 00:14:41.003955 kernel: audit: type=1131 audit(1747095280.998:61): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.998000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.987982 systemd[1]: Stopped systemd-networkd.service. May 13 00:14:40.989099 systemd[1]: systemd-resolved.service: Deactivated successfully. May 13 00:14:40.989185 systemd[1]: Stopped systemd-resolved.service. May 13 00:14:40.990688 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 13 00:14:40.990720 systemd[1]: Closed systemd-networkd.socket. May 13 00:14:40.992607 systemd[1]: Stopping network-cleanup.service... May 13 00:14:41.009000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.993330 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 13 00:14:41.013740 kernel: audit: type=1131 audit(1747095281.009:62): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.993424 systemd[1]: Stopped parse-ip-for-networkd.service. May 13 00:14:41.013000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.995863 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 13 00:14:41.021073 kernel: audit: type=1131 audit(1747095281.013:63): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.021094 kernel: audit: type=1131 audit(1747095281.017:64): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.017000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.995908 systemd[1]: Stopped systemd-sysctl.service. May 13 00:14:40.998044 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 13 00:14:40.998086 systemd[1]: Stopped systemd-modules-load.service. May 13 00:14:41.024000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:40.999101 systemd[1]: Stopping systemd-udevd.service... May 13 00:14:41.031190 kernel: audit: type=1131 audit(1747095281.024:65): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.031213 kernel: audit: type=1131 audit(1747095281.027:66): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.027000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.005558 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 13 00:14:41.035388 kernel: audit: type=1131 audit(1747095281.031:67): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.031000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.008592 systemd[1]: network-cleanup.service: Deactivated successfully. May 13 00:14:41.035000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.008691 systemd[1]: Stopped network-cleanup.service. May 13 00:14:41.040988 kernel: audit: type=1131 audit(1747095281.035:68): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.012890 systemd[1]: systemd-udevd.service: Deactivated successfully. May 13 00:14:41.041000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.013011 systemd[1]: Stopped systemd-udevd.service. May 13 00:14:41.014750 systemd[1]: sysroot-boot.service: Deactivated successfully. May 13 00:14:41.051074 kernel: audit: type=1131 audit(1747095281.041:69): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.051097 kernel: audit: type=1131 audit(1747095281.047:70): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.047000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.014836 systemd[1]: Stopped sysroot-boot.service. May 13 00:14:41.051000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.018182 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 13 00:14:41.018218 systemd[1]: Closed systemd-udevd-control.socket. May 13 00:14:41.053000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.053000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:41.021848 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 13 00:14:41.021880 systemd[1]: Closed systemd-udevd-kernel.socket. May 13 00:14:41.023092 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 13 00:14:41.023135 systemd[1]: Stopped dracut-pre-udev.service. May 13 00:14:41.024657 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 13 00:14:41.024695 systemd[1]: Stopped dracut-cmdline.service. May 13 00:14:41.028317 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 13 00:14:41.028386 systemd[1]: Stopped dracut-cmdline-ask.service. May 13 00:14:41.031962 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 13 00:14:41.032002 systemd[1]: Stopped initrd-setup-root.service. May 13 00:14:41.036996 systemd[1]: Starting initrd-udevadm-cleanup-db.service... May 13 00:14:41.064000 audit: BPF prog-id=5 op=UNLOAD May 13 00:14:41.064000 audit: BPF prog-id=4 op=UNLOAD May 13 00:14:41.064000 audit: BPF prog-id=3 op=UNLOAD May 13 00:14:41.040264 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 13 00:14:41.065000 audit: BPF prog-id=8 op=UNLOAD May 13 00:14:41.065000 audit: BPF prog-id=7 op=UNLOAD May 13 00:14:41.040319 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service. May 13 00:14:41.046497 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 13 00:14:41.046542 systemd[1]: Stopped kmod-static-nodes.service. May 13 00:14:41.047836 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 13 00:14:41.047877 systemd[1]: Stopped systemd-vconsole-setup.service. May 13 00:14:41.052742 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. May 13 00:14:41.053193 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 13 00:14:41.053286 systemd[1]: Finished initrd-udevadm-cleanup-db.service. May 13 00:14:41.054454 systemd[1]: Reached target initrd-switch-root.target. May 13 00:14:41.056498 systemd[1]: Starting initrd-switch-root.service... May 13 00:14:41.062739 systemd[1]: Switching root. May 13 00:14:41.086808 iscsid[745]: iscsid shutting down. May 13 00:14:41.087484 systemd-journald[289]: Received SIGTERM from PID 1 (systemd). May 13 00:14:41.087537 systemd-journald[289]: Journal stopped May 13 00:14:43.228975 kernel: SELinux: Class mctp_socket not defined in policy. May 13 00:14:43.229025 kernel: SELinux: Class anon_inode not defined in policy. May 13 00:14:43.229041 kernel: SELinux: the above unknown classes and permissions will be allowed May 13 00:14:43.229051 kernel: SELinux: policy capability network_peer_controls=1 May 13 00:14:43.229061 kernel: SELinux: policy capability open_perms=1 May 13 00:14:43.229075 kernel: SELinux: policy capability extended_socket_class=1 May 13 00:14:43.229089 kernel: SELinux: policy capability always_check_network=0 May 13 00:14:43.229098 kernel: SELinux: policy capability cgroup_seclabel=1 May 13 00:14:43.229111 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 13 00:14:43.229120 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 13 00:14:43.229131 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 13 00:14:43.229145 systemd[1]: Successfully loaded SELinux policy in 36.133ms. May 13 00:14:43.229166 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 8.746ms. May 13 00:14:43.229181 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) May 13 00:14:43.229194 systemd[1]: Detected virtualization kvm. May 13 00:14:43.229204 systemd[1]: Detected architecture arm64. May 13 00:14:43.229214 systemd[1]: Detected first boot. May 13 00:14:43.229227 systemd[1]: Initializing machine ID from VM UUID. May 13 00:14:43.229237 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). May 13 00:14:43.229247 systemd[1]: Populated /etc with preset unit settings. May 13 00:14:43.229258 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 13 00:14:43.229269 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 13 00:14:43.229281 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 13 00:14:43.229292 systemd[1]: Queued start job for default target multi-user.target. May 13 00:14:43.229304 systemd[1]: Unnecessary job was removed for dev-vda6.device. May 13 00:14:43.229315 systemd[1]: Created slice system-addon\x2dconfig.slice. May 13 00:14:43.229325 systemd[1]: Created slice system-addon\x2drun.slice. May 13 00:14:43.229362 systemd[1]: Created slice system-getty.slice. May 13 00:14:43.229376 systemd[1]: Created slice system-modprobe.slice. May 13 00:14:43.229387 systemd[1]: Created slice system-serial\x2dgetty.slice. May 13 00:14:43.229398 systemd[1]: Created slice system-system\x2dcloudinit.slice. May 13 00:14:43.229409 systemd[1]: Created slice system-systemd\x2dfsck.slice. May 13 00:14:43.229421 systemd[1]: Created slice user.slice. May 13 00:14:43.229432 systemd[1]: Started systemd-ask-password-console.path. May 13 00:14:43.229442 systemd[1]: Started systemd-ask-password-wall.path. May 13 00:14:43.229452 systemd[1]: Set up automount boot.automount. May 13 00:14:43.229464 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. May 13 00:14:43.229475 systemd[1]: Reached target integritysetup.target. May 13 00:14:43.229485 systemd[1]: Reached target remote-cryptsetup.target. May 13 00:14:43.229496 systemd[1]: Reached target remote-fs.target. May 13 00:14:43.229510 systemd[1]: Reached target slices.target. May 13 00:14:43.229525 systemd[1]: Reached target swap.target. May 13 00:14:43.229536 systemd[1]: Reached target torcx.target. May 13 00:14:43.229546 systemd[1]: Reached target veritysetup.target. May 13 00:14:43.229558 systemd[1]: Listening on systemd-coredump.socket. May 13 00:14:43.229568 systemd[1]: Listening on systemd-initctl.socket. May 13 00:14:43.229592 systemd[1]: Listening on systemd-journald-audit.socket. May 13 00:14:43.229603 systemd[1]: Listening on systemd-journald-dev-log.socket. May 13 00:14:43.229614 systemd[1]: Listening on systemd-journald.socket. May 13 00:14:43.229624 systemd[1]: Listening on systemd-networkd.socket. May 13 00:14:43.229637 systemd[1]: Listening on systemd-udevd-control.socket. May 13 00:14:43.229648 systemd[1]: Listening on systemd-udevd-kernel.socket. May 13 00:14:43.229658 systemd[1]: Listening on systemd-userdbd.socket. May 13 00:14:43.229669 systemd[1]: Mounting dev-hugepages.mount... May 13 00:14:43.229680 systemd[1]: Mounting dev-mqueue.mount... May 13 00:14:43.229690 systemd[1]: Mounting media.mount... May 13 00:14:43.229700 systemd[1]: Mounting sys-kernel-debug.mount... May 13 00:14:43.229711 systemd[1]: Mounting sys-kernel-tracing.mount... May 13 00:14:43.229721 systemd[1]: Mounting tmp.mount... May 13 00:14:43.229733 systemd[1]: Starting flatcar-tmpfiles.service... May 13 00:14:43.229743 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 13 00:14:43.229754 systemd[1]: Starting kmod-static-nodes.service... May 13 00:14:43.229764 systemd[1]: Starting modprobe@configfs.service... May 13 00:14:43.229774 systemd[1]: Starting modprobe@dm_mod.service... May 13 00:14:43.229785 systemd[1]: Starting modprobe@drm.service... May 13 00:14:43.229795 systemd[1]: Starting modprobe@efi_pstore.service... May 13 00:14:43.229806 systemd[1]: Starting modprobe@fuse.service... May 13 00:14:43.229816 systemd[1]: Starting modprobe@loop.service... May 13 00:14:43.229829 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 13 00:14:43.229840 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. May 13 00:14:43.229850 systemd[1]: (This warning is only shown for the first unit using IP firewalling.) May 13 00:14:43.229861 systemd[1]: Starting systemd-journald.service... May 13 00:14:43.229873 kernel: fuse: init (API version 7.34) May 13 00:14:43.229883 systemd[1]: Starting systemd-modules-load.service... May 13 00:14:43.229894 systemd[1]: Starting systemd-network-generator.service... May 13 00:14:43.229904 kernel: loop: module loaded May 13 00:14:43.229914 systemd[1]: Starting systemd-remount-fs.service... May 13 00:14:43.229926 systemd[1]: Starting systemd-udev-trigger.service... May 13 00:14:43.229937 systemd[1]: Mounted dev-hugepages.mount. May 13 00:14:43.229948 systemd[1]: Mounted dev-mqueue.mount. May 13 00:14:43.229959 systemd[1]: Mounted media.mount. May 13 00:14:43.229969 systemd[1]: Mounted sys-kernel-debug.mount. May 13 00:14:43.229980 systemd[1]: Mounted sys-kernel-tracing.mount. May 13 00:14:43.229991 systemd[1]: Mounted tmp.mount. May 13 00:14:43.230001 systemd[1]: Finished kmod-static-nodes.service. May 13 00:14:43.230012 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 13 00:14:43.230024 systemd[1]: Finished modprobe@configfs.service. May 13 00:14:43.230034 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 13 00:14:43.230045 systemd[1]: Finished modprobe@dm_mod.service. May 13 00:14:43.230055 systemd[1]: modprobe@drm.service: Deactivated successfully. May 13 00:14:43.230066 systemd[1]: Finished modprobe@drm.service. May 13 00:14:43.230076 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 13 00:14:43.230087 systemd[1]: Finished modprobe@efi_pstore.service. May 13 00:14:43.230101 systemd-journald[1033]: Journal started May 13 00:14:43.230145 systemd-journald[1033]: Runtime Journal (/run/log/journal/7878dbb6338145bcab4321800b19da4e) is 6.0M, max 48.7M, 42.6M free. May 13 00:14:43.104000 audit[1]: AVC avc: denied { audit_read } for pid=1 comm="systemd" capability=37 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 May 13 00:14:43.104000 audit[1]: EVENT_LISTENER pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 May 13 00:14:43.217000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.220000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.220000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.224000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.224000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.227000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.227000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.227000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 May 13 00:14:43.227000 audit[1033]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=60 a0=6 a1=ffffd30e2a20 a2=4000 a3=1 items=0 ppid=1 pid=1033 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:43.227000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" May 13 00:14:43.230000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.230000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.232644 systemd[1]: Started systemd-journald.service. May 13 00:14:43.232000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.233940 systemd[1]: Finished flatcar-tmpfiles.service. May 13 00:14:43.234000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.235082 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 13 00:14:43.235279 systemd[1]: Finished modprobe@fuse.service. May 13 00:14:43.235000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.235000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.236434 systemd[1]: modprobe@loop.service: Deactivated successfully. May 13 00:14:43.236664 systemd[1]: Finished modprobe@loop.service. May 13 00:14:43.237000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.237000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.237890 systemd[1]: Finished systemd-modules-load.service. May 13 00:14:43.238000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.239219 systemd[1]: Finished systemd-network-generator.service. May 13 00:14:43.239000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.240623 systemd[1]: Finished systemd-remount-fs.service. May 13 00:14:43.241000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.241993 systemd[1]: Reached target network-pre.target. May 13 00:14:43.244076 systemd[1]: Mounting sys-fs-fuse-connections.mount... May 13 00:14:43.245928 systemd[1]: Mounting sys-kernel-config.mount... May 13 00:14:43.246670 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 13 00:14:43.248511 systemd[1]: Starting systemd-hwdb-update.service... May 13 00:14:43.251081 systemd[1]: Starting systemd-journal-flush.service... May 13 00:14:43.252334 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 13 00:14:43.257103 systemd[1]: Starting systemd-random-seed.service... May 13 00:14:43.257994 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 13 00:14:43.259163 systemd[1]: Starting systemd-sysctl.service... May 13 00:14:43.261220 systemd[1]: Starting systemd-sysusers.service... May 13 00:14:43.265000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.265001 systemd[1]: Finished systemd-udev-trigger.service. May 13 00:14:43.266096 systemd[1]: Mounted sys-fs-fuse-connections.mount. May 13 00:14:43.267457 systemd[1]: Mounted sys-kernel-config.mount. May 13 00:14:43.268765 systemd-journald[1033]: Time spent on flushing to /var/log/journal/7878dbb6338145bcab4321800b19da4e is 11.512ms for 939 entries. May 13 00:14:43.268765 systemd-journald[1033]: System Journal (/var/log/journal/7878dbb6338145bcab4321800b19da4e) is 8.0M, max 195.6M, 187.6M free. May 13 00:14:43.381963 systemd-journald[1033]: Received client request to flush runtime journal. May 13 00:14:43.286000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.295000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.305000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.322000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.269751 systemd[1]: Starting systemd-udev-settle.service... May 13 00:14:43.382483 udevadm[1079]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. May 13 00:14:43.285485 systemd[1]: Finished systemd-sysctl.service. May 13 00:14:43.295089 systemd[1]: Finished systemd-sysusers.service. May 13 00:14:43.297249 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... May 13 00:14:43.304867 systemd[1]: Finished systemd-random-seed.service. May 13 00:14:43.305942 systemd[1]: Reached target first-boot-complete.target. May 13 00:14:43.321426 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. May 13 00:14:43.382995 systemd[1]: Finished systemd-journal-flush.service. May 13 00:14:43.383000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.621614 systemd[1]: Finished systemd-hwdb-update.service. May 13 00:14:43.622000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.623733 systemd[1]: Starting systemd-udevd.service... May 13 00:14:43.641818 systemd-udevd[1091]: Using default interface naming scheme 'v252'. May 13 00:14:43.654566 systemd[1]: Started systemd-udevd.service. May 13 00:14:43.655000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.658127 systemd[1]: Starting systemd-networkd.service... May 13 00:14:43.663290 systemd[1]: Starting systemd-userdbd.service... May 13 00:14:43.679896 systemd[1]: Found device dev-ttyAMA0.device. May 13 00:14:43.720108 systemd[1]: Started systemd-userdbd.service. May 13 00:14:43.720000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.748177 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. May 13 00:14:43.785681 systemd[1]: Finished systemd-udev-settle.service. May 13 00:14:43.786014 systemd-networkd[1102]: lo: Link UP May 13 00:14:43.786265 systemd-networkd[1102]: lo: Gained carrier May 13 00:14:43.786000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.786830 systemd-networkd[1102]: Enumeration completed May 13 00:14:43.787016 systemd-networkd[1102]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 13 00:14:43.789000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.788021 systemd[1]: Starting lvm2-activation-early.service... May 13 00:14:43.789074 systemd[1]: Started systemd-networkd.service. May 13 00:14:43.794282 systemd-networkd[1102]: eth0: Link UP May 13 00:14:43.794417 systemd-networkd[1102]: eth0: Gained carrier May 13 00:14:43.805433 lvm[1125]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 13 00:14:43.815474 systemd-networkd[1102]: eth0: DHCPv4 address 10.0.0.8/16, gateway 10.0.0.1 acquired from 10.0.0.1 May 13 00:14:43.830241 systemd[1]: Finished lvm2-activation-early.service. May 13 00:14:43.830000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.831304 systemd[1]: Reached target cryptsetup.target. May 13 00:14:43.833317 systemd[1]: Starting lvm2-activation.service... May 13 00:14:43.837010 lvm[1127]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 13 00:14:43.861272 systemd[1]: Finished lvm2-activation.service. May 13 00:14:43.861000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.862252 systemd[1]: Reached target local-fs-pre.target. May 13 00:14:43.863118 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 13 00:14:43.863146 systemd[1]: Reached target local-fs.target. May 13 00:14:43.863910 systemd[1]: Reached target machines.target. May 13 00:14:43.865858 systemd[1]: Starting ldconfig.service... May 13 00:14:43.866862 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 13 00:14:43.866910 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 13 00:14:43.867922 systemd[1]: Starting systemd-boot-update.service... May 13 00:14:43.869754 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... May 13 00:14:43.871998 systemd[1]: Starting systemd-machine-id-commit.service... May 13 00:14:43.873992 systemd[1]: Starting systemd-sysext.service... May 13 00:14:43.875066 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1130 (bootctl) May 13 00:14:43.876218 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... May 13 00:14:43.881216 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. May 13 00:14:43.882000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.886426 systemd[1]: Unmounting usr-share-oem.mount... May 13 00:14:43.890240 systemd[1]: usr-share-oem.mount: Deactivated successfully. May 13 00:14:43.890513 systemd[1]: Unmounted usr-share-oem.mount. May 13 00:14:43.934372 kernel: loop0: detected capacity change from 0 to 194096 May 13 00:14:43.941003 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 13 00:14:43.941696 systemd[1]: Finished systemd-machine-id-commit.service. May 13 00:14:43.942000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.949416 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 13 00:14:43.957575 systemd-fsck[1142]: fsck.fat 4.2 (2021-01-31) May 13 00:14:43.957575 systemd-fsck[1142]: /dev/vda1: 236 files, 117310/258078 clusters May 13 00:14:43.959902 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. May 13 00:14:43.961000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:43.962898 systemd[1]: Mounting boot.mount... May 13 00:14:43.969546 systemd[1]: Mounted boot.mount. May 13 00:14:43.977587 systemd[1]: Finished systemd-boot-update.service. May 13 00:14:43.978000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.017469 kernel: loop1: detected capacity change from 0 to 194096 May 13 00:14:44.021607 (sd-sysext)[1152]: Using extensions 'kubernetes'. May 13 00:14:44.022023 (sd-sysext)[1152]: Merged extensions into '/usr'. May 13 00:14:44.039244 ldconfig[1129]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 13 00:14:44.039264 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 13 00:14:44.040499 systemd[1]: Starting modprobe@dm_mod.service... May 13 00:14:44.042334 systemd[1]: Starting modprobe@efi_pstore.service... May 13 00:14:44.044251 systemd[1]: Starting modprobe@loop.service... May 13 00:14:44.045069 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 13 00:14:44.045205 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 13 00:14:44.046060 systemd[1]: Finished ldconfig.service. May 13 00:14:44.046000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ldconfig comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.047203 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 13 00:14:44.047363 systemd[1]: Finished modprobe@dm_mod.service. May 13 00:14:44.047000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.047000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.048610 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 13 00:14:44.048740 systemd[1]: Finished modprobe@efi_pstore.service. May 13 00:14:44.049000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.049000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.049974 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 13 00:14:44.051104 systemd[1]: modprobe@loop.service: Deactivated successfully. May 13 00:14:44.051263 systemd[1]: Finished modprobe@loop.service. May 13 00:14:44.051000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.051000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.052467 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 13 00:14:44.208543 systemd[1]: Mounting usr-share-oem.mount... May 13 00:14:44.213854 systemd[1]: Mounted usr-share-oem.mount. May 13 00:14:44.215812 systemd[1]: Finished systemd-sysext.service. May 13 00:14:44.216000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.217937 systemd[1]: Starting ensure-sysext.service... May 13 00:14:44.219802 systemd[1]: Starting systemd-tmpfiles-setup.service... May 13 00:14:44.224274 systemd[1]: Reloading. May 13 00:14:44.229049 systemd-tmpfiles[1167]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. May 13 00:14:44.230243 systemd-tmpfiles[1167]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 13 00:14:44.231579 systemd-tmpfiles[1167]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 13 00:14:44.265431 /usr/lib/systemd/system-generators/torcx-generator[1187]: time="2025-05-13T00:14:44Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 13 00:14:44.265457 /usr/lib/systemd/system-generators/torcx-generator[1187]: time="2025-05-13T00:14:44Z" level=info msg="torcx already run" May 13 00:14:44.343769 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 13 00:14:44.343790 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 13 00:14:44.360639 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 13 00:14:44.402133 systemd[1]: Finished systemd-tmpfiles-setup.service. May 13 00:14:44.403000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.406464 systemd[1]: Starting audit-rules.service... May 13 00:14:44.408721 systemd[1]: Starting clean-ca-certificates.service... May 13 00:14:44.410899 systemd[1]: Starting systemd-journal-catalog-update.service... May 13 00:14:44.413758 systemd[1]: Starting systemd-resolved.service... May 13 00:14:44.416239 systemd[1]: Starting systemd-timesyncd.service... May 13 00:14:44.418763 systemd[1]: Starting systemd-update-utmp.service... May 13 00:14:44.420303 systemd[1]: Finished clean-ca-certificates.service. May 13 00:14:44.421000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.423481 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 13 00:14:44.425041 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 13 00:14:44.426296 systemd[1]: Starting modprobe@dm_mod.service... May 13 00:14:44.430000 audit[1245]: SYSTEM_BOOT pid=1245 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' May 13 00:14:44.431134 systemd[1]: Starting modprobe@efi_pstore.service... May 13 00:14:44.433187 systemd[1]: Starting modprobe@loop.service... May 13 00:14:44.434153 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 13 00:14:44.434287 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 13 00:14:44.434413 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 13 00:14:44.435250 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 13 00:14:44.435448 systemd[1]: Finished modprobe@dm_mod.service. May 13 00:14:44.436000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.436000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.436757 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 13 00:14:44.436904 systemd[1]: Finished modprobe@efi_pstore.service. May 13 00:14:44.437000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.437000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.438240 systemd[1]: modprobe@loop.service: Deactivated successfully. May 13 00:14:44.438443 systemd[1]: Finished modprobe@loop.service. May 13 00:14:44.439000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.439000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.442681 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 13 00:14:44.442821 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 13 00:14:44.444502 systemd[1]: Finished systemd-update-utmp.service. May 13 00:14:44.445000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.447016 systemd[1]: Finished systemd-journal-catalog-update.service. May 13 00:14:44.447000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.448509 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 13 00:14:44.449889 systemd[1]: Starting modprobe@dm_mod.service... May 13 00:14:44.451870 systemd[1]: Starting modprobe@efi_pstore.service... May 13 00:14:44.453936 systemd[1]: Starting modprobe@loop.service... May 13 00:14:44.454783 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 13 00:14:44.454914 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 13 00:14:44.456499 systemd[1]: Starting systemd-update-done.service... May 13 00:14:44.457412 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 13 00:14:44.458723 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 13 00:14:44.458909 systemd[1]: Finished modprobe@dm_mod.service. May 13 00:14:44.459000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.459000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.460157 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 13 00:14:44.461000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.461000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.462000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.462000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.460310 systemd[1]: Finished modprobe@efi_pstore.service. May 13 00:14:44.461764 systemd[1]: modprobe@loop.service: Deactivated successfully. May 13 00:14:44.461990 systemd[1]: Finished modprobe@loop.service. May 13 00:14:44.463187 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 13 00:14:44.463274 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 13 00:14:44.467779 systemd[1]: Finished systemd-update-done.service. May 13 00:14:44.468000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-done comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.471420 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 13 00:14:44.472900 systemd[1]: Starting modprobe@dm_mod.service... May 13 00:14:44.475049 systemd[1]: Starting modprobe@drm.service... May 13 00:14:44.478277 systemd[1]: Starting modprobe@efi_pstore.service... May 13 00:14:44.480591 systemd[1]: Starting modprobe@loop.service... May 13 00:14:44.481527 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 13 00:14:44.481697 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 13 00:14:44.483099 systemd[1]: Starting systemd-networkd-wait-online.service... May 13 00:14:44.484202 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 13 00:14:44.485422 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 13 00:14:44.485586 systemd[1]: Finished modprobe@dm_mod.service. May 13 00:14:44.486000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.486000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.486931 systemd[1]: modprobe@drm.service: Deactivated successfully. May 13 00:14:44.487087 systemd[1]: Finished modprobe@drm.service. May 13 00:14:44.488454 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 13 00:14:44.488615 systemd[1]: Finished modprobe@efi_pstore.service. May 13 00:14:44.487000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.487000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.489977 systemd[1]: modprobe@loop.service: Deactivated successfully. May 13 00:14:44.490152 systemd[1]: Finished modprobe@loop.service. May 13 00:14:44.489000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.489000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.490000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.490000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.491631 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 13 00:14:44.491726 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 13 00:14:44.492964 systemd[1]: Finished ensure-sysext.service. May 13 00:14:44.493000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ensure-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:44.504000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 May 13 00:14:44.504000 audit[1283]: SYSCALL arch=c00000b7 syscall=206 success=yes exit=1056 a0=3 a1=ffffdd042940 a2=420 a3=0 items=0 ppid=1233 pid=1283 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:44.504000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 May 13 00:14:44.505507 augenrules[1283]: No rules May 13 00:14:44.505561 systemd[1]: Started systemd-timesyncd.service. May 13 00:14:44.069532 systemd-timesyncd[1239]: Contacted time server 10.0.0.1:123 (10.0.0.1). May 13 00:14:44.095810 systemd-journald[1033]: Time jumped backwards, rotating. May 13 00:14:44.069622 systemd-timesyncd[1239]: Initial clock synchronization to Tue 2025-05-13 00:14:44.069438 UTC. May 13 00:14:44.071149 systemd[1]: Finished audit-rules.service. May 13 00:14:44.072261 systemd[1]: Reached target time-set.target. May 13 00:14:44.080397 systemd-resolved[1238]: Positive Trust Anchors: May 13 00:14:44.080405 systemd-resolved[1238]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 13 00:14:44.080432 systemd-resolved[1238]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test May 13 00:14:44.108959 systemd-resolved[1238]: Defaulting to hostname 'linux'. May 13 00:14:44.110543 systemd[1]: Started systemd-resolved.service. May 13 00:14:44.111477 systemd[1]: Reached target network.target. May 13 00:14:44.112337 systemd[1]: Reached target nss-lookup.target. May 13 00:14:44.113187 systemd[1]: Reached target sysinit.target. May 13 00:14:44.114076 systemd[1]: Started motdgen.path. May 13 00:14:44.114831 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. May 13 00:14:44.116154 systemd[1]: Started logrotate.timer. May 13 00:14:44.117030 systemd[1]: Started mdadm.timer. May 13 00:14:44.117760 systemd[1]: Started systemd-tmpfiles-clean.timer. May 13 00:14:44.118634 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 13 00:14:44.118664 systemd[1]: Reached target paths.target. May 13 00:14:44.119404 systemd[1]: Reached target timers.target. May 13 00:14:44.120703 systemd[1]: Listening on dbus.socket. May 13 00:14:44.122654 systemd[1]: Starting docker.socket... May 13 00:14:44.124418 systemd[1]: Listening on sshd.socket. May 13 00:14:44.125290 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 13 00:14:44.125778 systemd[1]: Listening on docker.socket. May 13 00:14:44.126559 systemd[1]: Reached target sockets.target. May 13 00:14:44.127320 systemd[1]: Reached target basic.target. May 13 00:14:44.128272 systemd[1]: System is tainted: cgroupsv1 May 13 00:14:44.128324 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. May 13 00:14:44.128346 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. May 13 00:14:44.129500 systemd[1]: Starting containerd.service... May 13 00:14:44.131567 systemd[1]: Starting dbus.service... May 13 00:14:44.133344 systemd[1]: Starting enable-oem-cloudinit.service... May 13 00:14:44.135612 systemd[1]: Starting extend-filesystems.service... May 13 00:14:44.136554 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). May 13 00:14:44.138112 systemd[1]: Starting motdgen.service... May 13 00:14:44.142408 systemd[1]: Starting prepare-helm.service... May 13 00:14:44.143764 jq[1296]: false May 13 00:14:44.146878 systemd[1]: Starting ssh-key-proc-cmdline.service... May 13 00:14:44.149028 systemd[1]: Starting sshd-keygen.service... May 13 00:14:44.151739 systemd[1]: Starting systemd-logind.service... May 13 00:14:44.152967 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 13 00:14:44.153070 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 13 00:14:44.157128 systemd[1]: Starting update-engine.service... May 13 00:14:44.159264 systemd[1]: Starting update-ssh-keys-after-ignition.service... May 13 00:14:44.162059 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 13 00:14:44.162740 jq[1316]: true May 13 00:14:44.162327 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. May 13 00:14:44.162727 systemd[1]: motdgen.service: Deactivated successfully. May 13 00:14:44.162954 systemd[1]: Finished motdgen.service. May 13 00:14:44.170554 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 13 00:14:44.170864 systemd[1]: Finished ssh-key-proc-cmdline.service. May 13 00:14:44.179472 extend-filesystems[1297]: Found loop1 May 13 00:14:44.179472 extend-filesystems[1297]: Found vda May 13 00:14:44.179472 extend-filesystems[1297]: Found vda1 May 13 00:14:44.179472 extend-filesystems[1297]: Found vda2 May 13 00:14:44.179472 extend-filesystems[1297]: Found vda3 May 13 00:14:44.179472 extend-filesystems[1297]: Found usr May 13 00:14:44.179472 extend-filesystems[1297]: Found vda4 May 13 00:14:44.179472 extend-filesystems[1297]: Found vda6 May 13 00:14:44.179472 extend-filesystems[1297]: Found vda7 May 13 00:14:44.179472 extend-filesystems[1297]: Found vda9 May 13 00:14:44.179472 extend-filesystems[1297]: Checking size of /dev/vda9 May 13 00:14:44.208817 jq[1321]: true May 13 00:14:44.209113 tar[1320]: linux-arm64/helm May 13 00:14:44.210234 dbus-daemon[1295]: [system] SELinux support is enabled May 13 00:14:44.210578 systemd[1]: Started dbus.service. May 13 00:14:44.214184 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 13 00:14:44.214208 systemd[1]: Reached target system-config.target. May 13 00:14:44.215195 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 13 00:14:44.215221 systemd[1]: Reached target user-config.target. May 13 00:14:44.222642 extend-filesystems[1297]: Resized partition /dev/vda9 May 13 00:14:44.233448 extend-filesystems[1347]: resize2fs 1.46.5 (30-Dec-2021) May 13 00:14:44.250227 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks May 13 00:14:44.264783 systemd-logind[1306]: Watching system buttons on /dev/input/event0 (Power Button) May 13 00:14:44.265037 systemd-logind[1306]: New seat seat0. May 13 00:14:44.270588 systemd[1]: Started systemd-logind.service. May 13 00:14:44.277121 update_engine[1314]: I0513 00:14:44.273051 1314 main.cc:92] Flatcar Update Engine starting May 13 00:14:44.280524 systemd[1]: Started update-engine.service. May 13 00:14:44.280788 update_engine[1314]: I0513 00:14:44.280764 1314 update_check_scheduler.cc:74] Next update check in 8m2s May 13 00:14:44.283253 kernel: EXT4-fs (vda9): resized filesystem to 1864699 May 13 00:14:44.295178 systemd[1]: Started locksmithd.service. May 13 00:14:44.296894 extend-filesystems[1347]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required May 13 00:14:44.296894 extend-filesystems[1347]: old_desc_blocks = 1, new_desc_blocks = 1 May 13 00:14:44.296894 extend-filesystems[1347]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. May 13 00:14:44.302923 extend-filesystems[1297]: Resized filesystem in /dev/vda9 May 13 00:14:44.298328 systemd[1]: extend-filesystems.service: Deactivated successfully. May 13 00:14:44.304416 bash[1348]: Updated "/home/core/.ssh/authorized_keys" May 13 00:14:44.298611 systemd[1]: Finished extend-filesystems.service. May 13 00:14:44.305092 systemd[1]: Finished update-ssh-keys-after-ignition.service. May 13 00:14:44.348558 env[1323]: time="2025-05-13T00:14:44.348429053Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 May 13 00:14:44.376027 env[1323]: time="2025-05-13T00:14:44.375975253Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 May 13 00:14:44.376299 env[1323]: time="2025-05-13T00:14:44.376276773Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 May 13 00:14:44.377699 env[1323]: time="2025-05-13T00:14:44.377654093Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.181-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 May 13 00:14:44.377699 env[1323]: time="2025-05-13T00:14:44.377692653Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 May 13 00:14:44.377987 env[1323]: time="2025-05-13T00:14:44.377960013Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 May 13 00:14:44.377987 env[1323]: time="2025-05-13T00:14:44.377985213Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 May 13 00:14:44.378046 env[1323]: time="2025-05-13T00:14:44.378000213Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" May 13 00:14:44.378046 env[1323]: time="2025-05-13T00:14:44.378010973Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 May 13 00:14:44.378097 env[1323]: time="2025-05-13T00:14:44.378079773Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 May 13 00:14:44.378435 env[1323]: time="2025-05-13T00:14:44.378411853Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 May 13 00:14:44.378618 env[1323]: time="2025-05-13T00:14:44.378594013Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 May 13 00:14:44.378654 env[1323]: time="2025-05-13T00:14:44.378616773Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 May 13 00:14:44.378691 env[1323]: time="2025-05-13T00:14:44.378671173Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" May 13 00:14:44.378691 env[1323]: time="2025-05-13T00:14:44.378689493Z" level=info msg="metadata content store policy set" policy=shared May 13 00:14:44.399647 env[1323]: time="2025-05-13T00:14:44.396917533Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 May 13 00:14:44.399647 env[1323]: time="2025-05-13T00:14:44.396975373Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 May 13 00:14:44.399647 env[1323]: time="2025-05-13T00:14:44.396989413Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 May 13 00:14:44.399647 env[1323]: time="2025-05-13T00:14:44.397030853Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 May 13 00:14:44.399647 env[1323]: time="2025-05-13T00:14:44.397046613Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 May 13 00:14:44.399647 env[1323]: time="2025-05-13T00:14:44.397061453Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 May 13 00:14:44.399647 env[1323]: time="2025-05-13T00:14:44.397074293Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 May 13 00:14:44.399647 env[1323]: time="2025-05-13T00:14:44.397428293Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 May 13 00:14:44.399647 env[1323]: time="2025-05-13T00:14:44.397447653Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 May 13 00:14:44.399647 env[1323]: time="2025-05-13T00:14:44.397461613Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 May 13 00:14:44.399647 env[1323]: time="2025-05-13T00:14:44.397475453Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 May 13 00:14:44.399647 env[1323]: time="2025-05-13T00:14:44.397489053Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 May 13 00:14:44.399647 env[1323]: time="2025-05-13T00:14:44.397668613Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 May 13 00:14:44.399647 env[1323]: time="2025-05-13T00:14:44.397750773Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398049213Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398093533Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398109133Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398394613Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398407093Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398419413Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398431013Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398442973Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398455773Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398466853Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398481013Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398493373Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398645293Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398663453Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 May 13 00:14:44.399992 env[1323]: time="2025-05-13T00:14:44.398682253Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 May 13 00:14:44.401681 env[1323]: time="2025-05-13T00:14:44.398693893Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 May 13 00:14:44.401681 env[1323]: time="2025-05-13T00:14:44.398707933Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 May 13 00:14:44.401681 env[1323]: time="2025-05-13T00:14:44.398720813Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 May 13 00:14:44.401681 env[1323]: time="2025-05-13T00:14:44.398739293Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" May 13 00:14:44.401681 env[1323]: time="2025-05-13T00:14:44.398778413Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 May 13 00:14:44.400336 systemd[1]: Started containerd.service. May 13 00:14:44.401834 env[1323]: time="2025-05-13T00:14:44.398975253Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" May 13 00:14:44.401834 env[1323]: time="2025-05-13T00:14:44.399030293Z" level=info msg="Connect containerd service" May 13 00:14:44.401834 env[1323]: time="2025-05-13T00:14:44.399064933Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" May 13 00:14:44.401834 env[1323]: time="2025-05-13T00:14:44.399762133Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 13 00:14:44.401834 env[1323]: time="2025-05-13T00:14:44.400128173Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 13 00:14:44.401834 env[1323]: time="2025-05-13T00:14:44.400172333Z" level=info msg=serving... address=/run/containerd/containerd.sock May 13 00:14:44.401834 env[1323]: time="2025-05-13T00:14:44.400226693Z" level=info msg="containerd successfully booted in 0.060480s" May 13 00:14:44.403764 env[1323]: time="2025-05-13T00:14:44.402624813Z" level=info msg="Start subscribing containerd event" May 13 00:14:44.403764 env[1323]: time="2025-05-13T00:14:44.402681453Z" level=info msg="Start recovering state" May 13 00:14:44.403764 env[1323]: time="2025-05-13T00:14:44.402761093Z" level=info msg="Start event monitor" May 13 00:14:44.403764 env[1323]: time="2025-05-13T00:14:44.402783053Z" level=info msg="Start snapshots syncer" May 13 00:14:44.403764 env[1323]: time="2025-05-13T00:14:44.402793653Z" level=info msg="Start cni network conf syncer for default" May 13 00:14:44.403764 env[1323]: time="2025-05-13T00:14:44.402805133Z" level=info msg="Start streaming server" May 13 00:14:44.408791 locksmithd[1356]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 13 00:14:44.586645 systemd-networkd[1102]: eth0: Gained IPv6LL May 13 00:14:44.588522 systemd[1]: Finished systemd-networkd-wait-online.service. May 13 00:14:44.589878 systemd[1]: Reached target network-online.target. May 13 00:14:44.592429 systemd[1]: Starting kubelet.service... May 13 00:14:44.639805 tar[1320]: linux-arm64/LICENSE May 13 00:14:44.639904 tar[1320]: linux-arm64/README.md May 13 00:14:44.645324 systemd[1]: Finished prepare-helm.service. May 13 00:14:45.200482 systemd[1]: Started kubelet.service. May 13 00:14:45.731414 kubelet[1381]: E0513 00:14:45.731363 1381 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 13 00:14:45.733348 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 13 00:14:45.733494 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 13 00:14:47.669268 sshd_keygen[1326]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 13 00:14:47.686357 systemd[1]: Finished sshd-keygen.service. May 13 00:14:47.688879 systemd[1]: Starting issuegen.service... May 13 00:14:47.693504 systemd[1]: issuegen.service: Deactivated successfully. May 13 00:14:47.693738 systemd[1]: Finished issuegen.service. May 13 00:14:47.696046 systemd[1]: Starting systemd-user-sessions.service... May 13 00:14:47.701841 systemd[1]: Finished systemd-user-sessions.service. May 13 00:14:47.704237 systemd[1]: Started getty@tty1.service. May 13 00:14:47.706387 systemd[1]: Started serial-getty@ttyAMA0.service. May 13 00:14:47.707501 systemd[1]: Reached target getty.target. May 13 00:14:47.708561 systemd[1]: Reached target multi-user.target. May 13 00:14:47.710747 systemd[1]: Starting systemd-update-utmp-runlevel.service... May 13 00:14:47.716906 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. May 13 00:14:47.717143 systemd[1]: Finished systemd-update-utmp-runlevel.service. May 13 00:14:47.718358 systemd[1]: Startup finished in 5.171s (kernel) + 7.014s (userspace) = 12.186s. May 13 00:14:48.917887 systemd[1]: Created slice system-sshd.slice. May 13 00:14:48.920016 systemd[1]: Started sshd@0-10.0.0.8:22-10.0.0.1:49756.service. May 13 00:14:48.974495 sshd[1408]: Accepted publickey for core from 10.0.0.1 port 49756 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:14:48.976488 sshd[1408]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:14:48.987720 systemd-logind[1306]: New session 1 of user core. May 13 00:14:48.988298 systemd[1]: Created slice user-500.slice. May 13 00:14:48.989461 systemd[1]: Starting user-runtime-dir@500.service... May 13 00:14:48.998788 systemd[1]: Finished user-runtime-dir@500.service. May 13 00:14:49.000486 systemd[1]: Starting user@500.service... May 13 00:14:49.004118 (systemd)[1413]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 13 00:14:49.069239 systemd[1413]: Queued start job for default target default.target. May 13 00:14:49.069476 systemd[1413]: Reached target paths.target. May 13 00:14:49.069491 systemd[1413]: Reached target sockets.target. May 13 00:14:49.069502 systemd[1413]: Reached target timers.target. May 13 00:14:49.069513 systemd[1413]: Reached target basic.target. May 13 00:14:49.069590 systemd[1413]: Reached target default.target. May 13 00:14:49.069614 systemd[1413]: Startup finished in 59ms. May 13 00:14:49.069684 systemd[1]: Started user@500.service. May 13 00:14:49.070630 systemd[1]: Started session-1.scope. May 13 00:14:49.121384 systemd[1]: Started sshd@1-10.0.0.8:22-10.0.0.1:49770.service. May 13 00:14:49.157117 sshd[1422]: Accepted publickey for core from 10.0.0.1 port 49770 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:14:49.158392 sshd[1422]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:14:49.162691 systemd-logind[1306]: New session 2 of user core. May 13 00:14:49.163391 systemd[1]: Started session-2.scope. May 13 00:14:49.220153 sshd[1422]: pam_unix(sshd:session): session closed for user core May 13 00:14:49.222431 systemd[1]: Started sshd@2-10.0.0.8:22-10.0.0.1:49784.service. May 13 00:14:49.223072 systemd[1]: sshd@1-10.0.0.8:22-10.0.0.1:49770.service: Deactivated successfully. May 13 00:14:49.223967 systemd-logind[1306]: Session 2 logged out. Waiting for processes to exit. May 13 00:14:49.224033 systemd[1]: session-2.scope: Deactivated successfully. May 13 00:14:49.225157 systemd-logind[1306]: Removed session 2. May 13 00:14:49.254064 sshd[1427]: Accepted publickey for core from 10.0.0.1 port 49784 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:14:49.255227 sshd[1427]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:14:49.258580 systemd-logind[1306]: New session 3 of user core. May 13 00:14:49.259317 systemd[1]: Started session-3.scope. May 13 00:14:49.311237 sshd[1427]: pam_unix(sshd:session): session closed for user core May 13 00:14:49.314137 systemd[1]: Started sshd@3-10.0.0.8:22-10.0.0.1:49794.service. May 13 00:14:49.314602 systemd[1]: sshd@2-10.0.0.8:22-10.0.0.1:49784.service: Deactivated successfully. May 13 00:14:49.315511 systemd[1]: session-3.scope: Deactivated successfully. May 13 00:14:49.315512 systemd-logind[1306]: Session 3 logged out. Waiting for processes to exit. May 13 00:14:49.316747 systemd-logind[1306]: Removed session 3. May 13 00:14:49.345683 sshd[1435]: Accepted publickey for core from 10.0.0.1 port 49794 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:14:49.346943 sshd[1435]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:14:49.350459 systemd-logind[1306]: New session 4 of user core. May 13 00:14:49.351018 systemd[1]: Started session-4.scope. May 13 00:14:49.404166 sshd[1435]: pam_unix(sshd:session): session closed for user core May 13 00:14:49.407001 systemd[1]: Started sshd@4-10.0.0.8:22-10.0.0.1:49808.service. May 13 00:14:49.407591 systemd[1]: sshd@3-10.0.0.8:22-10.0.0.1:49794.service: Deactivated successfully. May 13 00:14:49.408560 systemd-logind[1306]: Session 4 logged out. Waiting for processes to exit. May 13 00:14:49.408741 systemd[1]: session-4.scope: Deactivated successfully. May 13 00:14:49.409651 systemd-logind[1306]: Removed session 4. May 13 00:14:49.439228 sshd[1442]: Accepted publickey for core from 10.0.0.1 port 49808 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:14:49.440519 sshd[1442]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:14:49.443802 systemd-logind[1306]: New session 5 of user core. May 13 00:14:49.444753 systemd[1]: Started session-5.scope. May 13 00:14:49.503958 sudo[1447]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 May 13 00:14:49.504510 sudo[1447]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) May 13 00:14:49.518696 dbus-daemon[1295]: avc: received setenforce notice (enforcing=1) May 13 00:14:49.519587 sudo[1447]: pam_unix(sudo:session): session closed for user root May 13 00:14:49.530946 sshd[1442]: pam_unix(sshd:session): session closed for user core May 13 00:14:49.534009 systemd[1]: Started sshd@5-10.0.0.8:22-10.0.0.1:49820.service. May 13 00:14:49.535883 systemd[1]: sshd@4-10.0.0.8:22-10.0.0.1:49808.service: Deactivated successfully. May 13 00:14:49.537094 systemd-logind[1306]: Session 5 logged out. Waiting for processes to exit. May 13 00:14:49.537171 systemd[1]: session-5.scope: Deactivated successfully. May 13 00:14:49.537921 systemd-logind[1306]: Removed session 5. May 13 00:14:49.568957 sshd[1449]: Accepted publickey for core from 10.0.0.1 port 49820 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:14:49.570328 sshd[1449]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:14:49.573594 systemd-logind[1306]: New session 6 of user core. May 13 00:14:49.574407 systemd[1]: Started session-6.scope. May 13 00:14:49.627950 sudo[1456]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules May 13 00:14:49.628175 sudo[1456]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) May 13 00:14:49.630853 sudo[1456]: pam_unix(sudo:session): session closed for user root May 13 00:14:49.635560 sudo[1455]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules May 13 00:14:49.635786 sudo[1455]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) May 13 00:14:49.645025 systemd[1]: Stopping audit-rules.service... May 13 00:14:49.645000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 May 13 00:14:49.646718 auditctl[1459]: No rules May 13 00:14:49.647161 kernel: kauditd_printk_skb: 97 callbacks suppressed May 13 00:14:49.647199 kernel: audit: type=1305 audit(1747095289.645:157): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 May 13 00:14:49.647131 systemd[1]: audit-rules.service: Deactivated successfully. May 13 00:14:49.647342 systemd[1]: Stopped audit-rules.service. May 13 00:14:49.645000 audit[1459]: SYSCALL arch=c00000b7 syscall=206 success=yes exit=1056 a0=3 a1=ffffeca84390 a2=420 a3=0 items=0 ppid=1 pid=1459 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:49.648961 systemd[1]: Starting audit-rules.service... May 13 00:14:49.652854 kernel: audit: type=1300 audit(1747095289.645:157): arch=c00000b7 syscall=206 success=yes exit=1056 a0=3 a1=ffffeca84390 a2=420 a3=0 items=0 ppid=1 pid=1459 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:49.652925 kernel: audit: type=1327 audit(1747095289.645:157): proctitle=2F7362696E2F617564697463746C002D44 May 13 00:14:49.645000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D44 May 13 00:14:49.646000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:49.656477 kernel: audit: type=1131 audit(1747095289.646:158): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:49.666881 augenrules[1477]: No rules May 13 00:14:49.667874 systemd[1]: Finished audit-rules.service. May 13 00:14:49.667000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:49.668753 sudo[1455]: pam_unix(sudo:session): session closed for user root May 13 00:14:49.670365 sshd[1449]: pam_unix(sshd:session): session closed for user core May 13 00:14:49.668000 audit[1455]: USER_END pid=1455 uid=500 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 13 00:14:49.674569 kernel: audit: type=1130 audit(1747095289.667:159): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:49.674627 kernel: audit: type=1106 audit(1747095289.668:160): pid=1455 uid=500 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 13 00:14:49.674663 kernel: audit: type=1104 audit(1747095289.668:161): pid=1455 uid=500 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 13 00:14:49.668000 audit[1455]: CRED_DISP pid=1455 uid=500 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 13 00:14:49.672699 systemd[1]: Started sshd@6-10.0.0.8:22-10.0.0.1:49822.service. May 13 00:14:49.672000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-10.0.0.8:22-10.0.0.1:49822 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:49.680559 kernel: audit: type=1130 audit(1747095289.672:162): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-10.0.0.8:22-10.0.0.1:49822 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:49.680614 kernel: audit: type=1106 audit(1747095289.677:163): pid=1449 uid=0 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:14:49.677000 audit[1449]: USER_END pid=1449 uid=0 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:14:49.679745 systemd[1]: sshd@5-10.0.0.8:22-10.0.0.1:49820.service: Deactivated successfully. May 13 00:14:49.681161 systemd[1]: session-6.scope: Deactivated successfully. May 13 00:14:49.681771 systemd-logind[1306]: Session 6 logged out. Waiting for processes to exit. May 13 00:14:49.682646 systemd-logind[1306]: Removed session 6. May 13 00:14:49.677000 audit[1449]: CRED_DISP pid=1449 uid=0 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:14:49.687331 kernel: audit: type=1104 audit(1747095289.677:164): pid=1449 uid=0 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:14:49.679000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@5-10.0.0.8:22-10.0.0.1:49820 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:49.706000 audit[1482]: USER_ACCT pid=1482 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:14:49.706733 sshd[1482]: Accepted publickey for core from 10.0.0.1 port 49822 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:14:49.707000 audit[1482]: CRED_ACQ pid=1482 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:14:49.707000 audit[1482]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffffea91730 a2=3 a3=1 items=0 ppid=1 pid=1482 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=7 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:49.707000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:14:49.708189 sshd[1482]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:14:49.712550 systemd[1]: Started session-7.scope. May 13 00:14:49.712738 systemd-logind[1306]: New session 7 of user core. May 13 00:14:49.716000 audit[1482]: USER_START pid=1482 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:14:49.717000 audit[1487]: CRED_ACQ pid=1487 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:14:49.764000 audit[1488]: USER_ACCT pid=1488 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 13 00:14:49.764000 audit[1488]: CRED_REFR pid=1488 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 13 00:14:49.764716 sudo[1488]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 13 00:14:49.764931 sudo[1488]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) May 13 00:14:49.767000 audit[1488]: USER_START pid=1488 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 13 00:14:49.839189 systemd[1]: Starting docker.service... May 13 00:14:49.921558 env[1500]: time="2025-05-13T00:14:49.921494693Z" level=info msg="Starting up" May 13 00:14:49.923181 env[1500]: time="2025-05-13T00:14:49.923151173Z" level=info msg="parsed scheme: \"unix\"" module=grpc May 13 00:14:49.923264 env[1500]: time="2025-05-13T00:14:49.923250253Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc May 13 00:14:49.923337 env[1500]: time="2025-05-13T00:14:49.923321453Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc May 13 00:14:49.923405 env[1500]: time="2025-05-13T00:14:49.923392533Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc May 13 00:14:49.925418 env[1500]: time="2025-05-13T00:14:49.925377893Z" level=info msg="parsed scheme: \"unix\"" module=grpc May 13 00:14:49.925418 env[1500]: time="2025-05-13T00:14:49.925405373Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc May 13 00:14:49.925418 env[1500]: time="2025-05-13T00:14:49.925420333Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc May 13 00:14:49.925546 env[1500]: time="2025-05-13T00:14:49.925429933Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc May 13 00:14:50.104225 env[1500]: time="2025-05-13T00:14:50.103873573Z" level=warning msg="Your kernel does not support cgroup blkio weight" May 13 00:14:50.104225 env[1500]: time="2025-05-13T00:14:50.103902413Z" level=warning msg="Your kernel does not support cgroup blkio weight_device" May 13 00:14:50.104225 env[1500]: time="2025-05-13T00:14:50.104029773Z" level=info msg="Loading containers: start." May 13 00:14:50.158000 audit[1533]: NETFILTER_CFG table=nat:2 family=2 entries=2 op=nft_register_chain pid=1533 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.158000 audit[1533]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=116 a0=3 a1=ffffc78911c0 a2=0 a3=1 items=0 ppid=1500 pid=1533 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.158000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 May 13 00:14:50.159000 audit[1535]: NETFILTER_CFG table=filter:3 family=2 entries=2 op=nft_register_chain pid=1535 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.159000 audit[1535]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=124 a0=3 a1=ffffc9c54920 a2=0 a3=1 items=0 ppid=1500 pid=1535 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.159000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 May 13 00:14:50.161000 audit[1537]: NETFILTER_CFG table=filter:4 family=2 entries=1 op=nft_register_chain pid=1537 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.161000 audit[1537]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=112 a0=3 a1=fffff5d1dce0 a2=0 a3=1 items=0 ppid=1500 pid=1537 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.161000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 May 13 00:14:50.164000 audit[1539]: NETFILTER_CFG table=filter:5 family=2 entries=1 op=nft_register_chain pid=1539 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.164000 audit[1539]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=112 a0=3 a1=ffffd5efa470 a2=0 a3=1 items=0 ppid=1500 pid=1539 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.164000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 May 13 00:14:50.171000 audit[1541]: NETFILTER_CFG table=filter:6 family=2 entries=1 op=nft_register_rule pid=1541 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.171000 audit[1541]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=228 a0=3 a1=fffffadb84d0 a2=0 a3=1 items=0 ppid=1500 pid=1541 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.171000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4100444F434B45522D49534F4C4154494F4E2D53544147452D31002D6A0052455455524E May 13 00:14:50.191000 audit[1546]: NETFILTER_CFG table=filter:7 family=2 entries=1 op=nft_register_rule pid=1546 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.191000 audit[1546]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=228 a0=3 a1=ffffdb791ec0 a2=0 a3=1 items=0 ppid=1500 pid=1546 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.191000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4100444F434B45522D49534F4C4154494F4E2D53544147452D32002D6A0052455455524E May 13 00:14:50.199000 audit[1548]: NETFILTER_CFG table=filter:8 family=2 entries=1 op=nft_register_chain pid=1548 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.199000 audit[1548]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=96 a0=3 a1=ffffd56895f0 a2=0 a3=1 items=0 ppid=1500 pid=1548 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.199000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 May 13 00:14:50.200000 audit[1550]: NETFILTER_CFG table=filter:9 family=2 entries=1 op=nft_register_rule pid=1550 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.200000 audit[1550]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=212 a0=3 a1=ffffdda32450 a2=0 a3=1 items=0 ppid=1500 pid=1550 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.200000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E May 13 00:14:50.202000 audit[1552]: NETFILTER_CFG table=filter:10 family=2 entries=2 op=nft_register_chain pid=1552 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.202000 audit[1552]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=308 a0=3 a1=ffffec1a2da0 a2=0 a3=1 items=0 ppid=1500 pid=1552 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.202000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 May 13 00:14:50.210000 audit[1556]: NETFILTER_CFG table=filter:11 family=2 entries=1 op=nft_unregister_rule pid=1556 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.210000 audit[1556]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=216 a0=3 a1=fffffcee0230 a2=0 a3=1 items=0 ppid=1500 pid=1556 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.210000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4400464F5257415244002D6A00444F434B45522D55534552 May 13 00:14:50.219000 audit[1557]: NETFILTER_CFG table=filter:12 family=2 entries=1 op=nft_register_rule pid=1557 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.219000 audit[1557]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=224 a0=3 a1=fffff5956760 a2=0 a3=1 items=0 ppid=1500 pid=1557 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.219000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 May 13 00:14:50.228556 kernel: Initializing XFRM netlink socket May 13 00:14:50.262121 env[1500]: time="2025-05-13T00:14:50.262079253Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" May 13 00:14:50.280000 audit[1565]: NETFILTER_CFG table=nat:13 family=2 entries=2 op=nft_register_chain pid=1565 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.280000 audit[1565]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=492 a0=3 a1=fffff5cdfc20 a2=0 a3=1 items=0 ppid=1500 pid=1565 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.280000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D4900504F5354524F5554494E47002D73003137322E31372E302E302F31360000002D6F00646F636B657230002D6A004D415351554552414445 May 13 00:14:50.295000 audit[1568]: NETFILTER_CFG table=nat:14 family=2 entries=1 op=nft_register_rule pid=1568 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.295000 audit[1568]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=288 a0=3 a1=ffffcfbc92e0 a2=0 a3=1 items=0 ppid=1500 pid=1568 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.295000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D4900444F434B4552002D6900646F636B657230002D6A0052455455524E May 13 00:14:50.299000 audit[1571]: NETFILTER_CFG table=filter:15 family=2 entries=1 op=nft_register_rule pid=1571 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.299000 audit[1571]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=376 a0=3 a1=ffffe31c7790 a2=0 a3=1 items=0 ppid=1500 pid=1571 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.299000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6900646F636B657230002D6F00646F636B657230002D6A00414343455054 May 13 00:14:50.301000 audit[1573]: NETFILTER_CFG table=filter:16 family=2 entries=1 op=nft_register_rule pid=1573 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.301000 audit[1573]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=376 a0=3 a1=ffffd5176d70 a2=0 a3=1 items=0 ppid=1500 pid=1573 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.301000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6900646F636B6572300000002D6F00646F636B657230002D6A00414343455054 May 13 00:14:50.303000 audit[1575]: NETFILTER_CFG table=nat:17 family=2 entries=2 op=nft_register_chain pid=1575 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.303000 audit[1575]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=356 a0=3 a1=fffff3878720 a2=0 a3=1 items=0 ppid=1500 pid=1575 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.303000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 May 13 00:14:50.306000 audit[1577]: NETFILTER_CFG table=nat:18 family=2 entries=2 op=nft_register_chain pid=1577 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.306000 audit[1577]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=444 a0=3 a1=ffffed8dda60 a2=0 a3=1 items=0 ppid=1500 pid=1577 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.306000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003132372E302E302E302F38 May 13 00:14:50.308000 audit[1579]: NETFILTER_CFG table=filter:19 family=2 entries=1 op=nft_register_rule pid=1579 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.308000 audit[1579]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=304 a0=3 a1=ffffd1f7c660 a2=0 a3=1 items=0 ppid=1500 pid=1579 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.308000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6F00646F636B657230002D6A00444F434B4552 May 13 00:14:50.316000 audit[1582]: NETFILTER_CFG table=filter:20 family=2 entries=1 op=nft_register_rule pid=1582 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.316000 audit[1582]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=508 a0=3 a1=fffffb064270 a2=0 a3=1 items=0 ppid=1500 pid=1582 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.316000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6F00646F636B657230002D6D00636F6E6E747261636B002D2D637473746174650052454C415445442C45535441424C4953484544002D6A00414343455054 May 13 00:14:50.318000 audit[1584]: NETFILTER_CFG table=filter:21 family=2 entries=1 op=nft_register_rule pid=1584 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.318000 audit[1584]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=240 a0=3 a1=ffffd1293a30 a2=0 a3=1 items=0 ppid=1500 pid=1584 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.318000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 May 13 00:14:50.320000 audit[1586]: NETFILTER_CFG table=filter:22 family=2 entries=1 op=nft_register_rule pid=1586 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.320000 audit[1586]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=428 a0=3 a1=ffffdbdd7e60 a2=0 a3=1 items=0 ppid=1500 pid=1586 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.320000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4900444F434B45522D49534F4C4154494F4E2D53544147452D31002D6900646F636B6572300000002D6F00646F636B657230002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D32 May 13 00:14:50.322000 audit[1588]: NETFILTER_CFG table=filter:23 family=2 entries=1 op=nft_register_rule pid=1588 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.322000 audit[1588]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=312 a0=3 a1=ffffdc8b7100 a2=0 a3=1 items=0 ppid=1500 pid=1588 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.322000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4900444F434B45522D49534F4C4154494F4E2D53544147452D32002D6F00646F636B657230002D6A0044524F50 May 13 00:14:50.323313 systemd-networkd[1102]: docker0: Link UP May 13 00:14:50.334000 audit[1592]: NETFILTER_CFG table=filter:24 family=2 entries=1 op=nft_unregister_rule pid=1592 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.334000 audit[1592]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=228 a0=3 a1=ffffded6dd90 a2=0 a3=1 items=0 ppid=1500 pid=1592 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.334000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4400464F5257415244002D6A00444F434B45522D55534552 May 13 00:14:50.347000 audit[1593]: NETFILTER_CFG table=filter:25 family=2 entries=1 op=nft_register_rule pid=1593 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:14:50.347000 audit[1593]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=224 a0=3 a1=fffff59fdae0 a2=0 a3=1 items=0 ppid=1500 pid=1593 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:14:50.347000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 May 13 00:14:50.348183 env[1500]: time="2025-05-13T00:14:50.348155333Z" level=info msg="Loading containers: done." May 13 00:14:50.388147 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1456325557-merged.mount: Deactivated successfully. May 13 00:14:50.395159 env[1500]: time="2025-05-13T00:14:50.395104173Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 13 00:14:50.395324 env[1500]: time="2025-05-13T00:14:50.395302093Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 May 13 00:14:50.395420 env[1500]: time="2025-05-13T00:14:50.395401853Z" level=info msg="Daemon has completed initialization" May 13 00:14:50.423583 systemd[1]: Started docker.service. May 13 00:14:50.423000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=docker comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:50.424343 env[1500]: time="2025-05-13T00:14:50.424196293Z" level=info msg="API listen on /run/docker.sock" May 13 00:14:51.137271 env[1323]: time="2025-05-13T00:14:51.137221893Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.12\"" May 13 00:14:51.811664 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2676521777.mount: Deactivated successfully. May 13 00:14:53.440720 env[1323]: time="2025-05-13T00:14:53.440661573Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:53.442142 env[1323]: time="2025-05-13T00:14:53.442105853Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:afbe230ec4abc2c9e87f7fbe7814bde21dbe30f03252c8861c4ca9510cb43ec6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:53.444280 env[1323]: time="2025-05-13T00:14:53.444237453Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:53.446257 env[1323]: time="2025-05-13T00:14:53.446219293Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:4878682f7a044274d42399a6316ef452c5411aafd4ad99cc57de7235ca490e4e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:53.447008 env[1323]: time="2025-05-13T00:14:53.446972093Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.12\" returns image reference \"sha256:afbe230ec4abc2c9e87f7fbe7814bde21dbe30f03252c8861c4ca9510cb43ec6\"" May 13 00:14:53.456348 env[1323]: time="2025-05-13T00:14:53.456289253Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.12\"" May 13 00:14:55.114136 env[1323]: time="2025-05-13T00:14:55.114076853Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:55.116232 env[1323]: time="2025-05-13T00:14:55.116201173Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3df23260c56ff58d759f8a841c67846184e97ce81a269549ca8d14b36da14c14,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:55.118362 env[1323]: time="2025-05-13T00:14:55.118319853Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:55.140989 env[1323]: time="2025-05-13T00:14:55.136358053Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:3a36711d0409d565b370a18d0c19339e93d4f1b1f2b3fd382eb31c714c463b74,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:55.140989 env[1323]: time="2025-05-13T00:14:55.136720293Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.12\" returns image reference \"sha256:3df23260c56ff58d759f8a841c67846184e97ce81a269549ca8d14b36da14c14\"" May 13 00:14:55.150008 env[1323]: time="2025-05-13T00:14:55.149965933Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.12\"" May 13 00:14:55.984376 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 13 00:14:55.984000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:55.984591 systemd[1]: Stopped kubelet.service. May 13 00:14:55.986332 systemd[1]: Starting kubelet.service... May 13 00:14:55.987958 kernel: kauditd_printk_skb: 84 callbacks suppressed May 13 00:14:55.988016 kernel: audit: type=1130 audit(1747095295.984:199): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:55.988049 kernel: audit: type=1131 audit(1747095295.984:200): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:55.984000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:56.084347 systemd[1]: Started kubelet.service. May 13 00:14:56.084000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:56.088563 kernel: audit: type=1130 audit(1747095296.084:201): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:14:56.140751 kubelet[1661]: E0513 00:14:56.140710 1661 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 13 00:14:56.143460 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 13 00:14:56.143630 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 13 00:14:56.143000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' May 13 00:14:56.147561 kernel: audit: type=1131 audit(1747095296.143:202): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' May 13 00:14:56.463795 env[1323]: time="2025-05-13T00:14:56.463332013Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:56.464894 env[1323]: time="2025-05-13T00:14:56.464863533Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:fb0f5dac5fa74463b801d11598454c00462609b582d17052195012e5f682c2ba,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:56.467048 env[1323]: time="2025-05-13T00:14:56.467020973Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:56.469047 env[1323]: time="2025-05-13T00:14:56.469024133Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:521c843d01025be7d4e246ddee8cde74556eb9813c606d6db9f0f03236f6d029,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:56.469718 env[1323]: time="2025-05-13T00:14:56.469688293Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.12\" returns image reference \"sha256:fb0f5dac5fa74463b801d11598454c00462609b582d17052195012e5f682c2ba\"" May 13 00:14:56.479558 env[1323]: time="2025-05-13T00:14:56.479326133Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.12\"" May 13 00:14:57.647641 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1729361937.mount: Deactivated successfully. May 13 00:14:58.306765 env[1323]: time="2025-05-13T00:14:58.306714773Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:58.311585 env[1323]: time="2025-05-13T00:14:58.311546093Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:b4250a9efcae16f8d20358e204a159844e2b7e854edad08aee8791774acbdaed,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:58.312679 env[1323]: time="2025-05-13T00:14:58.312632493Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:58.314558 env[1323]: time="2025-05-13T00:14:58.314517693Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:ea8c7d5392acf6b0c11ebba78301e1a6c2dc6abcd7544102ed578e49d1c82f15,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:58.314824 env[1323]: time="2025-05-13T00:14:58.314791653Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.12\" returns image reference \"sha256:b4250a9efcae16f8d20358e204a159844e2b7e854edad08aee8791774acbdaed\"" May 13 00:14:58.323529 env[1323]: time="2025-05-13T00:14:58.323485133Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" May 13 00:14:58.876281 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount552183595.mount: Deactivated successfully. May 13 00:14:59.785330 env[1323]: time="2025-05-13T00:14:59.785258093Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.11.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:59.786709 env[1323]: time="2025-05-13T00:14:59.786671093Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:59.788851 env[1323]: time="2025-05-13T00:14:59.788814893Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.11.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:59.791400 env[1323]: time="2025-05-13T00:14:59.791370573Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:14:59.792142 env[1323]: time="2025-05-13T00:14:59.792105533Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\"" May 13 00:14:59.803034 env[1323]: time="2025-05-13T00:14:59.803005573Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" May 13 00:15:00.244194 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount134593340.mount: Deactivated successfully. May 13 00:15:00.250754 env[1323]: time="2025-05-13T00:15:00.250699173Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:00.253444 env[1323]: time="2025-05-13T00:15:00.253410333Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:00.254823 env[1323]: time="2025-05-13T00:15:00.254794813Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:00.257447 env[1323]: time="2025-05-13T00:15:00.257412653Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:00.258116 env[1323]: time="2025-05-13T00:15:00.258079933Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\"" May 13 00:15:00.266309 env[1323]: time="2025-05-13T00:15:00.266263733Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" May 13 00:15:00.774565 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount665714399.mount: Deactivated successfully. May 13 00:15:03.266636 env[1323]: time="2025-05-13T00:15:03.266586653Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.12-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:03.286287 env[1323]: time="2025-05-13T00:15:03.286237213Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:03.289339 env[1323]: time="2025-05-13T00:15:03.289303893Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.12-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:03.294286 env[1323]: time="2025-05-13T00:15:03.294245653Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:03.295349 env[1323]: time="2025-05-13T00:15:03.295297213Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\"" May 13 00:15:06.394468 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. May 13 00:15:06.394660 systemd[1]: Stopped kubelet.service. May 13 00:15:06.394000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:06.396072 systemd[1]: Starting kubelet.service... May 13 00:15:06.394000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:06.400288 kernel: audit: type=1130 audit(1747095306.394:203): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:06.400383 kernel: audit: type=1131 audit(1747095306.394:204): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:06.478075 systemd[1]: Started kubelet.service. May 13 00:15:06.477000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:06.483581 kernel: audit: type=1130 audit(1747095306.477:205): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:06.526162 kubelet[1777]: E0513 00:15:06.526114 1777 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 13 00:15:06.528309 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 13 00:15:06.528454 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 13 00:15:06.528000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' May 13 00:15:06.531560 kernel: audit: type=1131 audit(1747095306.528:206): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' May 13 00:15:08.881023 systemd[1]: Stopped kubelet.service. May 13 00:15:08.880000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:08.883081 systemd[1]: Starting kubelet.service... May 13 00:15:08.880000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:08.886667 kernel: audit: type=1130 audit(1747095308.880:207): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:08.886728 kernel: audit: type=1131 audit(1747095308.880:208): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:08.903285 systemd[1]: Reloading. May 13 00:15:08.951096 /usr/lib/systemd/system-generators/torcx-generator[1813]: time="2025-05-13T00:15:08Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 13 00:15:08.951125 /usr/lib/systemd/system-generators/torcx-generator[1813]: time="2025-05-13T00:15:08Z" level=info msg="torcx already run" May 13 00:15:09.051919 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 13 00:15:09.052115 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 13 00:15:09.069504 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 13 00:15:09.127421 systemd[1]: Started kubelet.service. May 13 00:15:09.127000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:09.130574 kernel: audit: type=1130 audit(1747095309.127:209): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:09.131346 systemd[1]: Stopping kubelet.service... May 13 00:15:09.132346 systemd[1]: kubelet.service: Deactivated successfully. May 13 00:15:09.132698 systemd[1]: Stopped kubelet.service. May 13 00:15:09.132000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:09.134329 systemd[1]: Starting kubelet.service... May 13 00:15:09.135590 kernel: audit: type=1131 audit(1747095309.132:210): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:09.225096 systemd[1]: Started kubelet.service. May 13 00:15:09.233491 kernel: audit: type=1130 audit(1747095309.225:211): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:09.225000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:09.267020 kubelet[1876]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 13 00:15:09.267020 kubelet[1876]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 13 00:15:09.267020 kubelet[1876]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 13 00:15:09.269034 kubelet[1876]: I0513 00:15:09.268975 1876 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 13 00:15:10.334003 kubelet[1876]: I0513 00:15:10.333952 1876 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" May 13 00:15:10.334003 kubelet[1876]: I0513 00:15:10.333988 1876 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 13 00:15:10.334352 kubelet[1876]: I0513 00:15:10.334219 1876 server.go:927] "Client rotation is on, will bootstrap in background" May 13 00:15:10.391191 kubelet[1876]: E0513 00:15:10.391146 1876 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.8:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:10.391791 kubelet[1876]: I0513 00:15:10.391768 1876 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 13 00:15:10.404643 kubelet[1876]: I0513 00:15:10.404607 1876 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 13 00:15:10.405558 kubelet[1876]: I0513 00:15:10.405508 1876 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 13 00:15:10.405719 kubelet[1876]: I0513 00:15:10.405560 1876 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} May 13 00:15:10.405902 kubelet[1876]: I0513 00:15:10.405891 1876 topology_manager.go:138] "Creating topology manager with none policy" May 13 00:15:10.405902 kubelet[1876]: I0513 00:15:10.405902 1876 container_manager_linux.go:301] "Creating device plugin manager" May 13 00:15:10.406239 kubelet[1876]: I0513 00:15:10.406229 1876 state_mem.go:36] "Initialized new in-memory state store" May 13 00:15:10.408013 kubelet[1876]: I0513 00:15:10.407862 1876 kubelet.go:400] "Attempting to sync node with API server" May 13 00:15:10.408013 kubelet[1876]: I0513 00:15:10.407884 1876 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" May 13 00:15:10.408353 kubelet[1876]: I0513 00:15:10.408338 1876 kubelet.go:312] "Adding apiserver pod source" May 13 00:15:10.408652 kubelet[1876]: I0513 00:15:10.408638 1876 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 13 00:15:10.409639 kubelet[1876]: W0513 00:15:10.409583 1876 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.8:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:10.410144 kubelet[1876]: E0513 00:15:10.409953 1876 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.8:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:10.410144 kubelet[1876]: W0513 00:15:10.409584 1876 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.8:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:10.410144 kubelet[1876]: E0513 00:15:10.410010 1876 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.8:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:10.410503 kubelet[1876]: I0513 00:15:10.410485 1876 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" May 13 00:15:10.411411 kubelet[1876]: I0513 00:15:10.411396 1876 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 13 00:15:10.411779 kubelet[1876]: W0513 00:15:10.411766 1876 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 13 00:15:10.413328 kubelet[1876]: I0513 00:15:10.413307 1876 server.go:1264] "Started kubelet" May 13 00:15:10.413480 kubelet[1876]: I0513 00:15:10.413453 1876 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 13 00:15:10.413000 audit[1876]: AVC avc: denied { mac_admin } for pid=1876 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:15:10.414627 kubelet[1876]: I0513 00:15:10.414477 1876 kubelet.go:1419] "Unprivileged containerized plugins might not work, could not set selinux context on plugin registration dir" path="/var/lib/kubelet/plugins_registry" err="setxattr /var/lib/kubelet/plugins_registry: invalid argument" May 13 00:15:10.414627 kubelet[1876]: I0513 00:15:10.414514 1876 kubelet.go:1423] "Unprivileged containerized plugins might not work, could not set selinux context on plugins dir" path="/var/lib/kubelet/plugins" err="setxattr /var/lib/kubelet/plugins: invalid argument" May 13 00:15:10.414797 kubelet[1876]: I0513 00:15:10.414777 1876 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 13 00:15:10.415677 kubelet[1876]: I0513 00:15:10.415655 1876 server.go:455] "Adding debug handlers to kubelet server" May 13 00:15:10.416671 kubelet[1876]: I0513 00:15:10.416623 1876 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 13 00:15:10.416941 kubelet[1876]: I0513 00:15:10.416925 1876 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 13 00:15:10.413000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 13 00:15:10.413000 audit[1876]: SYSCALL arch=c00000b7 syscall=5 success=no exit=-22 a0=40009810e0 a1=40006f0e28 a2=40009810b0 a3=25 items=0 ppid=1 pid=1876 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.417654 kernel: audit: type=1400 audit(1747095310.413:212): avc: denied { mac_admin } for pid=1876 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:15:10.413000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 13 00:15:10.413000 audit[1876]: AVC avc: denied { mac_admin } for pid=1876 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:15:10.413000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 13 00:15:10.413000 audit[1876]: SYSCALL arch=c00000b7 syscall=5 success=no exit=-22 a0=400010fdc0 a1=40006f0e40 a2=4000981170 a3=25 items=0 ppid=1 pid=1876 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.413000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 13 00:15:10.419000 audit[1888]: NETFILTER_CFG table=mangle:26 family=2 entries=2 op=nft_register_chain pid=1888 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:10.419000 audit[1888]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=136 a0=3 a1=fffff237ab20 a2=0 a3=1 items=0 ppid=1876 pid=1888 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.419000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 May 13 00:15:10.420000 audit[1889]: NETFILTER_CFG table=filter:27 family=2 entries=1 op=nft_register_chain pid=1889 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:10.420000 audit[1889]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffd51c1110 a2=0 a3=1 items=0 ppid=1876 pid=1889 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.420000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 May 13 00:15:10.425403 kubelet[1876]: E0513 00:15:10.425382 1876 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 13 00:15:10.425747 kubelet[1876]: I0513 00:15:10.425731 1876 volume_manager.go:291] "Starting Kubelet Volume Manager" May 13 00:15:10.425920 kubelet[1876]: I0513 00:15:10.425904 1876 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 13 00:15:10.426043 kubelet[1876]: I0513 00:15:10.426032 1876 reconciler.go:26] "Reconciler: start to sync state" May 13 00:15:10.426525 kubelet[1876]: W0513 00:15:10.426477 1876 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.8:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:10.426663 kubelet[1876]: E0513 00:15:10.426647 1876 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.8:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:10.427000 audit[1891]: NETFILTER_CFG table=filter:28 family=2 entries=2 op=nft_register_chain pid=1891 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:10.427000 audit[1891]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=312 a0=3 a1=ffffed77ad40 a2=0 a3=1 items=0 ppid=1876 pid=1891 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.427000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C May 13 00:15:10.428521 kubelet[1876]: E0513 00:15:10.428234 1876 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.8:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.8:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.183eededa5ef3155 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-05-13 00:15:10.413279573 +0000 UTC m=+1.184651961,LastTimestamp:2025-05-13 00:15:10.413279573 +0000 UTC m=+1.184651961,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" May 13 00:15:10.428683 kubelet[1876]: I0513 00:15:10.428657 1876 factory.go:221] Registration of the systemd container factory successfully May 13 00:15:10.428765 kubelet[1876]: I0513 00:15:10.428746 1876 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 13 00:15:10.429809 kubelet[1876]: E0513 00:15:10.429777 1876 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 13 00:15:10.429923 kubelet[1876]: E0513 00:15:10.429886 1876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.8:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.8:6443: connect: connection refused" interval="200ms" May 13 00:15:10.429970 kubelet[1876]: I0513 00:15:10.429949 1876 factory.go:221] Registration of the containerd container factory successfully May 13 00:15:10.430000 audit[1893]: NETFILTER_CFG table=filter:29 family=2 entries=2 op=nft_register_chain pid=1893 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:10.430000 audit[1893]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=312 a0=3 a1=ffffffb3e770 a2=0 a3=1 items=0 ppid=1876 pid=1893 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.430000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C May 13 00:15:10.445000 audit[1901]: NETFILTER_CFG table=filter:30 family=2 entries=1 op=nft_register_rule pid=1901 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:10.445000 audit[1901]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=924 a0=3 a1=fffff2efdba0 a2=0 a3=1 items=0 ppid=1876 pid=1901 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.445000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D41004B5542452D4649524557414C4C002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E7400626C6F636B20696E636F6D696E67206C6F63616C6E657420636F6E6E656374696F6E73002D2D647374003132372E302E302E302F38 May 13 00:15:10.446210 kubelet[1876]: I0513 00:15:10.446159 1876 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 13 00:15:10.446000 audit[1903]: NETFILTER_CFG table=mangle:31 family=10 entries=2 op=nft_register_chain pid=1903 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:10.446000 audit[1903]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=136 a0=3 a1=ffffdf696ea0 a2=0 a3=1 items=0 ppid=1876 pid=1903 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.446000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 May 13 00:15:10.447364 kubelet[1876]: I0513 00:15:10.447345 1876 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 13 00:15:10.447715 kubelet[1876]: I0513 00:15:10.447703 1876 status_manager.go:217] "Starting to sync pod status with apiserver" May 13 00:15:10.447790 kubelet[1876]: I0513 00:15:10.447780 1876 kubelet.go:2337] "Starting kubelet main sync loop" May 13 00:15:10.447895 kubelet[1876]: E0513 00:15:10.447877 1876 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 13 00:15:10.447000 audit[1905]: NETFILTER_CFG table=mangle:32 family=2 entries=1 op=nft_register_chain pid=1905 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:10.447000 audit[1905]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffc8d50ff0 a2=0 a3=1 items=0 ppid=1876 pid=1905 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.447000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 May 13 00:15:10.447000 audit[1906]: NETFILTER_CFG table=mangle:33 family=10 entries=1 op=nft_register_chain pid=1906 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:10.447000 audit[1906]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffc1c1e4e0 a2=0 a3=1 items=0 ppid=1876 pid=1906 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.447000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 May 13 00:15:10.448824 kubelet[1876]: W0513 00:15:10.448780 1876 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.8:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:10.448920 kubelet[1876]: E0513 00:15:10.448904 1876 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.8:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:10.448000 audit[1908]: NETFILTER_CFG table=nat:34 family=10 entries=2 op=nft_register_chain pid=1908 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:10.448000 audit[1908]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=128 a0=3 a1=ffffe2396d10 a2=0 a3=1 items=0 ppid=1876 pid=1908 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.448000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 May 13 00:15:10.449000 audit[1907]: NETFILTER_CFG table=nat:35 family=2 entries=1 op=nft_register_chain pid=1907 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:10.449000 audit[1907]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffdfe3cfc0 a2=0 a3=1 items=0 ppid=1876 pid=1907 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.449000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 May 13 00:15:10.450000 audit[1909]: NETFILTER_CFG table=filter:36 family=10 entries=2 op=nft_register_chain pid=1909 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:10.450000 audit[1909]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=136 a0=3 a1=fffff45e6d20 a2=0 a3=1 items=0 ppid=1876 pid=1909 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.450000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 May 13 00:15:10.450838 kubelet[1876]: I0513 00:15:10.450797 1876 cpu_manager.go:214] "Starting CPU manager" policy="none" May 13 00:15:10.450838 kubelet[1876]: I0513 00:15:10.450813 1876 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 13 00:15:10.450838 kubelet[1876]: I0513 00:15:10.450832 1876 state_mem.go:36] "Initialized new in-memory state store" May 13 00:15:10.451000 audit[1910]: NETFILTER_CFG table=filter:37 family=2 entries=1 op=nft_register_chain pid=1910 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:10.451000 audit[1910]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=fffffa5618a0 a2=0 a3=1 items=0 ppid=1876 pid=1910 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.451000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 May 13 00:15:10.527603 kubelet[1876]: I0513 00:15:10.527571 1876 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 13 00:15:10.527969 kubelet[1876]: E0513 00:15:10.527945 1876 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.8:6443/api/v1/nodes\": dial tcp 10.0.0.8:6443: connect: connection refused" node="localhost" May 13 00:15:10.543153 kubelet[1876]: I0513 00:15:10.543126 1876 policy_none.go:49] "None policy: Start" May 13 00:15:10.543904 kubelet[1876]: I0513 00:15:10.543884 1876 memory_manager.go:170] "Starting memorymanager" policy="None" May 13 00:15:10.543985 kubelet[1876]: I0513 00:15:10.543914 1876 state_mem.go:35] "Initializing new in-memory state store" May 13 00:15:10.548316 kubelet[1876]: E0513 00:15:10.548293 1876 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 13 00:15:10.551874 kubelet[1876]: I0513 00:15:10.551849 1876 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 13 00:15:10.551000 audit[1876]: AVC avc: denied { mac_admin } for pid=1876 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:15:10.551000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 13 00:15:10.551000 audit[1876]: SYSCALL arch=c00000b7 syscall=5 success=no exit=-22 a0=4000955170 a1=4000956348 a2=4000955140 a3=25 items=0 ppid=1 pid=1876 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:10.551000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 13 00:15:10.552084 kubelet[1876]: I0513 00:15:10.551931 1876 server.go:88] "Unprivileged containerized plugins might not work. Could not set selinux context on socket dir" path="/var/lib/kubelet/device-plugins/" err="setxattr /var/lib/kubelet/device-plugins/: invalid argument" May 13 00:15:10.552084 kubelet[1876]: I0513 00:15:10.552035 1876 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 13 00:15:10.552130 kubelet[1876]: I0513 00:15:10.552125 1876 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 13 00:15:10.553686 kubelet[1876]: E0513 00:15:10.553666 1876 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" May 13 00:15:10.631438 kubelet[1876]: E0513 00:15:10.631331 1876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.8:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.8:6443: connect: connection refused" interval="400ms" May 13 00:15:10.729717 kubelet[1876]: I0513 00:15:10.729675 1876 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 13 00:15:10.730045 kubelet[1876]: E0513 00:15:10.729996 1876 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.8:6443/api/v1/nodes\": dial tcp 10.0.0.8:6443: connect: connection refused" node="localhost" May 13 00:15:10.749211 kubelet[1876]: I0513 00:15:10.749147 1876 topology_manager.go:215] "Topology Admit Handler" podUID="e84797d13280e3c51f319a646d598d49" podNamespace="kube-system" podName="kube-apiserver-localhost" May 13 00:15:10.750320 kubelet[1876]: I0513 00:15:10.750289 1876 topology_manager.go:215] "Topology Admit Handler" podUID="b20b39a8540dba87b5883a6f0f602dba" podNamespace="kube-system" podName="kube-controller-manager-localhost" May 13 00:15:10.751097 kubelet[1876]: I0513 00:15:10.751071 1876 topology_manager.go:215] "Topology Admit Handler" podUID="6ece95f10dbffa04b25ec3439a115512" podNamespace="kube-system" podName="kube-scheduler-localhost" May 13 00:15:10.827912 kubelet[1876]: I0513 00:15:10.827868 1876 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6ece95f10dbffa04b25ec3439a115512-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6ece95f10dbffa04b25ec3439a115512\") " pod="kube-system/kube-scheduler-localhost" May 13 00:15:10.827912 kubelet[1876]: I0513 00:15:10.827916 1876 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e84797d13280e3c51f319a646d598d49-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"e84797d13280e3c51f319a646d598d49\") " pod="kube-system/kube-apiserver-localhost" May 13 00:15:10.828065 kubelet[1876]: I0513 00:15:10.827938 1876 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 13 00:15:10.828065 kubelet[1876]: I0513 00:15:10.827955 1876 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 13 00:15:10.828065 kubelet[1876]: I0513 00:15:10.827975 1876 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e84797d13280e3c51f319a646d598d49-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"e84797d13280e3c51f319a646d598d49\") " pod="kube-system/kube-apiserver-localhost" May 13 00:15:10.828065 kubelet[1876]: I0513 00:15:10.827992 1876 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e84797d13280e3c51f319a646d598d49-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"e84797d13280e3c51f319a646d598d49\") " pod="kube-system/kube-apiserver-localhost" May 13 00:15:10.828065 kubelet[1876]: I0513 00:15:10.828006 1876 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 13 00:15:10.828208 kubelet[1876]: I0513 00:15:10.828021 1876 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 13 00:15:10.828208 kubelet[1876]: I0513 00:15:10.828036 1876 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 13 00:15:11.032431 kubelet[1876]: E0513 00:15:11.032326 1876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.8:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.8:6443: connect: connection refused" interval="800ms" May 13 00:15:11.054555 kubelet[1876]: E0513 00:15:11.054512 1876 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:11.055700 kubelet[1876]: E0513 00:15:11.055679 1876 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:11.056034 kubelet[1876]: E0513 00:15:11.056016 1876 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:11.060130 env[1323]: time="2025-05-13T00:15:11.058392173Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:e84797d13280e3c51f319a646d598d49,Namespace:kube-system,Attempt:0,}" May 13 00:15:11.060459 env[1323]: time="2025-05-13T00:15:11.060107373Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6ece95f10dbffa04b25ec3439a115512,Namespace:kube-system,Attempt:0,}" May 13 00:15:11.060850 env[1323]: time="2025-05-13T00:15:11.060706133Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b20b39a8540dba87b5883a6f0f602dba,Namespace:kube-system,Attempt:0,}" May 13 00:15:11.131375 kubelet[1876]: I0513 00:15:11.131327 1876 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 13 00:15:11.131762 kubelet[1876]: E0513 00:15:11.131730 1876 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.8:6443/api/v1/nodes\": dial tcp 10.0.0.8:6443: connect: connection refused" node="localhost" May 13 00:15:11.530765 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3514315185.mount: Deactivated successfully. May 13 00:15:11.535831 env[1323]: time="2025-05-13T00:15:11.535784813Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:11.540061 env[1323]: time="2025-05-13T00:15:11.540017453Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:11.541452 env[1323]: time="2025-05-13T00:15:11.541414853Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:11.543244 env[1323]: time="2025-05-13T00:15:11.543205133Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:11.543998 env[1323]: time="2025-05-13T00:15:11.543970133Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:11.547031 kubelet[1876]: W0513 00:15:11.546927 1876 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.8:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:11.547031 kubelet[1876]: E0513 00:15:11.546989 1876 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.8:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:11.548070 env[1323]: time="2025-05-13T00:15:11.548033773Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:11.550228 env[1323]: time="2025-05-13T00:15:11.550170533Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:11.553186 env[1323]: time="2025-05-13T00:15:11.553117093Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:11.554970 env[1323]: time="2025-05-13T00:15:11.554933933Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:11.556769 env[1323]: time="2025-05-13T00:15:11.556738293Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:11.558551 env[1323]: time="2025-05-13T00:15:11.558512373Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:11.560393 env[1323]: time="2025-05-13T00:15:11.560362213Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:11.595365 env[1323]: time="2025-05-13T00:15:11.595299773Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 13 00:15:11.595365 env[1323]: time="2025-05-13T00:15:11.595340413Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 13 00:15:11.595680 env[1323]: time="2025-05-13T00:15:11.595643133Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 13 00:15:11.596674 env[1323]: time="2025-05-13T00:15:11.596501973Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/4e96c9694a4afacb904b0cdd2e3949f232e2ebf64306791a726e045444020c63 pid=1927 runtime=io.containerd.runc.v2 May 13 00:15:11.599991 env[1323]: time="2025-05-13T00:15:11.599811173Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 13 00:15:11.599991 env[1323]: time="2025-05-13T00:15:11.599846933Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 13 00:15:11.599991 env[1323]: time="2025-05-13T00:15:11.599857653Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 13 00:15:11.600145 env[1323]: time="2025-05-13T00:15:11.600028253Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/fdcd6b3407881bd262e64628f4976ce0b3d2e2e403b9666f1d6e026eb0032e1b pid=1946 runtime=io.containerd.runc.v2 May 13 00:15:11.601136 env[1323]: time="2025-05-13T00:15:11.600968973Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 13 00:15:11.601136 env[1323]: time="2025-05-13T00:15:11.600998853Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 13 00:15:11.601136 env[1323]: time="2025-05-13T00:15:11.601009453Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 13 00:15:11.601248 env[1323]: time="2025-05-13T00:15:11.601179453Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/8bd301cd8177bcb8b0beac3ef85e8ab1d37d3019bfe20bb5cbb271ecb51c5b97 pid=1944 runtime=io.containerd.runc.v2 May 13 00:15:11.631014 kubelet[1876]: W0513 00:15:11.630950 1876 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.8:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:11.631014 kubelet[1876]: E0513 00:15:11.630989 1876 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.8:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:11.685562 env[1323]: time="2025-05-13T00:15:11.685486493Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:e84797d13280e3c51f319a646d598d49,Namespace:kube-system,Attempt:0,} returns sandbox id \"4e96c9694a4afacb904b0cdd2e3949f232e2ebf64306791a726e045444020c63\"" May 13 00:15:11.686683 kubelet[1876]: E0513 00:15:11.686435 1876 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:11.687530 env[1323]: time="2025-05-13T00:15:11.687462693Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6ece95f10dbffa04b25ec3439a115512,Namespace:kube-system,Attempt:0,} returns sandbox id \"fdcd6b3407881bd262e64628f4976ce0b3d2e2e403b9666f1d6e026eb0032e1b\"" May 13 00:15:11.687530 env[1323]: time="2025-05-13T00:15:11.687504373Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b20b39a8540dba87b5883a6f0f602dba,Namespace:kube-system,Attempt:0,} returns sandbox id \"8bd301cd8177bcb8b0beac3ef85e8ab1d37d3019bfe20bb5cbb271ecb51c5b97\"" May 13 00:15:11.689124 kubelet[1876]: E0513 00:15:11.688948 1876 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:11.691573 kubelet[1876]: E0513 00:15:11.689747 1876 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:11.691650 env[1323]: time="2025-05-13T00:15:11.691126573Z" level=info msg="CreateContainer within sandbox \"fdcd6b3407881bd262e64628f4976ce0b3d2e2e403b9666f1d6e026eb0032e1b\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 13 00:15:11.691650 env[1323]: time="2025-05-13T00:15:11.691357413Z" level=info msg="CreateContainer within sandbox \"8bd301cd8177bcb8b0beac3ef85e8ab1d37d3019bfe20bb5cbb271ecb51c5b97\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 13 00:15:11.691995 env[1323]: time="2025-05-13T00:15:11.691967213Z" level=info msg="CreateContainer within sandbox \"4e96c9694a4afacb904b0cdd2e3949f232e2ebf64306791a726e045444020c63\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 13 00:15:11.713061 env[1323]: time="2025-05-13T00:15:11.713001693Z" level=info msg="CreateContainer within sandbox \"fdcd6b3407881bd262e64628f4976ce0b3d2e2e403b9666f1d6e026eb0032e1b\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"39b32163e601d1f909c72721953a1563bdfa9ef03f93efce4aeec437904b3e80\"" May 13 00:15:11.713832 env[1323]: time="2025-05-13T00:15:11.713804213Z" level=info msg="StartContainer for \"39b32163e601d1f909c72721953a1563bdfa9ef03f93efce4aeec437904b3e80\"" May 13 00:15:11.722046 env[1323]: time="2025-05-13T00:15:11.721950093Z" level=info msg="CreateContainer within sandbox \"8bd301cd8177bcb8b0beac3ef85e8ab1d37d3019bfe20bb5cbb271ecb51c5b97\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"6b5b83a4845a6d31e40a0d188df0f06b82cb8588e792a13504eb5506330ad126\"" May 13 00:15:11.722712 env[1323]: time="2025-05-13T00:15:11.722660253Z" level=info msg="StartContainer for \"6b5b83a4845a6d31e40a0d188df0f06b82cb8588e792a13504eb5506330ad126\"" May 13 00:15:11.722786 env[1323]: time="2025-05-13T00:15:11.722694493Z" level=info msg="CreateContainer within sandbox \"4e96c9694a4afacb904b0cdd2e3949f232e2ebf64306791a726e045444020c63\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"59b1169b18d43d09995e776261cb5c9fae73afe286fa36f0245a03b75e75e1dc\"" May 13 00:15:11.723177 env[1323]: time="2025-05-13T00:15:11.723143493Z" level=info msg="StartContainer for \"59b1169b18d43d09995e776261cb5c9fae73afe286fa36f0245a03b75e75e1dc\"" May 13 00:15:11.727954 kubelet[1876]: W0513 00:15:11.727858 1876 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.8:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:11.727954 kubelet[1876]: E0513 00:15:11.727915 1876 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.8:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:11.826862 env[1323]: time="2025-05-13T00:15:11.826757173Z" level=info msg="StartContainer for \"39b32163e601d1f909c72721953a1563bdfa9ef03f93efce4aeec437904b3e80\" returns successfully" May 13 00:15:11.834468 kubelet[1876]: E0513 00:15:11.834405 1876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.8:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.8:6443: connect: connection refused" interval="1.6s" May 13 00:15:11.846912 kubelet[1876]: W0513 00:15:11.846842 1876 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.8:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:11.846912 kubelet[1876]: E0513 00:15:11.846912 1876 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.8:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.8:6443: connect: connection refused May 13 00:15:11.886353 env[1323]: time="2025-05-13T00:15:11.877068133Z" level=info msg="StartContainer for \"6b5b83a4845a6d31e40a0d188df0f06b82cb8588e792a13504eb5506330ad126\" returns successfully" May 13 00:15:11.886353 env[1323]: time="2025-05-13T00:15:11.880393973Z" level=info msg="StartContainer for \"59b1169b18d43d09995e776261cb5c9fae73afe286fa36f0245a03b75e75e1dc\" returns successfully" May 13 00:15:11.935654 kubelet[1876]: I0513 00:15:11.935615 1876 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 13 00:15:11.935980 kubelet[1876]: E0513 00:15:11.935933 1876 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.8:6443/api/v1/nodes\": dial tcp 10.0.0.8:6443: connect: connection refused" node="localhost" May 13 00:15:12.456467 kubelet[1876]: E0513 00:15:12.456192 1876 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:12.456621 kubelet[1876]: E0513 00:15:12.456556 1876 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:12.457613 kubelet[1876]: E0513 00:15:12.457591 1876 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:13.446593 kubelet[1876]: E0513 00:15:13.446558 1876 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" May 13 00:15:13.459664 kubelet[1876]: E0513 00:15:13.459623 1876 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:13.459664 kubelet[1876]: E0513 00:15:13.459666 1876 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:13.537502 kubelet[1876]: I0513 00:15:13.537464 1876 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 13 00:15:13.546618 kubelet[1876]: I0513 00:15:13.546574 1876 kubelet_node_status.go:76] "Successfully registered node" node="localhost" May 13 00:15:13.554051 kubelet[1876]: E0513 00:15:13.554007 1876 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 13 00:15:13.654219 kubelet[1876]: E0513 00:15:13.654173 1876 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 13 00:15:13.755076 kubelet[1876]: E0513 00:15:13.754970 1876 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 13 00:15:13.855342 kubelet[1876]: E0513 00:15:13.855302 1876 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 13 00:15:13.956073 kubelet[1876]: E0513 00:15:13.956025 1876 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 13 00:15:14.056553 kubelet[1876]: E0513 00:15:14.056390 1876 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 13 00:15:14.156823 kubelet[1876]: E0513 00:15:14.156775 1876 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 13 00:15:14.257183 kubelet[1876]: E0513 00:15:14.257139 1876 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 13 00:15:14.357972 kubelet[1876]: E0513 00:15:14.357866 1876 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 13 00:15:14.410778 kubelet[1876]: I0513 00:15:14.410737 1876 apiserver.go:52] "Watching apiserver" May 13 00:15:14.426351 kubelet[1876]: I0513 00:15:14.426316 1876 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 13 00:15:15.078550 kubelet[1876]: E0513 00:15:15.078502 1876 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:15.461473 kubelet[1876]: E0513 00:15:15.461364 1876 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:15.694727 systemd[1]: Reloading. May 13 00:15:15.750648 /usr/lib/systemd/system-generators/torcx-generator[2176]: time="2025-05-13T00:15:15Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 13 00:15:15.750676 /usr/lib/systemd/system-generators/torcx-generator[2176]: time="2025-05-13T00:15:15Z" level=info msg="torcx already run" May 13 00:15:15.828854 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 13 00:15:15.829000 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 13 00:15:15.846427 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 13 00:15:15.910528 systemd[1]: Stopping kubelet.service... May 13 00:15:15.929997 systemd[1]: kubelet.service: Deactivated successfully. May 13 00:15:15.930346 systemd[1]: Stopped kubelet.service. May 13 00:15:15.929000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:15.931139 kernel: kauditd_printk_skb: 47 callbacks suppressed May 13 00:15:15.931195 kernel: audit: type=1131 audit(1747095315.929:227): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:15.932521 systemd[1]: Starting kubelet.service... May 13 00:15:16.017000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:16.017350 systemd[1]: Started kubelet.service. May 13 00:15:16.021603 kernel: audit: type=1130 audit(1747095316.017:228): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:16.063901 kubelet[2230]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 13 00:15:16.063901 kubelet[2230]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 13 00:15:16.063901 kubelet[2230]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 13 00:15:16.064264 kubelet[2230]: I0513 00:15:16.063967 2230 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 13 00:15:16.068196 kubelet[2230]: I0513 00:15:16.068160 2230 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" May 13 00:15:16.068196 kubelet[2230]: I0513 00:15:16.068186 2230 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 13 00:15:16.068411 kubelet[2230]: I0513 00:15:16.068381 2230 server.go:927] "Client rotation is on, will bootstrap in background" May 13 00:15:16.069712 kubelet[2230]: I0513 00:15:16.069688 2230 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". May 13 00:15:16.070904 kubelet[2230]: I0513 00:15:16.070871 2230 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 13 00:15:16.079370 kubelet[2230]: I0513 00:15:16.079345 2230 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 13 00:15:16.079843 kubelet[2230]: I0513 00:15:16.079817 2230 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 13 00:15:16.079999 kubelet[2230]: I0513 00:15:16.079847 2230 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} May 13 00:15:16.080492 kubelet[2230]: I0513 00:15:16.080005 2230 topology_manager.go:138] "Creating topology manager with none policy" May 13 00:15:16.080492 kubelet[2230]: I0513 00:15:16.080014 2230 container_manager_linux.go:301] "Creating device plugin manager" May 13 00:15:16.080492 kubelet[2230]: I0513 00:15:16.080045 2230 state_mem.go:36] "Initialized new in-memory state store" May 13 00:15:16.080492 kubelet[2230]: I0513 00:15:16.080129 2230 kubelet.go:400] "Attempting to sync node with API server" May 13 00:15:16.080492 kubelet[2230]: I0513 00:15:16.080142 2230 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" May 13 00:15:16.080492 kubelet[2230]: I0513 00:15:16.080167 2230 kubelet.go:312] "Adding apiserver pod source" May 13 00:15:16.080492 kubelet[2230]: I0513 00:15:16.080182 2230 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 13 00:15:16.094064 kernel: audit: type=1400 audit(1747095316.083:229): avc: denied { mac_admin } for pid=2230 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:15:16.094161 kernel: audit: type=1401 audit(1747095316.083:229): op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 13 00:15:16.094179 kernel: audit: type=1300 audit(1747095316.083:229): arch=c00000b7 syscall=5 success=no exit=-22 a0=40008095f0 a1=4000a3e8b8 a2=40008095c0 a3=25 items=0 ppid=1 pid=2230 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:16.083000 audit[2230]: AVC avc: denied { mac_admin } for pid=2230 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:15:16.083000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 13 00:15:16.083000 audit[2230]: SYSCALL arch=c00000b7 syscall=5 success=no exit=-22 a0=40008095f0 a1=4000a3e8b8 a2=40008095c0 a3=25 items=0 ppid=1 pid=2230 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:16.094332 kubelet[2230]: I0513 00:15:16.081158 2230 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" May 13 00:15:16.094332 kubelet[2230]: I0513 00:15:16.081415 2230 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 13 00:15:16.094332 kubelet[2230]: I0513 00:15:16.081881 2230 server.go:1264] "Started kubelet" May 13 00:15:16.094332 kubelet[2230]: I0513 00:15:16.082453 2230 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 13 00:15:16.094332 kubelet[2230]: I0513 00:15:16.083032 2230 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 13 00:15:16.094332 kubelet[2230]: I0513 00:15:16.083084 2230 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 13 00:15:16.094332 kubelet[2230]: I0513 00:15:16.084110 2230 server.go:455] "Adding debug handlers to kubelet server" May 13 00:15:16.094332 kubelet[2230]: I0513 00:15:16.089146 2230 kubelet.go:1419] "Unprivileged containerized plugins might not work, could not set selinux context on plugin registration dir" path="/var/lib/kubelet/plugins_registry" err="setxattr /var/lib/kubelet/plugins_registry: invalid argument" May 13 00:15:16.094332 kubelet[2230]: I0513 00:15:16.089207 2230 kubelet.go:1423] "Unprivileged containerized plugins might not work, could not set selinux context on plugins dir" path="/var/lib/kubelet/plugins" err="setxattr /var/lib/kubelet/plugins: invalid argument" May 13 00:15:16.094332 kubelet[2230]: I0513 00:15:16.089235 2230 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 13 00:15:16.083000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 13 00:15:16.102469 kernel: audit: type=1327 audit(1747095316.083:229): proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 13 00:15:16.102616 kubelet[2230]: I0513 00:15:16.100679 2230 volume_manager.go:291] "Starting Kubelet Volume Manager" May 13 00:15:16.102616 kubelet[2230]: I0513 00:15:16.100778 2230 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 13 00:15:16.102616 kubelet[2230]: I0513 00:15:16.100917 2230 reconciler.go:26] "Reconciler: start to sync state" May 13 00:15:16.102616 kubelet[2230]: I0513 00:15:16.102525 2230 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 13 00:15:16.088000 audit[2230]: AVC avc: denied { mac_admin } for pid=2230 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:15:16.105381 kubelet[2230]: E0513 00:15:16.105359 2230 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 13 00:15:16.105655 kernel: audit: type=1400 audit(1747095316.088:230): avc: denied { mac_admin } for pid=2230 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:15:16.107764 kernel: audit: type=1401 audit(1747095316.088:230): op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 13 00:15:16.088000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 13 00:15:16.107844 kubelet[2230]: I0513 00:15:16.105727 2230 factory.go:221] Registration of the containerd container factory successfully May 13 00:15:16.107844 kubelet[2230]: I0513 00:15:16.105739 2230 factory.go:221] Registration of the systemd container factory successfully May 13 00:15:16.088000 audit[2230]: SYSCALL arch=c00000b7 syscall=5 success=no exit=-22 a0=40006c36c0 a1=4000a3e8d0 a2=4000809680 a3=25 items=0 ppid=1 pid=2230 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:16.113068 kernel: audit: type=1300 audit(1747095316.088:230): arch=c00000b7 syscall=5 success=no exit=-22 a0=40006c36c0 a1=4000a3e8d0 a2=4000809680 a3=25 items=0 ppid=1 pid=2230 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:16.088000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 13 00:15:16.116528 kernel: audit: type=1327 audit(1747095316.088:230): proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 13 00:15:16.127829 kubelet[2230]: I0513 00:15:16.127790 2230 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 13 00:15:16.128780 kubelet[2230]: I0513 00:15:16.128751 2230 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 13 00:15:16.128780 kubelet[2230]: I0513 00:15:16.128784 2230 status_manager.go:217] "Starting to sync pod status with apiserver" May 13 00:15:16.128861 kubelet[2230]: I0513 00:15:16.128801 2230 kubelet.go:2337] "Starting kubelet main sync loop" May 13 00:15:16.128861 kubelet[2230]: E0513 00:15:16.128840 2230 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 13 00:15:16.152087 kubelet[2230]: I0513 00:15:16.152058 2230 cpu_manager.go:214] "Starting CPU manager" policy="none" May 13 00:15:16.152087 kubelet[2230]: I0513 00:15:16.152075 2230 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 13 00:15:16.152087 kubelet[2230]: I0513 00:15:16.152096 2230 state_mem.go:36] "Initialized new in-memory state store" May 13 00:15:16.152260 kubelet[2230]: I0513 00:15:16.152249 2230 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 13 00:15:16.152283 kubelet[2230]: I0513 00:15:16.152261 2230 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 13 00:15:16.152283 kubelet[2230]: I0513 00:15:16.152280 2230 policy_none.go:49] "None policy: Start" May 13 00:15:16.152834 kubelet[2230]: I0513 00:15:16.152820 2230 memory_manager.go:170] "Starting memorymanager" policy="None" May 13 00:15:16.152919 kubelet[2230]: I0513 00:15:16.152909 2230 state_mem.go:35] "Initializing new in-memory state store" May 13 00:15:16.153097 kubelet[2230]: I0513 00:15:16.153083 2230 state_mem.go:75] "Updated machine memory state" May 13 00:15:16.154230 kubelet[2230]: I0513 00:15:16.154205 2230 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 13 00:15:16.153000 audit[2230]: AVC avc: denied { mac_admin } for pid=2230 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:15:16.153000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 13 00:15:16.153000 audit[2230]: SYSCALL arch=c00000b7 syscall=5 success=no exit=-22 a0=400132d170 a1=40013289d8 a2=400132d140 a3=25 items=0 ppid=1 pid=2230 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:16.153000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 13 00:15:16.154599 kubelet[2230]: I0513 00:15:16.154578 2230 server.go:88] "Unprivileged containerized plugins might not work. Could not set selinux context on socket dir" path="/var/lib/kubelet/device-plugins/" err="setxattr /var/lib/kubelet/device-plugins/: invalid argument" May 13 00:15:16.154813 kubelet[2230]: I0513 00:15:16.154778 2230 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 13 00:15:16.154962 kubelet[2230]: I0513 00:15:16.154951 2230 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 13 00:15:16.204250 kubelet[2230]: I0513 00:15:16.204198 2230 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 13 00:15:16.211055 kubelet[2230]: I0513 00:15:16.211022 2230 kubelet_node_status.go:112] "Node was previously registered" node="localhost" May 13 00:15:16.211253 kubelet[2230]: I0513 00:15:16.211241 2230 kubelet_node_status.go:76] "Successfully registered node" node="localhost" May 13 00:15:16.229265 kubelet[2230]: I0513 00:15:16.229191 2230 topology_manager.go:215] "Topology Admit Handler" podUID="b20b39a8540dba87b5883a6f0f602dba" podNamespace="kube-system" podName="kube-controller-manager-localhost" May 13 00:15:16.229413 kubelet[2230]: I0513 00:15:16.229334 2230 topology_manager.go:215] "Topology Admit Handler" podUID="6ece95f10dbffa04b25ec3439a115512" podNamespace="kube-system" podName="kube-scheduler-localhost" May 13 00:15:16.229413 kubelet[2230]: I0513 00:15:16.229385 2230 topology_manager.go:215] "Topology Admit Handler" podUID="e84797d13280e3c51f319a646d598d49" podNamespace="kube-system" podName="kube-apiserver-localhost" May 13 00:15:16.236892 kubelet[2230]: E0513 00:15:16.236836 2230 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" May 13 00:15:16.301434 kubelet[2230]: I0513 00:15:16.301331 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 13 00:15:16.301434 kubelet[2230]: I0513 00:15:16.301430 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 13 00:15:16.301608 kubelet[2230]: I0513 00:15:16.301455 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6ece95f10dbffa04b25ec3439a115512-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6ece95f10dbffa04b25ec3439a115512\") " pod="kube-system/kube-scheduler-localhost" May 13 00:15:16.301608 kubelet[2230]: I0513 00:15:16.301498 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e84797d13280e3c51f319a646d598d49-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"e84797d13280e3c51f319a646d598d49\") " pod="kube-system/kube-apiserver-localhost" May 13 00:15:16.301608 kubelet[2230]: I0513 00:15:16.301518 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e84797d13280e3c51f319a646d598d49-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"e84797d13280e3c51f319a646d598d49\") " pod="kube-system/kube-apiserver-localhost" May 13 00:15:16.301608 kubelet[2230]: I0513 00:15:16.301567 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e84797d13280e3c51f319a646d598d49-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"e84797d13280e3c51f319a646d598d49\") " pod="kube-system/kube-apiserver-localhost" May 13 00:15:16.301608 kubelet[2230]: I0513 00:15:16.301588 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 13 00:15:16.301727 kubelet[2230]: I0513 00:15:16.301603 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 13 00:15:16.301727 kubelet[2230]: I0513 00:15:16.301631 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 13 00:15:16.537923 kubelet[2230]: E0513 00:15:16.537886 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:16.538068 kubelet[2230]: E0513 00:15:16.538043 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:16.538247 kubelet[2230]: E0513 00:15:16.538230 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:17.080940 kubelet[2230]: I0513 00:15:17.080903 2230 apiserver.go:52] "Watching apiserver" May 13 00:15:17.103507 kubelet[2230]: I0513 00:15:17.101111 2230 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 13 00:15:17.142086 kubelet[2230]: E0513 00:15:17.139883 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:17.149405 kubelet[2230]: E0513 00:15:17.149359 2230 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" May 13 00:15:17.149692 kubelet[2230]: E0513 00:15:17.149670 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:17.149777 kubelet[2230]: E0513 00:15:17.149760 2230 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" May 13 00:15:17.150481 kubelet[2230]: E0513 00:15:17.150454 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:17.164871 kubelet[2230]: I0513 00:15:17.164810 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=2.164747498 podStartE2EDuration="2.164747498s" podCreationTimestamp="2025-05-13 00:15:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-13 00:15:17.164461097 +0000 UTC m=+1.142683845" watchObservedRunningTime="2025-05-13 00:15:17.164747498 +0000 UTC m=+1.142970246" May 13 00:15:17.177495 kubelet[2230]: I0513 00:15:17.177428 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.177402997 podStartE2EDuration="1.177402997s" podCreationTimestamp="2025-05-13 00:15:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-13 00:15:17.176257712 +0000 UTC m=+1.154480460" watchObservedRunningTime="2025-05-13 00:15:17.177402997 +0000 UTC m=+1.155625705" May 13 00:15:17.184252 kubelet[2230]: I0513 00:15:17.184193 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.184177869 podStartE2EDuration="1.184177869s" podCreationTimestamp="2025-05-13 00:15:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-13 00:15:17.183141304 +0000 UTC m=+1.161364012" watchObservedRunningTime="2025-05-13 00:15:17.184177869 +0000 UTC m=+1.162400617" May 13 00:15:18.140619 kubelet[2230]: E0513 00:15:18.140575 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:18.141291 kubelet[2230]: E0513 00:15:18.141255 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:19.142107 kubelet[2230]: E0513 00:15:19.142077 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:20.978210 sudo[1488]: pam_unix(sudo:session): session closed for user root May 13 00:15:20.977000 audit[1488]: USER_END pid=1488 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 13 00:15:20.979033 kernel: kauditd_printk_skb: 4 callbacks suppressed May 13 00:15:20.979066 kernel: audit: type=1106 audit(1747095320.977:232): pid=1488 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 13 00:15:20.980099 sshd[1482]: pam_unix(sshd:session): session closed for user core May 13 00:15:20.977000 audit[1488]: CRED_DISP pid=1488 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 13 00:15:20.985032 kernel: audit: type=1104 audit(1747095320.977:233): pid=1488 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 13 00:15:20.985000 audit[1482]: USER_END pid=1482 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:20.987752 systemd-logind[1306]: Session 7 logged out. Waiting for processes to exit. May 13 00:15:20.988337 systemd[1]: sshd@6-10.0.0.8:22-10.0.0.1:49822.service: Deactivated successfully. May 13 00:15:20.989146 systemd[1]: session-7.scope: Deactivated successfully. May 13 00:15:20.990005 systemd-logind[1306]: Removed session 7. May 13 00:15:20.985000 audit[1482]: CRED_DISP pid=1482 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:20.993193 kernel: audit: type=1106 audit(1747095320.985:234): pid=1482 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:20.993251 kernel: audit: type=1104 audit(1747095320.985:235): pid=1482 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:20.988000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-10.0.0.8:22-10.0.0.1:49822 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:20.996172 kernel: audit: type=1131 audit(1747095320.988:236): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-10.0.0.8:22-10.0.0.1:49822 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:25.723580 kubelet[2230]: E0513 00:15:25.722024 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:26.154780 kubelet[2230]: E0513 00:15:26.154682 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:27.060184 kubelet[2230]: E0513 00:15:27.059726 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:28.284060 kubelet[2230]: E0513 00:15:28.282098 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:29.532941 update_engine[1314]: I0513 00:15:29.532712 1314 update_attempter.cc:509] Updating boot flags... May 13 00:15:31.456498 kubelet[2230]: I0513 00:15:31.456426 2230 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 13 00:15:31.457172 env[1323]: time="2025-05-13T00:15:31.457085908Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 13 00:15:31.457787 kubelet[2230]: I0513 00:15:31.457766 2230 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 13 00:15:31.882728 kubelet[2230]: I0513 00:15:31.882675 2230 topology_manager.go:215] "Topology Admit Handler" podUID="9d638da6-3528-4b32-93e9-8e5baa883fb3" podNamespace="tigera-operator" podName="tigera-operator-797db67f8-zrgmk" May 13 00:15:31.913361 kubelet[2230]: I0513 00:15:31.913288 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/9d638da6-3528-4b32-93e9-8e5baa883fb3-var-lib-calico\") pod \"tigera-operator-797db67f8-zrgmk\" (UID: \"9d638da6-3528-4b32-93e9-8e5baa883fb3\") " pod="tigera-operator/tigera-operator-797db67f8-zrgmk" May 13 00:15:31.913361 kubelet[2230]: I0513 00:15:31.913336 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sfw2\" (UniqueName: \"kubernetes.io/projected/9d638da6-3528-4b32-93e9-8e5baa883fb3-kube-api-access-2sfw2\") pod \"tigera-operator-797db67f8-zrgmk\" (UID: \"9d638da6-3528-4b32-93e9-8e5baa883fb3\") " pod="tigera-operator/tigera-operator-797db67f8-zrgmk" May 13 00:15:32.022255 kubelet[2230]: E0513 00:15:32.022220 2230 projected.go:294] Couldn't get configMap tigera-operator/kube-root-ca.crt: configmap "kube-root-ca.crt" not found May 13 00:15:32.022423 kubelet[2230]: E0513 00:15:32.022399 2230 projected.go:200] Error preparing data for projected volume kube-api-access-2sfw2 for pod tigera-operator/tigera-operator-797db67f8-zrgmk: configmap "kube-root-ca.crt" not found May 13 00:15:32.022564 kubelet[2230]: E0513 00:15:32.022531 2230 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d638da6-3528-4b32-93e9-8e5baa883fb3-kube-api-access-2sfw2 podName:9d638da6-3528-4b32-93e9-8e5baa883fb3 nodeName:}" failed. No retries permitted until 2025-05-13 00:15:32.522508054 +0000 UTC m=+16.500730802 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-2sfw2" (UniqueName: "kubernetes.io/projected/9d638da6-3528-4b32-93e9-8e5baa883fb3-kube-api-access-2sfw2") pod "tigera-operator-797db67f8-zrgmk" (UID: "9d638da6-3528-4b32-93e9-8e5baa883fb3") : configmap "kube-root-ca.crt" not found May 13 00:15:32.211379 kubelet[2230]: I0513 00:15:32.211173 2230 topology_manager.go:215] "Topology Admit Handler" podUID="53103450-e7d3-428b-a4a4-54e003071863" podNamespace="kube-system" podName="kube-proxy-jvv2p" May 13 00:15:32.316429 kubelet[2230]: I0513 00:15:32.316370 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/53103450-e7d3-428b-a4a4-54e003071863-kube-proxy\") pod \"kube-proxy-jvv2p\" (UID: \"53103450-e7d3-428b-a4a4-54e003071863\") " pod="kube-system/kube-proxy-jvv2p" May 13 00:15:32.316429 kubelet[2230]: I0513 00:15:32.316425 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/53103450-e7d3-428b-a4a4-54e003071863-xtables-lock\") pod \"kube-proxy-jvv2p\" (UID: \"53103450-e7d3-428b-a4a4-54e003071863\") " pod="kube-system/kube-proxy-jvv2p" May 13 00:15:32.316429 kubelet[2230]: I0513 00:15:32.316448 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/53103450-e7d3-428b-a4a4-54e003071863-lib-modules\") pod \"kube-proxy-jvv2p\" (UID: \"53103450-e7d3-428b-a4a4-54e003071863\") " pod="kube-system/kube-proxy-jvv2p" May 13 00:15:32.316649 kubelet[2230]: I0513 00:15:32.316484 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfs8n\" (UniqueName: \"kubernetes.io/projected/53103450-e7d3-428b-a4a4-54e003071863-kube-api-access-wfs8n\") pod \"kube-proxy-jvv2p\" (UID: \"53103450-e7d3-428b-a4a4-54e003071863\") " pod="kube-system/kube-proxy-jvv2p" May 13 00:15:32.515028 kubelet[2230]: E0513 00:15:32.514929 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:32.515588 env[1323]: time="2025-05-13T00:15:32.515479807Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-jvv2p,Uid:53103450-e7d3-428b-a4a4-54e003071863,Namespace:kube-system,Attempt:0,}" May 13 00:15:32.529343 env[1323]: time="2025-05-13T00:15:32.529272271Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 13 00:15:32.529466 env[1323]: time="2025-05-13T00:15:32.529361231Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 13 00:15:32.529466 env[1323]: time="2025-05-13T00:15:32.529388631Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 13 00:15:32.529580 env[1323]: time="2025-05-13T00:15:32.529531992Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/e19689e88cbb6eea17a4b7f583d56026b59712a9150f88c5c797105c5b6165ce pid=2341 runtime=io.containerd.runc.v2 May 13 00:15:32.578250 env[1323]: time="2025-05-13T00:15:32.578199598Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-jvv2p,Uid:53103450-e7d3-428b-a4a4-54e003071863,Namespace:kube-system,Attempt:0,} returns sandbox id \"e19689e88cbb6eea17a4b7f583d56026b59712a9150f88c5c797105c5b6165ce\"" May 13 00:15:32.579257 kubelet[2230]: E0513 00:15:32.579169 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:32.582922 env[1323]: time="2025-05-13T00:15:32.582875846Z" level=info msg="CreateContainer within sandbox \"e19689e88cbb6eea17a4b7f583d56026b59712a9150f88c5c797105c5b6165ce\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 13 00:15:32.595528 env[1323]: time="2025-05-13T00:15:32.595482989Z" level=info msg="CreateContainer within sandbox \"e19689e88cbb6eea17a4b7f583d56026b59712a9150f88c5c797105c5b6165ce\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"0d9d514cfd870b1612b56e98c074d42f5e9fea4667ab996f87fd1830cfade8b3\"" May 13 00:15:32.596193 env[1323]: time="2025-05-13T00:15:32.596163710Z" level=info msg="StartContainer for \"0d9d514cfd870b1612b56e98c074d42f5e9fea4667ab996f87fd1830cfade8b3\"" May 13 00:15:32.681280 env[1323]: time="2025-05-13T00:15:32.681233460Z" level=info msg="StartContainer for \"0d9d514cfd870b1612b56e98c074d42f5e9fea4667ab996f87fd1830cfade8b3\" returns successfully" May 13 00:15:32.785880 env[1323]: time="2025-05-13T00:15:32.785780286Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-797db67f8-zrgmk,Uid:9d638da6-3528-4b32-93e9-8e5baa883fb3,Namespace:tigera-operator,Attempt:0,}" May 13 00:15:32.800006 env[1323]: time="2025-05-13T00:15:32.799901711Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 13 00:15:32.800006 env[1323]: time="2025-05-13T00:15:32.799944471Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 13 00:15:32.800006 env[1323]: time="2025-05-13T00:15:32.799955071Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 13 00:15:32.800590 env[1323]: time="2025-05-13T00:15:32.800385991Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/c53ee449c2f4523e473fd05fb456c6bc8797e30eb360047e630de3d3ea126e1a pid=2440 runtime=io.containerd.runc.v2 May 13 00:15:32.800000 audit[2452]: NETFILTER_CFG table=mangle:38 family=2 entries=1 op=nft_register_chain pid=2452 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.800000 audit[2452]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffc05203d0 a2=0 a3=1 items=0 ppid=2393 pid=2452 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.807790 kernel: audit: type=1325 audit(1747095332.800:237): table=mangle:38 family=2 entries=1 op=nft_register_chain pid=2452 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.807868 kernel: audit: type=1300 audit(1747095332.800:237): arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffc05203d0 a2=0 a3=1 items=0 ppid=2393 pid=2452 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.807886 kernel: audit: type=1327 audit(1747095332.800:237): proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 May 13 00:15:32.800000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 May 13 00:15:32.802000 audit[2453]: NETFILTER_CFG table=mangle:39 family=10 entries=1 op=nft_register_chain pid=2453 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.811939 kernel: audit: type=1325 audit(1747095332.802:238): table=mangle:39 family=10 entries=1 op=nft_register_chain pid=2453 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.812021 kernel: audit: type=1300 audit(1747095332.802:238): arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffeb58e5a0 a2=0 a3=1 items=0 ppid=2393 pid=2453 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.802000 audit[2453]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffeb58e5a0 a2=0 a3=1 items=0 ppid=2393 pid=2453 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.802000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 May 13 00:15:32.817593 kernel: audit: type=1327 audit(1747095332.802:238): proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 May 13 00:15:32.802000 audit[2456]: NETFILTER_CFG table=nat:40 family=2 entries=1 op=nft_register_chain pid=2456 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.819546 kernel: audit: type=1325 audit(1747095332.802:239): table=nat:40 family=2 entries=1 op=nft_register_chain pid=2456 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.819583 kernel: audit: type=1300 audit(1747095332.802:239): arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=fffffa5c0a10 a2=0 a3=1 items=0 ppid=2393 pid=2456 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.802000 audit[2456]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=fffffa5c0a10 a2=0 a3=1 items=0 ppid=2393 pid=2456 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.802000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 May 13 00:15:32.825559 kernel: audit: type=1327 audit(1747095332.802:239): proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 May 13 00:15:32.825618 kernel: audit: type=1325 audit(1747095332.803:240): table=nat:41 family=10 entries=1 op=nft_register_chain pid=2457 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.803000 audit[2457]: NETFILTER_CFG table=nat:41 family=10 entries=1 op=nft_register_chain pid=2457 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.803000 audit[2457]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffe48df610 a2=0 a3=1 items=0 ppid=2393 pid=2457 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.803000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 May 13 00:15:32.803000 audit[2458]: NETFILTER_CFG table=filter:42 family=2 entries=1 op=nft_register_chain pid=2458 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.803000 audit[2458]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffdb3457a0 a2=0 a3=1 items=0 ppid=2393 pid=2458 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.803000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 May 13 00:15:32.808000 audit[2462]: NETFILTER_CFG table=filter:43 family=10 entries=1 op=nft_register_chain pid=2462 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.808000 audit[2462]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffd48b62f0 a2=0 a3=1 items=0 ppid=2393 pid=2462 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.808000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 May 13 00:15:32.856237 env[1323]: time="2025-05-13T00:15:32.856186610Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-797db67f8-zrgmk,Uid:9d638da6-3528-4b32-93e9-8e5baa883fb3,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"c53ee449c2f4523e473fd05fb456c6bc8797e30eb360047e630de3d3ea126e1a\"" May 13 00:15:32.859073 env[1323]: time="2025-05-13T00:15:32.857842493Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\"" May 13 00:15:32.903000 audit[2483]: NETFILTER_CFG table=filter:44 family=2 entries=1 op=nft_register_chain pid=2483 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.903000 audit[2483]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=108 a0=3 a1=ffffc71c75c0 a2=0 a3=1 items=0 ppid=2393 pid=2483 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.903000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 May 13 00:15:32.907000 audit[2485]: NETFILTER_CFG table=filter:45 family=2 entries=1 op=nft_register_rule pid=2485 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.907000 audit[2485]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=752 a0=3 a1=ffffc0607f50 a2=0 a3=1 items=0 ppid=2393 pid=2485 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.907000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C652073657276696365 May 13 00:15:32.912000 audit[2488]: NETFILTER_CFG table=filter:46 family=2 entries=1 op=nft_register_rule pid=2488 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.912000 audit[2488]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=752 a0=3 a1=ffffcc5b7af0 a2=0 a3=1 items=0 ppid=2393 pid=2488 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.912000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669 May 13 00:15:32.913000 audit[2489]: NETFILTER_CFG table=filter:47 family=2 entries=1 op=nft_register_chain pid=2489 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.913000 audit[2489]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffd462e2f0 a2=0 a3=1 items=0 ppid=2393 pid=2489 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.913000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4E4F4445504F525453002D740066696C746572 May 13 00:15:32.915000 audit[2491]: NETFILTER_CFG table=filter:48 family=2 entries=1 op=nft_register_rule pid=2491 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.915000 audit[2491]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=528 a0=3 a1=ffffdcd8a1c0 a2=0 a3=1 items=0 ppid=2393 pid=2491 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.915000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 May 13 00:15:32.916000 audit[2492]: NETFILTER_CFG table=filter:49 family=2 entries=1 op=nft_register_chain pid=2492 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.916000 audit[2492]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffe2eb7850 a2=0 a3=1 items=0 ppid=2393 pid=2492 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.916000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D740066696C746572 May 13 00:15:32.919000 audit[2494]: NETFILTER_CFG table=filter:50 family=2 entries=1 op=nft_register_rule pid=2494 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.919000 audit[2494]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=744 a0=3 a1=ffffe0223e40 a2=0 a3=1 items=0 ppid=2393 pid=2494 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.919000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D May 13 00:15:32.922000 audit[2497]: NETFILTER_CFG table=filter:51 family=2 entries=1 op=nft_register_rule pid=2497 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.922000 audit[2497]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=744 a0=3 a1=ffffebae2290 a2=0 a3=1 items=0 ppid=2393 pid=2497 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.922000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D53 May 13 00:15:32.923000 audit[2498]: NETFILTER_CFG table=filter:52 family=2 entries=1 op=nft_register_chain pid=2498 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.923000 audit[2498]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffec561b40 a2=0 a3=1 items=0 ppid=2393 pid=2498 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.923000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D464F5257415244002D740066696C746572 May 13 00:15:32.925000 audit[2500]: NETFILTER_CFG table=filter:53 family=2 entries=1 op=nft_register_rule pid=2500 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.925000 audit[2500]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=528 a0=3 a1=ffffd359e460 a2=0 a3=1 items=0 ppid=2393 pid=2500 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.925000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 May 13 00:15:32.927000 audit[2501]: NETFILTER_CFG table=filter:54 family=2 entries=1 op=nft_register_chain pid=2501 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.927000 audit[2501]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffd7069480 a2=0 a3=1 items=0 ppid=2393 pid=2501 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.927000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 May 13 00:15:32.929000 audit[2503]: NETFILTER_CFG table=filter:55 family=2 entries=1 op=nft_register_rule pid=2503 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.929000 audit[2503]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=748 a0=3 a1=ffffe2f063a0 a2=0 a3=1 items=0 ppid=2393 pid=2503 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.929000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A May 13 00:15:32.932000 audit[2506]: NETFILTER_CFG table=filter:56 family=2 entries=1 op=nft_register_rule pid=2506 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.932000 audit[2506]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=748 a0=3 a1=ffffe35c40a0 a2=0 a3=1 items=0 ppid=2393 pid=2506 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.932000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A May 13 00:15:32.936000 audit[2509]: NETFILTER_CFG table=filter:57 family=2 entries=1 op=nft_register_rule pid=2509 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.936000 audit[2509]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=748 a0=3 a1=ffffe9c74820 a2=0 a3=1 items=0 ppid=2393 pid=2509 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.936000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D May 13 00:15:32.937000 audit[2510]: NETFILTER_CFG table=nat:58 family=2 entries=1 op=nft_register_chain pid=2510 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.937000 audit[2510]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=96 a0=3 a1=ffffede3a240 a2=0 a3=1 items=0 ppid=2393 pid=2510 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.937000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D74006E6174 May 13 00:15:32.939000 audit[2512]: NETFILTER_CFG table=nat:59 family=2 entries=1 op=nft_register_rule pid=2512 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.939000 audit[2512]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=524 a0=3 a1=ffffc24f6260 a2=0 a3=1 items=0 ppid=2393 pid=2512 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.939000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 May 13 00:15:32.942000 audit[2515]: NETFILTER_CFG table=nat:60 family=2 entries=1 op=nft_register_rule pid=2515 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.942000 audit[2515]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=528 a0=3 a1=ffffe6277d10 a2=0 a3=1 items=0 ppid=2393 pid=2515 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.942000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 May 13 00:15:32.943000 audit[2516]: NETFILTER_CFG table=nat:61 family=2 entries=1 op=nft_register_chain pid=2516 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.943000 audit[2516]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffdbd00790 a2=0 a3=1 items=0 ppid=2393 pid=2516 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.943000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D504F5354524F5554494E47002D74006E6174 May 13 00:15:32.945000 audit[2518]: NETFILTER_CFG table=nat:62 family=2 entries=1 op=nft_register_rule pid=2518 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 13 00:15:32.945000 audit[2518]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=532 a0=3 a1=ffffc4bd5ac0 a2=0 a3=1 items=0 ppid=2393 pid=2518 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.945000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 May 13 00:15:32.964000 audit[2524]: NETFILTER_CFG table=filter:63 family=2 entries=8 op=nft_register_rule pid=2524 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:15:32.964000 audit[2524]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=5164 a0=3 a1=fffff8c5e130 a2=0 a3=1 items=0 ppid=2393 pid=2524 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.964000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:32.973000 audit[2524]: NETFILTER_CFG table=nat:64 family=2 entries=14 op=nft_register_chain pid=2524 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:15:32.973000 audit[2524]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=5508 a0=3 a1=fffff8c5e130 a2=0 a3=1 items=0 ppid=2393 pid=2524 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.973000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:32.974000 audit[2530]: NETFILTER_CFG table=filter:65 family=10 entries=1 op=nft_register_chain pid=2530 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.974000 audit[2530]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=108 a0=3 a1=fffffe0f6a50 a2=0 a3=1 items=0 ppid=2393 pid=2530 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.974000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 May 13 00:15:32.976000 audit[2532]: NETFILTER_CFG table=filter:66 family=10 entries=2 op=nft_register_chain pid=2532 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.976000 audit[2532]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=836 a0=3 a1=ffffe1e561b0 a2=0 a3=1 items=0 ppid=2393 pid=2532 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.976000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C6520736572766963 May 13 00:15:32.980000 audit[2535]: NETFILTER_CFG table=filter:67 family=10 entries=2 op=nft_register_chain pid=2535 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.980000 audit[2535]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=836 a0=3 a1=ffffd81b1940 a2=0 a3=1 items=0 ppid=2393 pid=2535 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.980000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C652073657276 May 13 00:15:32.981000 audit[2536]: NETFILTER_CFG table=filter:68 family=10 entries=1 op=nft_register_chain pid=2536 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.981000 audit[2536]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffeba33f60 a2=0 a3=1 items=0 ppid=2393 pid=2536 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.981000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4E4F4445504F525453002D740066696C746572 May 13 00:15:32.983000 audit[2538]: NETFILTER_CFG table=filter:69 family=10 entries=1 op=nft_register_rule pid=2538 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.983000 audit[2538]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=528 a0=3 a1=ffffd79b4d10 a2=0 a3=1 items=0 ppid=2393 pid=2538 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.983000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 May 13 00:15:32.984000 audit[2539]: NETFILTER_CFG table=filter:70 family=10 entries=1 op=nft_register_chain pid=2539 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.984000 audit[2539]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=fffff9726930 a2=0 a3=1 items=0 ppid=2393 pid=2539 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.984000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D740066696C746572 May 13 00:15:32.987000 audit[2541]: NETFILTER_CFG table=filter:71 family=10 entries=1 op=nft_register_rule pid=2541 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.987000 audit[2541]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=744 a0=3 a1=ffffe9673570 a2=0 a3=1 items=0 ppid=2393 pid=2541 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.987000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B554245 May 13 00:15:32.991000 audit[2544]: NETFILTER_CFG table=filter:72 family=10 entries=2 op=nft_register_chain pid=2544 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.991000 audit[2544]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=828 a0=3 a1=ffffc542af50 a2=0 a3=1 items=0 ppid=2393 pid=2544 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.991000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D May 13 00:15:32.992000 audit[2545]: NETFILTER_CFG table=filter:73 family=10 entries=1 op=nft_register_chain pid=2545 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.992000 audit[2545]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffd830baa0 a2=0 a3=1 items=0 ppid=2393 pid=2545 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.992000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D464F5257415244002D740066696C746572 May 13 00:15:32.995000 audit[2547]: NETFILTER_CFG table=filter:74 family=10 entries=1 op=nft_register_rule pid=2547 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.995000 audit[2547]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=528 a0=3 a1=ffffc542c630 a2=0 a3=1 items=0 ppid=2393 pid=2547 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.995000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 May 13 00:15:32.996000 audit[2548]: NETFILTER_CFG table=filter:75 family=10 entries=1 op=nft_register_chain pid=2548 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:32.996000 audit[2548]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffc72e9780 a2=0 a3=1 items=0 ppid=2393 pid=2548 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:32.996000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 May 13 00:15:33.002000 audit[2550]: NETFILTER_CFG table=filter:76 family=10 entries=1 op=nft_register_rule pid=2550 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:33.002000 audit[2550]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=748 a0=3 a1=ffffea6859a0 a2=0 a3=1 items=0 ppid=2393 pid=2550 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:33.002000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A May 13 00:15:33.005000 audit[2553]: NETFILTER_CFG table=filter:77 family=10 entries=1 op=nft_register_rule pid=2553 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:33.005000 audit[2553]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=748 a0=3 a1=ffffcd813ec0 a2=0 a3=1 items=0 ppid=2393 pid=2553 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:33.005000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D May 13 00:15:33.008000 audit[2556]: NETFILTER_CFG table=filter:78 family=10 entries=1 op=nft_register_rule pid=2556 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:33.008000 audit[2556]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=748 a0=3 a1=ffffd49cae50 a2=0 a3=1 items=0 ppid=2393 pid=2556 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:33.008000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C May 13 00:15:33.009000 audit[2557]: NETFILTER_CFG table=nat:79 family=10 entries=1 op=nft_register_chain pid=2557 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:33.009000 audit[2557]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=96 a0=3 a1=ffffede26090 a2=0 a3=1 items=0 ppid=2393 pid=2557 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:33.009000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D74006E6174 May 13 00:15:33.012000 audit[2559]: NETFILTER_CFG table=nat:80 family=10 entries=2 op=nft_register_chain pid=2559 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:33.012000 audit[2559]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=600 a0=3 a1=ffffc2149380 a2=0 a3=1 items=0 ppid=2393 pid=2559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:33.012000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 May 13 00:15:33.015000 audit[2562]: NETFILTER_CFG table=nat:81 family=10 entries=2 op=nft_register_chain pid=2562 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:33.015000 audit[2562]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=608 a0=3 a1=fffff1408e80 a2=0 a3=1 items=0 ppid=2393 pid=2562 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:33.015000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 May 13 00:15:33.016000 audit[2563]: NETFILTER_CFG table=nat:82 family=10 entries=1 op=nft_register_chain pid=2563 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:33.016000 audit[2563]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffe9181150 a2=0 a3=1 items=0 ppid=2393 pid=2563 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:33.016000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D504F5354524F5554494E47002D74006E6174 May 13 00:15:33.018000 audit[2565]: NETFILTER_CFG table=nat:83 family=10 entries=2 op=nft_register_chain pid=2565 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:33.018000 audit[2565]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=612 a0=3 a1=ffffd7783410 a2=0 a3=1 items=0 ppid=2393 pid=2565 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:33.018000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 May 13 00:15:33.019000 audit[2566]: NETFILTER_CFG table=filter:84 family=10 entries=1 op=nft_register_chain pid=2566 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:33.019000 audit[2566]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffc19ac480 a2=0 a3=1 items=0 ppid=2393 pid=2566 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:33.019000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 May 13 00:15:33.023000 audit[2568]: NETFILTER_CFG table=filter:85 family=10 entries=1 op=nft_register_rule pid=2568 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:33.023000 audit[2568]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=228 a0=3 a1=ffffcb5e3e60 a2=0 a3=1 items=0 ppid=2393 pid=2568 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:33.023000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C May 13 00:15:33.026000 audit[2571]: NETFILTER_CFG table=filter:86 family=10 entries=1 op=nft_register_rule pid=2571 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 13 00:15:33.026000 audit[2571]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=228 a0=3 a1=ffffcd217110 a2=0 a3=1 items=0 ppid=2393 pid=2571 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:33.026000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C May 13 00:15:33.029000 audit[2573]: NETFILTER_CFG table=filter:87 family=10 entries=3 op=nft_register_rule pid=2573 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" May 13 00:15:33.029000 audit[2573]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2004 a0=3 a1=fffff3485020 a2=0 a3=1 items=0 ppid=2393 pid=2573 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:33.029000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:33.030000 audit[2573]: NETFILTER_CFG table=nat:88 family=10 entries=7 op=nft_register_chain pid=2573 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" May 13 00:15:33.030000 audit[2573]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2056 a0=3 a1=fffff3485020 a2=0 a3=1 items=0 ppid=2393 pid=2573 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:33.030000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:33.172817 kubelet[2230]: E0513 00:15:33.172791 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:33.182929 kubelet[2230]: I0513 00:15:33.182858 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-jvv2p" podStartSLOduration=1.182839969 podStartE2EDuration="1.182839969s" podCreationTimestamp="2025-05-13 00:15:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-13 00:15:33.181904207 +0000 UTC m=+17.160126955" watchObservedRunningTime="2025-05-13 00:15:33.182839969 +0000 UTC m=+17.161062757" May 13 00:15:33.805872 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2919245228.mount: Deactivated successfully. May 13 00:15:34.327821 env[1323]: time="2025-05-13T00:15:34.327772396Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/tigera/operator:v1.36.7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:34.329246 env[1323]: time="2025-05-13T00:15:34.329215159Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:27f7c2cfac802523e44ecd16453a4cc992f6c7d610c13054f2715a7cb4370565,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:34.331607 env[1323]: time="2025-05-13T00:15:34.331566002Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/tigera/operator:v1.36.7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:34.332610 env[1323]: time="2025-05-13T00:15:34.332582244Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:34.333143 env[1323]: time="2025-05-13T00:15:34.333114485Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\" returns image reference \"sha256:27f7c2cfac802523e44ecd16453a4cc992f6c7d610c13054f2715a7cb4370565\"" May 13 00:15:34.336113 env[1323]: time="2025-05-13T00:15:34.336078449Z" level=info msg="CreateContainer within sandbox \"c53ee449c2f4523e473fd05fb456c6bc8797e30eb360047e630de3d3ea126e1a\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" May 13 00:15:34.361591 env[1323]: time="2025-05-13T00:15:34.361503209Z" level=info msg="CreateContainer within sandbox \"c53ee449c2f4523e473fd05fb456c6bc8797e30eb360047e630de3d3ea126e1a\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"7068528d92479e9b02c19ef1496f90a3ad36a58c245170455fba5233cf594746\"" May 13 00:15:34.362408 env[1323]: time="2025-05-13T00:15:34.362350810Z" level=info msg="StartContainer for \"7068528d92479e9b02c19ef1496f90a3ad36a58c245170455fba5233cf594746\"" May 13 00:15:34.424895 env[1323]: time="2025-05-13T00:15:34.424840627Z" level=info msg="StartContainer for \"7068528d92479e9b02c19ef1496f90a3ad36a58c245170455fba5233cf594746\" returns successfully" May 13 00:15:37.703000 audit[2614]: NETFILTER_CFG table=filter:89 family=2 entries=15 op=nft_register_rule pid=2614 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:15:37.703000 audit[2614]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=5908 a0=3 a1=ffffd1fe08d0 a2=0 a3=1 items=0 ppid=2393 pid=2614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:37.703000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:37.730000 audit[2614]: NETFILTER_CFG table=nat:90 family=2 entries=12 op=nft_register_rule pid=2614 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:15:37.730000 audit[2614]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2700 a0=3 a1=ffffd1fe08d0 a2=0 a3=1 items=0 ppid=2393 pid=2614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:37.730000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:37.744000 audit[2616]: NETFILTER_CFG table=filter:91 family=2 entries=16 op=nft_register_rule pid=2616 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:15:37.744000 audit[2616]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=5908 a0=3 a1=ffffc590aae0 a2=0 a3=1 items=0 ppid=2393 pid=2616 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:37.744000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:37.754000 audit[2616]: NETFILTER_CFG table=nat:92 family=2 entries=12 op=nft_register_rule pid=2616 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:15:37.754000 audit[2616]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2700 a0=3 a1=ffffc590aae0 a2=0 a3=1 items=0 ppid=2393 pid=2616 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:37.754000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:38.285723 kubelet[2230]: I0513 00:15:38.285604 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-797db67f8-zrgmk" podStartSLOduration=5.808463023 podStartE2EDuration="7.285584498s" podCreationTimestamp="2025-05-13 00:15:31 +0000 UTC" firstStartedPulling="2025-05-13 00:15:32.857267572 +0000 UTC m=+16.835490320" lastFinishedPulling="2025-05-13 00:15:34.334389047 +0000 UTC m=+18.312611795" observedRunningTime="2025-05-13 00:15:35.183894551 +0000 UTC m=+19.162117299" watchObservedRunningTime="2025-05-13 00:15:38.285584498 +0000 UTC m=+22.263807246" May 13 00:15:38.286144 kubelet[2230]: I0513 00:15:38.285806 2230 topology_manager.go:215] "Topology Admit Handler" podUID="a74985bf-8cc3-4126-846b-e92cbfc687c7" podNamespace="calico-system" podName="calico-typha-85f4d64f96-sll44" May 13 00:15:38.341900 kubelet[2230]: I0513 00:15:38.341848 2230 topology_manager.go:215] "Topology Admit Handler" podUID="3aef28fa-07db-410f-a490-9ff3ec65460f" podNamespace="calico-system" podName="calico-node-bsdv5" May 13 00:15:38.458031 kubelet[2230]: I0513 00:15:38.457990 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a74985bf-8cc3-4126-846b-e92cbfc687c7-tigera-ca-bundle\") pod \"calico-typha-85f4d64f96-sll44\" (UID: \"a74985bf-8cc3-4126-846b-e92cbfc687c7\") " pod="calico-system/calico-typha-85f4d64f96-sll44" May 13 00:15:38.458188 kubelet[2230]: I0513 00:15:38.458083 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3aef28fa-07db-410f-a490-9ff3ec65460f-lib-modules\") pod \"calico-node-bsdv5\" (UID: \"3aef28fa-07db-410f-a490-9ff3ec65460f\") " pod="calico-system/calico-node-bsdv5" May 13 00:15:38.458188 kubelet[2230]: I0513 00:15:38.458105 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/3aef28fa-07db-410f-a490-9ff3ec65460f-flexvol-driver-host\") pod \"calico-node-bsdv5\" (UID: \"3aef28fa-07db-410f-a490-9ff3ec65460f\") " pod="calico-system/calico-node-bsdv5" May 13 00:15:38.458188 kubelet[2230]: I0513 00:15:38.458129 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvpqw\" (UniqueName: \"kubernetes.io/projected/a74985bf-8cc3-4126-846b-e92cbfc687c7-kube-api-access-zvpqw\") pod \"calico-typha-85f4d64f96-sll44\" (UID: \"a74985bf-8cc3-4126-846b-e92cbfc687c7\") " pod="calico-system/calico-typha-85f4d64f96-sll44" May 13 00:15:38.458188 kubelet[2230]: I0513 00:15:38.458146 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/3aef28fa-07db-410f-a490-9ff3ec65460f-var-run-calico\") pod \"calico-node-bsdv5\" (UID: \"3aef28fa-07db-410f-a490-9ff3ec65460f\") " pod="calico-system/calico-node-bsdv5" May 13 00:15:38.458188 kubelet[2230]: I0513 00:15:38.458188 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/3aef28fa-07db-410f-a490-9ff3ec65460f-cni-net-dir\") pod \"calico-node-bsdv5\" (UID: \"3aef28fa-07db-410f-a490-9ff3ec65460f\") " pod="calico-system/calico-node-bsdv5" May 13 00:15:38.458311 kubelet[2230]: I0513 00:15:38.458205 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn9jl\" (UniqueName: \"kubernetes.io/projected/3aef28fa-07db-410f-a490-9ff3ec65460f-kube-api-access-jn9jl\") pod \"calico-node-bsdv5\" (UID: \"3aef28fa-07db-410f-a490-9ff3ec65460f\") " pod="calico-system/calico-node-bsdv5" May 13 00:15:38.458311 kubelet[2230]: I0513 00:15:38.458255 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/3aef28fa-07db-410f-a490-9ff3ec65460f-cni-bin-dir\") pod \"calico-node-bsdv5\" (UID: \"3aef28fa-07db-410f-a490-9ff3ec65460f\") " pod="calico-system/calico-node-bsdv5" May 13 00:15:38.458311 kubelet[2230]: I0513 00:15:38.458275 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aef28fa-07db-410f-a490-9ff3ec65460f-tigera-ca-bundle\") pod \"calico-node-bsdv5\" (UID: \"3aef28fa-07db-410f-a490-9ff3ec65460f\") " pod="calico-system/calico-node-bsdv5" May 13 00:15:38.458311 kubelet[2230]: I0513 00:15:38.458305 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/3aef28fa-07db-410f-a490-9ff3ec65460f-node-certs\") pod \"calico-node-bsdv5\" (UID: \"3aef28fa-07db-410f-a490-9ff3ec65460f\") " pod="calico-system/calico-node-bsdv5" May 13 00:15:38.458417 kubelet[2230]: I0513 00:15:38.458319 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/3aef28fa-07db-410f-a490-9ff3ec65460f-var-lib-calico\") pod \"calico-node-bsdv5\" (UID: \"3aef28fa-07db-410f-a490-9ff3ec65460f\") " pod="calico-system/calico-node-bsdv5" May 13 00:15:38.458417 kubelet[2230]: I0513 00:15:38.458334 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/3aef28fa-07db-410f-a490-9ff3ec65460f-cni-log-dir\") pod \"calico-node-bsdv5\" (UID: \"3aef28fa-07db-410f-a490-9ff3ec65460f\") " pod="calico-system/calico-node-bsdv5" May 13 00:15:38.458417 kubelet[2230]: I0513 00:15:38.458381 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/3aef28fa-07db-410f-a490-9ff3ec65460f-policysync\") pod \"calico-node-bsdv5\" (UID: \"3aef28fa-07db-410f-a490-9ff3ec65460f\") " pod="calico-system/calico-node-bsdv5" May 13 00:15:38.458490 kubelet[2230]: I0513 00:15:38.458417 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/a74985bf-8cc3-4126-846b-e92cbfc687c7-typha-certs\") pod \"calico-typha-85f4d64f96-sll44\" (UID: \"a74985bf-8cc3-4126-846b-e92cbfc687c7\") " pod="calico-system/calico-typha-85f4d64f96-sll44" May 13 00:15:38.458490 kubelet[2230]: I0513 00:15:38.458433 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3aef28fa-07db-410f-a490-9ff3ec65460f-xtables-lock\") pod \"calico-node-bsdv5\" (UID: \"3aef28fa-07db-410f-a490-9ff3ec65460f\") " pod="calico-system/calico-node-bsdv5" May 13 00:15:38.478906 kubelet[2230]: I0513 00:15:38.478862 2230 topology_manager.go:215] "Topology Admit Handler" podUID="838e7283-9b17-4532-a9f4-4b17f2b5ca3b" podNamespace="calico-system" podName="csi-node-driver-srb86" May 13 00:15:38.479155 kubelet[2230]: E0513 00:15:38.479124 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-srb86" podUID="838e7283-9b17-4532-a9f4-4b17f2b5ca3b" May 13 00:15:38.561760 kubelet[2230]: E0513 00:15:38.561603 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.561760 kubelet[2230]: W0513 00:15:38.561636 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.561760 kubelet[2230]: E0513 00:15:38.561659 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.562499 kubelet[2230]: E0513 00:15:38.562406 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.562499 kubelet[2230]: W0513 00:15:38.562421 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.562499 kubelet[2230]: E0513 00:15:38.562432 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.562679 kubelet[2230]: E0513 00:15:38.562590 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.562679 kubelet[2230]: W0513 00:15:38.562598 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.562679 kubelet[2230]: E0513 00:15:38.562606 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.562752 kubelet[2230]: E0513 00:15:38.562745 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.562752 kubelet[2230]: W0513 00:15:38.562752 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.562799 kubelet[2230]: E0513 00:15:38.562758 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.562923 kubelet[2230]: E0513 00:15:38.562901 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.562923 kubelet[2230]: W0513 00:15:38.562918 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.562987 kubelet[2230]: E0513 00:15:38.562927 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.563045 kubelet[2230]: E0513 00:15:38.563031 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.563045 kubelet[2230]: W0513 00:15:38.563042 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.563103 kubelet[2230]: E0513 00:15:38.563049 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.563162 kubelet[2230]: E0513 00:15:38.563150 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.563162 kubelet[2230]: W0513 00:15:38.563160 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.563219 kubelet[2230]: E0513 00:15:38.563167 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.563280 kubelet[2230]: E0513 00:15:38.563268 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.563280 kubelet[2230]: W0513 00:15:38.563273 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.563280 kubelet[2230]: E0513 00:15:38.563280 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.563599 kubelet[2230]: E0513 00:15:38.563506 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.563599 kubelet[2230]: W0513 00:15:38.563523 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.563599 kubelet[2230]: E0513 00:15:38.563553 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.567697 kubelet[2230]: E0513 00:15:38.567674 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.567697 kubelet[2230]: W0513 00:15:38.567691 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.567869 kubelet[2230]: E0513 00:15:38.567849 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.567935 kubelet[2230]: E0513 00:15:38.567861 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.567994 kubelet[2230]: W0513 00:15:38.567981 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.568138 kubelet[2230]: E0513 00:15:38.568125 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.568318 kubelet[2230]: E0513 00:15:38.568307 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.568398 kubelet[2230]: W0513 00:15:38.568376 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.568531 kubelet[2230]: E0513 00:15:38.568520 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.570780 kubelet[2230]: E0513 00:15:38.570758 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.570878 kubelet[2230]: W0513 00:15:38.570863 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.571027 kubelet[2230]: E0513 00:15:38.570998 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.572209 kubelet[2230]: E0513 00:15:38.572187 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.572395 kubelet[2230]: W0513 00:15:38.572367 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.572563 kubelet[2230]: E0513 00:15:38.572523 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.572768 kubelet[2230]: E0513 00:15:38.572751 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.572840 kubelet[2230]: W0513 00:15:38.572827 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.572989 kubelet[2230]: E0513 00:15:38.572962 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.573168 kubelet[2230]: E0513 00:15:38.573153 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.573243 kubelet[2230]: W0513 00:15:38.573230 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.573462 kubelet[2230]: E0513 00:15:38.573450 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.573574 kubelet[2230]: W0513 00:15:38.573520 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.573793 kubelet[2230]: E0513 00:15:38.573780 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.573872 kubelet[2230]: W0513 00:15:38.573859 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.574074 kubelet[2230]: E0513 00:15:38.574062 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.574139 kubelet[2230]: W0513 00:15:38.574128 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.574357 kubelet[2230]: E0513 00:15:38.574345 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.574441 kubelet[2230]: W0513 00:15:38.574427 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.574669 kubelet[2230]: E0513 00:15:38.574656 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.574749 kubelet[2230]: W0513 00:15:38.574736 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.574944 kubelet[2230]: E0513 00:15:38.574933 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.575007 kubelet[2230]: W0513 00:15:38.574996 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.575196 kubelet[2230]: E0513 00:15:38.575185 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.575296 kubelet[2230]: W0513 00:15:38.575283 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.575352 kubelet[2230]: E0513 00:15:38.575341 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.575595 kubelet[2230]: E0513 00:15:38.575582 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.575664 kubelet[2230]: W0513 00:15:38.575652 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.575729 kubelet[2230]: E0513 00:15:38.575718 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.576006 kubelet[2230]: E0513 00:15:38.575994 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.576080 kubelet[2230]: W0513 00:15:38.576067 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.576140 kubelet[2230]: E0513 00:15:38.576129 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.576224 kubelet[2230]: E0513 00:15:38.576194 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.576311 kubelet[2230]: E0513 00:15:38.576273 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.576366 kubelet[2230]: E0513 00:15:38.576260 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.576520 kubelet[2230]: E0513 00:15:38.576290 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.576604 kubelet[2230]: E0513 00:15:38.576298 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.576671 kubelet[2230]: E0513 00:15:38.576304 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.576782 kubelet[2230]: E0513 00:15:38.576403 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.576849 kubelet[2230]: W0513 00:15:38.576834 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.579609 kubelet[2230]: E0513 00:15:38.577421 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.579609 kubelet[2230]: E0513 00:15:38.576206 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.579609 kubelet[2230]: E0513 00:15:38.577742 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.579609 kubelet[2230]: W0513 00:15:38.577755 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.579609 kubelet[2230]: E0513 00:15:38.577774 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.579609 kubelet[2230]: E0513 00:15:38.577935 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.579609 kubelet[2230]: W0513 00:15:38.577945 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.579609 kubelet[2230]: E0513 00:15:38.577954 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.579609 kubelet[2230]: E0513 00:15:38.578173 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.579609 kubelet[2230]: W0513 00:15:38.578183 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.579969 kubelet[2230]: E0513 00:15:38.578269 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.579969 kubelet[2230]: E0513 00:15:38.578572 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.579969 kubelet[2230]: W0513 00:15:38.578594 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.579969 kubelet[2230]: E0513 00:15:38.578675 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.580194 kubelet[2230]: E0513 00:15:38.580176 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.580262 kubelet[2230]: W0513 00:15:38.580250 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.580462 kubelet[2230]: E0513 00:15:38.580447 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.586852 kubelet[2230]: E0513 00:15:38.586793 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.587075 kubelet[2230]: W0513 00:15:38.587034 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.587947 kubelet[2230]: E0513 00:15:38.587913 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.588265 kubelet[2230]: E0513 00:15:38.588246 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.588439 kubelet[2230]: W0513 00:15:38.588336 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.588624 kubelet[2230]: E0513 00:15:38.588594 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.588844 kubelet[2230]: E0513 00:15:38.588829 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.588924 kubelet[2230]: W0513 00:15:38.588911 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.589107 kubelet[2230]: E0513 00:15:38.589090 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.589494 kubelet[2230]: E0513 00:15:38.589475 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.589631 kubelet[2230]: W0513 00:15:38.589616 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.589741 kubelet[2230]: E0513 00:15:38.589712 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.590019 kubelet[2230]: E0513 00:15:38.590001 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.590107 kubelet[2230]: W0513 00:15:38.590093 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.590241 kubelet[2230]: E0513 00:15:38.590214 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.590377 kubelet[2230]: E0513 00:15:38.590362 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.590467 kubelet[2230]: W0513 00:15:38.590452 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.590715 kubelet[2230]: E0513 00:15:38.590697 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.591025 kubelet[2230]: E0513 00:15:38.591008 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.591131 kubelet[2230]: W0513 00:15:38.591116 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.591257 kubelet[2230]: E0513 00:15:38.591241 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.591689 kubelet[2230]: E0513 00:15:38.591672 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.591791 kubelet[2230]: W0513 00:15:38.591777 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.591941 kubelet[2230]: E0513 00:15:38.591928 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.593000 kubelet[2230]: E0513 00:15:38.592964 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.593105 kubelet[2230]: W0513 00:15:38.593091 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.593352 kubelet[2230]: E0513 00:15:38.593190 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.593502 kubelet[2230]: E0513 00:15:38.593487 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.593502 kubelet[2230]: W0513 00:15:38.593503 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.593662 kubelet[2230]: E0513 00:15:38.593519 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.593723 kubelet[2230]: E0513 00:15:38.593711 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.593723 kubelet[2230]: W0513 00:15:38.593722 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.593782 kubelet[2230]: E0513 00:15:38.593734 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.594165 kubelet[2230]: E0513 00:15:38.594150 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.594211 kubelet[2230]: W0513 00:15:38.594193 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.594211 kubelet[2230]: E0513 00:15:38.594210 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.594939 kubelet[2230]: E0513 00:15:38.594924 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.594939 kubelet[2230]: W0513 00:15:38.594940 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.595517 kubelet[2230]: E0513 00:15:38.594952 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.595517 kubelet[2230]: E0513 00:15:38.595112 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.595517 kubelet[2230]: W0513 00:15:38.595119 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.595517 kubelet[2230]: E0513 00:15:38.595127 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.595517 kubelet[2230]: E0513 00:15:38.595274 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.595517 kubelet[2230]: W0513 00:15:38.595282 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.595517 kubelet[2230]: E0513 00:15:38.595289 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.596592 kubelet[2230]: E0513 00:15:38.596561 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.596592 kubelet[2230]: W0513 00:15:38.596589 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.596686 kubelet[2230]: E0513 00:15:38.596603 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.596806 kubelet[2230]: E0513 00:15:38.596788 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:38.596971 kubelet[2230]: E0513 00:15:38.596948 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.597030 kubelet[2230]: W0513 00:15:38.596979 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.597069 kubelet[2230]: E0513 00:15:38.597031 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.598900 kubelet[2230]: E0513 00:15:38.598569 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.598900 kubelet[2230]: W0513 00:15:38.598589 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.598900 kubelet[2230]: E0513 00:15:38.598605 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.598900 kubelet[2230]: E0513 00:15:38.598883 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.598900 kubelet[2230]: W0513 00:15:38.598894 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.598900 kubelet[2230]: E0513 00:15:38.598904 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.599657 env[1323]: time="2025-05-13T00:15:38.598802434Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-85f4d64f96-sll44,Uid:a74985bf-8cc3-4126-846b-e92cbfc687c7,Namespace:calico-system,Attempt:0,}" May 13 00:15:38.599919 kubelet[2230]: E0513 00:15:38.599168 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.599919 kubelet[2230]: W0513 00:15:38.599219 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.599919 kubelet[2230]: E0513 00:15:38.599415 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.599919 kubelet[2230]: E0513 00:15:38.599663 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.599919 kubelet[2230]: W0513 00:15:38.599674 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.599919 kubelet[2230]: E0513 00:15:38.599683 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.599919 kubelet[2230]: E0513 00:15:38.599816 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.599919 kubelet[2230]: W0513 00:15:38.599823 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.599919 kubelet[2230]: E0513 00:15:38.599830 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.600133 kubelet[2230]: E0513 00:15:38.599948 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.600133 kubelet[2230]: W0513 00:15:38.599955 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.600133 kubelet[2230]: E0513 00:15:38.599962 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.615581 env[1323]: time="2025-05-13T00:15:38.615453814Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 13 00:15:38.615581 env[1323]: time="2025-05-13T00:15:38.615516934Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 13 00:15:38.615581 env[1323]: time="2025-05-13T00:15:38.615528334Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 13 00:15:38.615875 env[1323]: time="2025-05-13T00:15:38.615761815Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/d60b9b69d98c5b3f8a5412b548c794bf90fb66766586bdf4ae59348be85dcbd8 pid=2682 runtime=io.containerd.runc.v2 May 13 00:15:38.645233 kubelet[2230]: E0513 00:15:38.645197 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:38.647946 env[1323]: time="2025-05-13T00:15:38.647910933Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-bsdv5,Uid:3aef28fa-07db-410f-a490-9ff3ec65460f,Namespace:calico-system,Attempt:0,}" May 13 00:15:38.660001 kubelet[2230]: E0513 00:15:38.659974 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.660001 kubelet[2230]: W0513 00:15:38.659996 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.660001 kubelet[2230]: E0513 00:15:38.660014 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.660178 kubelet[2230]: I0513 00:15:38.660044 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxfkm\" (UniqueName: \"kubernetes.io/projected/838e7283-9b17-4532-a9f4-4b17f2b5ca3b-kube-api-access-fxfkm\") pod \"csi-node-driver-srb86\" (UID: \"838e7283-9b17-4532-a9f4-4b17f2b5ca3b\") " pod="calico-system/csi-node-driver-srb86" May 13 00:15:38.660236 kubelet[2230]: E0513 00:15:38.660220 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.660236 kubelet[2230]: W0513 00:15:38.660235 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.660295 kubelet[2230]: E0513 00:15:38.660246 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.660295 kubelet[2230]: I0513 00:15:38.660260 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/838e7283-9b17-4532-a9f4-4b17f2b5ca3b-registration-dir\") pod \"csi-node-driver-srb86\" (UID: \"838e7283-9b17-4532-a9f4-4b17f2b5ca3b\") " pod="calico-system/csi-node-driver-srb86" May 13 00:15:38.660484 kubelet[2230]: E0513 00:15:38.660472 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.660484 kubelet[2230]: W0513 00:15:38.660484 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.660573 kubelet[2230]: E0513 00:15:38.660494 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.660573 kubelet[2230]: I0513 00:15:38.660509 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/838e7283-9b17-4532-a9f4-4b17f2b5ca3b-varrun\") pod \"csi-node-driver-srb86\" (UID: \"838e7283-9b17-4532-a9f4-4b17f2b5ca3b\") " pod="calico-system/csi-node-driver-srb86" May 13 00:15:38.660707 kubelet[2230]: E0513 00:15:38.660683 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.660707 kubelet[2230]: W0513 00:15:38.660698 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.660707 kubelet[2230]: E0513 00:15:38.660708 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.660795 kubelet[2230]: I0513 00:15:38.660722 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/838e7283-9b17-4532-a9f4-4b17f2b5ca3b-kubelet-dir\") pod \"csi-node-driver-srb86\" (UID: \"838e7283-9b17-4532-a9f4-4b17f2b5ca3b\") " pod="calico-system/csi-node-driver-srb86" May 13 00:15:38.661015 kubelet[2230]: E0513 00:15:38.660887 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.661015 kubelet[2230]: W0513 00:15:38.660898 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.661015 kubelet[2230]: E0513 00:15:38.660907 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.661015 kubelet[2230]: I0513 00:15:38.660920 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/838e7283-9b17-4532-a9f4-4b17f2b5ca3b-socket-dir\") pod \"csi-node-driver-srb86\" (UID: \"838e7283-9b17-4532-a9f4-4b17f2b5ca3b\") " pod="calico-system/csi-node-driver-srb86" May 13 00:15:38.661318 kubelet[2230]: E0513 00:15:38.661200 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.661318 kubelet[2230]: W0513 00:15:38.661217 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.661318 kubelet[2230]: E0513 00:15:38.661238 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.661636 kubelet[2230]: E0513 00:15:38.661491 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.661636 kubelet[2230]: W0513 00:15:38.661503 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.661636 kubelet[2230]: E0513 00:15:38.661577 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.661921 kubelet[2230]: E0513 00:15:38.661812 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.661921 kubelet[2230]: W0513 00:15:38.661825 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.661921 kubelet[2230]: E0513 00:15:38.661882 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.662195 kubelet[2230]: E0513 00:15:38.662084 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.662195 kubelet[2230]: W0513 00:15:38.662096 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.662195 kubelet[2230]: E0513 00:15:38.662107 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.662478 kubelet[2230]: E0513 00:15:38.662350 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.662478 kubelet[2230]: W0513 00:15:38.662362 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.662478 kubelet[2230]: E0513 00:15:38.662372 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.662767 kubelet[2230]: E0513 00:15:38.662660 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.662767 kubelet[2230]: W0513 00:15:38.662673 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.662767 kubelet[2230]: E0513 00:15:38.662684 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.663120 kubelet[2230]: E0513 00:15:38.662992 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.663120 kubelet[2230]: W0513 00:15:38.663005 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.663120 kubelet[2230]: E0513 00:15:38.663015 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.663423 kubelet[2230]: E0513 00:15:38.663293 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.663423 kubelet[2230]: W0513 00:15:38.663306 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.663423 kubelet[2230]: E0513 00:15:38.663316 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.663725 kubelet[2230]: E0513 00:15:38.663606 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.663725 kubelet[2230]: W0513 00:15:38.663619 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.663725 kubelet[2230]: E0513 00:15:38.663630 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.663951 kubelet[2230]: E0513 00:15:38.663899 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.663951 kubelet[2230]: W0513 00:15:38.663912 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.663951 kubelet[2230]: E0513 00:15:38.663922 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.704755 env[1323]: time="2025-05-13T00:15:38.704713122Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-85f4d64f96-sll44,Uid:a74985bf-8cc3-4126-846b-e92cbfc687c7,Namespace:calico-system,Attempt:0,} returns sandbox id \"d60b9b69d98c5b3f8a5412b548c794bf90fb66766586bdf4ae59348be85dcbd8\"" May 13 00:15:38.705409 kubelet[2230]: E0513 00:15:38.705382 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:38.706614 env[1323]: time="2025-05-13T00:15:38.706181323Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\"" May 13 00:15:38.761960 kubelet[2230]: E0513 00:15:38.761927 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.761960 kubelet[2230]: W0513 00:15:38.761949 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.761960 kubelet[2230]: E0513 00:15:38.761971 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.762442 kubelet[2230]: E0513 00:15:38.762230 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.762442 kubelet[2230]: W0513 00:15:38.762237 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.762442 kubelet[2230]: E0513 00:15:38.762253 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.762631 kubelet[2230]: E0513 00:15:38.762456 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.762631 kubelet[2230]: W0513 00:15:38.762465 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.762631 kubelet[2230]: E0513 00:15:38.762474 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.762714 kubelet[2230]: E0513 00:15:38.762657 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.762714 kubelet[2230]: W0513 00:15:38.762665 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.762714 kubelet[2230]: E0513 00:15:38.762673 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.762929 kubelet[2230]: E0513 00:15:38.762814 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.762929 kubelet[2230]: W0513 00:15:38.762828 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.762929 kubelet[2230]: E0513 00:15:38.762837 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.763199 kubelet[2230]: E0513 00:15:38.763092 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.763199 kubelet[2230]: W0513 00:15:38.763109 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.763199 kubelet[2230]: E0513 00:15:38.763128 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.763443 kubelet[2230]: E0513 00:15:38.763345 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.763443 kubelet[2230]: W0513 00:15:38.763356 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.763443 kubelet[2230]: E0513 00:15:38.763383 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.763698 kubelet[2230]: E0513 00:15:38.763615 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.763698 kubelet[2230]: W0513 00:15:38.763627 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.763698 kubelet[2230]: E0513 00:15:38.763648 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.764095 kubelet[2230]: E0513 00:15:38.764014 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.764095 kubelet[2230]: W0513 00:15:38.764027 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.764095 kubelet[2230]: E0513 00:15:38.764049 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.764437 kubelet[2230]: E0513 00:15:38.764350 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.764437 kubelet[2230]: W0513 00:15:38.764362 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.764437 kubelet[2230]: E0513 00:15:38.764394 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.764724 kubelet[2230]: E0513 00:15:38.764638 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.764724 kubelet[2230]: W0513 00:15:38.764650 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.764724 kubelet[2230]: E0513 00:15:38.764671 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.764962 kubelet[2230]: E0513 00:15:38.764882 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.764962 kubelet[2230]: W0513 00:15:38.764893 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.764962 kubelet[2230]: E0513 00:15:38.764915 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.765696 kubelet[2230]: E0513 00:15:38.765597 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.765696 kubelet[2230]: W0513 00:15:38.765610 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.765696 kubelet[2230]: E0513 00:15:38.765664 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.765963 kubelet[2230]: E0513 00:15:38.765876 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.765963 kubelet[2230]: W0513 00:15:38.765888 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.765963 kubelet[2230]: E0513 00:15:38.765912 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.766194 kubelet[2230]: E0513 00:15:38.766112 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.766194 kubelet[2230]: W0513 00:15:38.766124 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.766194 kubelet[2230]: E0513 00:15:38.766143 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.766438 kubelet[2230]: E0513 00:15:38.766344 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.766438 kubelet[2230]: W0513 00:15:38.766355 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.766438 kubelet[2230]: E0513 00:15:38.766376 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.766739 kubelet[2230]: E0513 00:15:38.766629 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.766739 kubelet[2230]: W0513 00:15:38.766641 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.766739 kubelet[2230]: E0513 00:15:38.766664 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.767003 kubelet[2230]: E0513 00:15:38.766892 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.767003 kubelet[2230]: W0513 00:15:38.766905 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.767003 kubelet[2230]: E0513 00:15:38.766920 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.767567 kubelet[2230]: E0513 00:15:38.767465 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.767567 kubelet[2230]: W0513 00:15:38.767481 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.767567 kubelet[2230]: E0513 00:15:38.767499 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.769022 kubelet[2230]: E0513 00:15:38.768999 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.770000 audit[2751]: NETFILTER_CFG table=filter:93 family=2 entries=17 op=nft_register_rule pid=2751 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:15:38.772521 kubelet[2230]: W0513 00:15:38.772494 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.772927 kubelet[2230]: E0513 00:15:38.772901 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.773893 kubelet[2230]: E0513 00:15:38.773877 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.773893 kubelet[2230]: W0513 00:15:38.773895 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.774067 kubelet[2230]: E0513 00:15:38.774054 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.774067 kubelet[2230]: W0513 00:15:38.774061 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.774143 kernel: kauditd_printk_skb: 155 callbacks suppressed May 13 00:15:38.774197 kernel: audit: type=1325 audit(1747095338.770:292): table=filter:93 family=2 entries=17 op=nft_register_rule pid=2751 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:15:38.774230 kernel: audit: type=1300 audit(1747095338.770:292): arch=c00000b7 syscall=211 success=yes exit=6652 a0=3 a1=ffffec2807e0 a2=0 a3=1 items=0 ppid=2393 pid=2751 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:38.770000 audit[2751]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=6652 a0=3 a1=ffffec2807e0 a2=0 a3=1 items=0 ppid=2393 pid=2751 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:38.774397 kubelet[2230]: E0513 00:15:38.774147 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.774397 kubelet[2230]: E0513 00:15:38.774202 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.774397 kubelet[2230]: W0513 00:15:38.774208 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.774397 kubelet[2230]: E0513 00:15:38.774221 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.774506 kubelet[2230]: E0513 00:15:38.774407 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.774506 kubelet[2230]: W0513 00:15:38.774423 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.774506 kubelet[2230]: E0513 00:15:38.774435 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.774613 kubelet[2230]: E0513 00:15:38.774594 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.774613 kubelet[2230]: W0513 00:15:38.774608 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.774680 kubelet[2230]: E0513 00:15:38.774617 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.774738 kubelet[2230]: E0513 00:15:38.774719 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.770000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:38.779919 kernel: audit: type=1327 audit(1747095338.770:292): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:38.784000 audit[2751]: NETFILTER_CFG table=nat:94 family=2 entries=12 op=nft_register_rule pid=2751 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:15:38.784000 audit[2751]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2700 a0=3 a1=ffffec2807e0 a2=0 a3=1 items=0 ppid=2393 pid=2751 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:38.792615 kernel: audit: type=1325 audit(1747095338.784:293): table=nat:94 family=2 entries=12 op=nft_register_rule pid=2751 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:15:38.792811 kernel: audit: type=1300 audit(1747095338.784:293): arch=c00000b7 syscall=211 success=yes exit=2700 a0=3 a1=ffffec2807e0 a2=0 a3=1 items=0 ppid=2393 pid=2751 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:38.792851 kubelet[2230]: E0513 00:15:38.788783 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:38.792851 kubelet[2230]: W0513 00:15:38.788801 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:38.792851 kubelet[2230]: E0513 00:15:38.788818 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:38.784000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:38.795393 kernel: audit: type=1327 audit(1747095338.784:293): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:38.808546 env[1323]: time="2025-05-13T00:15:38.808462166Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 13 00:15:38.808717 env[1323]: time="2025-05-13T00:15:38.808565327Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 13 00:15:38.808717 env[1323]: time="2025-05-13T00:15:38.808596927Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 13 00:15:38.808890 env[1323]: time="2025-05-13T00:15:38.808851807Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/37519f84ba282eff3dde55230441a8e793daa5adcee09704f8ec4e26b62acb33 pid=2766 runtime=io.containerd.runc.v2 May 13 00:15:38.852712 env[1323]: time="2025-05-13T00:15:38.852612340Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-bsdv5,Uid:3aef28fa-07db-410f-a490-9ff3ec65460f,Namespace:calico-system,Attempt:0,} returns sandbox id \"37519f84ba282eff3dde55230441a8e793daa5adcee09704f8ec4e26b62acb33\"" May 13 00:15:38.856198 kubelet[2230]: E0513 00:15:38.856132 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:40.130014 kubelet[2230]: E0513 00:15:40.129970 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-srb86" podUID="838e7283-9b17-4532-a9f4-4b17f2b5ca3b" May 13 00:15:40.257527 env[1323]: time="2025-05-13T00:15:40.257479877Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/typha:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:40.258681 env[1323]: time="2025-05-13T00:15:40.258650398Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:26e730979a07ea7452715da6ac48076016018bc982c06ebd32d5e095f42d3d54,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:40.260178 env[1323]: time="2025-05-13T00:15:40.260144879Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/typha:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:40.261432 env[1323]: time="2025-05-13T00:15:40.261410081Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:40.261818 env[1323]: time="2025-05-13T00:15:40.261790401Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\" returns image reference \"sha256:26e730979a07ea7452715da6ac48076016018bc982c06ebd32d5e095f42d3d54\"" May 13 00:15:40.273507 env[1323]: time="2025-05-13T00:15:40.273472493Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\"" May 13 00:15:40.287021 env[1323]: time="2025-05-13T00:15:40.286963708Z" level=info msg="CreateContainer within sandbox \"d60b9b69d98c5b3f8a5412b548c794bf90fb66766586bdf4ae59348be85dcbd8\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" May 13 00:15:40.302168 env[1323]: time="2025-05-13T00:15:40.302124084Z" level=info msg="CreateContainer within sandbox \"d60b9b69d98c5b3f8a5412b548c794bf90fb66766586bdf4ae59348be85dcbd8\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"806ff70fac795366b20072a5f8cd0a396560fd5c57d48f1ac0e69480ab6a6462\"" May 13 00:15:40.305488 env[1323]: time="2025-05-13T00:15:40.305458007Z" level=info msg="StartContainer for \"806ff70fac795366b20072a5f8cd0a396560fd5c57d48f1ac0e69480ab6a6462\"" May 13 00:15:40.386913 env[1323]: time="2025-05-13T00:15:40.386791813Z" level=info msg="StartContainer for \"806ff70fac795366b20072a5f8cd0a396560fd5c57d48f1ac0e69480ab6a6462\" returns successfully" May 13 00:15:41.136299 env[1323]: time="2025-05-13T00:15:41.136252037Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:41.137691 env[1323]: time="2025-05-13T00:15:41.137658398Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:dd8e710a588cc6f5834c4d84f7e12458efae593d3dfe527ca9e757c89239ecb8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:41.139403 env[1323]: time="2025-05-13T00:15:41.139361800Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:41.140984 env[1323]: time="2025-05-13T00:15:41.140951161Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:41.141355 env[1323]: time="2025-05-13T00:15:41.141318282Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" returns image reference \"sha256:dd8e710a588cc6f5834c4d84f7e12458efae593d3dfe527ca9e757c89239ecb8\"" May 13 00:15:41.143575 env[1323]: time="2025-05-13T00:15:41.143526524Z" level=info msg="CreateContainer within sandbox \"37519f84ba282eff3dde55230441a8e793daa5adcee09704f8ec4e26b62acb33\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" May 13 00:15:41.155915 env[1323]: time="2025-05-13T00:15:41.155868936Z" level=info msg="CreateContainer within sandbox \"37519f84ba282eff3dde55230441a8e793daa5adcee09704f8ec4e26b62acb33\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"a26e00899ff93af04beb4e01769b0f74e280179de1c42c028e8715f78a2758b1\"" May 13 00:15:41.157209 env[1323]: time="2025-05-13T00:15:41.156827337Z" level=info msg="StartContainer for \"a26e00899ff93af04beb4e01769b0f74e280179de1c42c028e8715f78a2758b1\"" May 13 00:15:41.198263 kubelet[2230]: E0513 00:15:41.198235 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:41.220961 kubelet[2230]: E0513 00:15:41.220936 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.220961 kubelet[2230]: W0513 00:15:41.220956 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.220961 kubelet[2230]: E0513 00:15:41.220973 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.221172 kubelet[2230]: E0513 00:15:41.221158 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.221172 kubelet[2230]: W0513 00:15:41.221168 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.221239 kubelet[2230]: E0513 00:15:41.221177 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.221332 kubelet[2230]: E0513 00:15:41.221320 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.221332 kubelet[2230]: W0513 00:15:41.221331 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.221413 kubelet[2230]: E0513 00:15:41.221340 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.221509 kubelet[2230]: E0513 00:15:41.221497 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.221567 kubelet[2230]: W0513 00:15:41.221512 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.221567 kubelet[2230]: E0513 00:15:41.221520 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.222088 kubelet[2230]: E0513 00:15:41.222073 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.222088 kubelet[2230]: W0513 00:15:41.222087 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.222181 kubelet[2230]: E0513 00:15:41.222097 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.223132 kubelet[2230]: E0513 00:15:41.223116 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.223132 kubelet[2230]: W0513 00:15:41.223129 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.223241 kubelet[2230]: E0513 00:15:41.223140 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.223461 kubelet[2230]: E0513 00:15:41.223440 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.223461 kubelet[2230]: W0513 00:15:41.223454 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.223556 kubelet[2230]: E0513 00:15:41.223464 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.223661 kubelet[2230]: E0513 00:15:41.223648 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.223661 kubelet[2230]: W0513 00:15:41.223660 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.223738 kubelet[2230]: E0513 00:15:41.223668 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.223867 kubelet[2230]: E0513 00:15:41.223853 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.223867 kubelet[2230]: W0513 00:15:41.223865 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.223950 kubelet[2230]: E0513 00:15:41.223876 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.224039 kubelet[2230]: E0513 00:15:41.224027 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.224039 kubelet[2230]: W0513 00:15:41.224037 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.224094 kubelet[2230]: E0513 00:15:41.224045 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.224176 kubelet[2230]: E0513 00:15:41.224165 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.224176 kubelet[2230]: W0513 00:15:41.224174 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.224226 kubelet[2230]: E0513 00:15:41.224181 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.224316 kubelet[2230]: E0513 00:15:41.224306 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.224341 kubelet[2230]: W0513 00:15:41.224316 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.224341 kubelet[2230]: E0513 00:15:41.224323 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.224487 kubelet[2230]: E0513 00:15:41.224476 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.224487 kubelet[2230]: W0513 00:15:41.224487 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.224586 kubelet[2230]: E0513 00:15:41.224495 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.224668 kubelet[2230]: E0513 00:15:41.224656 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.224696 kubelet[2230]: W0513 00:15:41.224668 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.224696 kubelet[2230]: E0513 00:15:41.224677 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.224805 kubelet[2230]: E0513 00:15:41.224796 2230 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 13 00:15:41.224833 kubelet[2230]: W0513 00:15:41.224806 2230 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 13 00:15:41.224833 kubelet[2230]: E0513 00:15:41.224812 2230 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 13 00:15:41.234331 env[1323]: time="2025-05-13T00:15:41.234289334Z" level=info msg="StartContainer for \"a26e00899ff93af04beb4e01769b0f74e280179de1c42c028e8715f78a2758b1\" returns successfully" May 13 00:15:41.296935 env[1323]: time="2025-05-13T00:15:41.296885116Z" level=info msg="shim disconnected" id=a26e00899ff93af04beb4e01769b0f74e280179de1c42c028e8715f78a2758b1 May 13 00:15:41.296935 env[1323]: time="2025-05-13T00:15:41.296930276Z" level=warning msg="cleaning up after shim disconnected" id=a26e00899ff93af04beb4e01769b0f74e280179de1c42c028e8715f78a2758b1 namespace=k8s.io May 13 00:15:41.296935 env[1323]: time="2025-05-13T00:15:41.296940236Z" level=info msg="cleaning up dead shim" May 13 00:15:41.304202 env[1323]: time="2025-05-13T00:15:41.304156243Z" level=warning msg="cleanup warnings time=\"2025-05-13T00:15:41Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2909 runtime=io.containerd.runc.v2\n" May 13 00:15:41.566703 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a26e00899ff93af04beb4e01769b0f74e280179de1c42c028e8715f78a2758b1-rootfs.mount: Deactivated successfully. May 13 00:15:42.130823 kubelet[2230]: E0513 00:15:42.130757 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-srb86" podUID="838e7283-9b17-4532-a9f4-4b17f2b5ca3b" May 13 00:15:42.201035 kubelet[2230]: I0513 00:15:42.200834 2230 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 13 00:15:42.201403 kubelet[2230]: E0513 00:15:42.201113 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:42.201472 kubelet[2230]: E0513 00:15:42.201449 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:42.202864 env[1323]: time="2025-05-13T00:15:42.202830281Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\"" May 13 00:15:42.218430 kubelet[2230]: I0513 00:15:42.218357 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-85f4d64f96-sll44" podStartSLOduration=2.652002126 podStartE2EDuration="4.218342415s" podCreationTimestamp="2025-05-13 00:15:38 +0000 UTC" firstStartedPulling="2025-05-13 00:15:38.705858483 +0000 UTC m=+22.684081231" lastFinishedPulling="2025-05-13 00:15:40.272198772 +0000 UTC m=+24.250421520" observedRunningTime="2025-05-13 00:15:41.214408874 +0000 UTC m=+25.192631622" watchObservedRunningTime="2025-05-13 00:15:42.218342415 +0000 UTC m=+26.196565163" May 13 00:15:44.130878 kubelet[2230]: E0513 00:15:44.130824 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-srb86" podUID="838e7283-9b17-4532-a9f4-4b17f2b5ca3b" May 13 00:15:45.009916 env[1323]: time="2025-05-13T00:15:45.009821917Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/cni:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:45.016163 env[1323]: time="2025-05-13T00:15:45.016120002Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:add6372545fb406bb017769f222d84c50549ce13e3b19f1fbaee3d8a4aaef627,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:45.019060 env[1323]: time="2025-05-13T00:15:45.019025044Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/cni:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:45.031194 env[1323]: time="2025-05-13T00:15:45.031133973Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:45.031938 env[1323]: time="2025-05-13T00:15:45.031898214Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\" returns image reference \"sha256:add6372545fb406bb017769f222d84c50549ce13e3b19f1fbaee3d8a4aaef627\"" May 13 00:15:45.036122 env[1323]: time="2025-05-13T00:15:45.036081017Z" level=info msg="CreateContainer within sandbox \"37519f84ba282eff3dde55230441a8e793daa5adcee09704f8ec4e26b62acb33\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" May 13 00:15:45.053062 env[1323]: time="2025-05-13T00:15:45.052995390Z" level=info msg="CreateContainer within sandbox \"37519f84ba282eff3dde55230441a8e793daa5adcee09704f8ec4e26b62acb33\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"462eb01974f067c6db86868ea8191a18420bc7e2bd1b84e6118ac041b73a7860\"" May 13 00:15:45.054722 env[1323]: time="2025-05-13T00:15:45.054671591Z" level=info msg="StartContainer for \"462eb01974f067c6db86868ea8191a18420bc7e2bd1b84e6118ac041b73a7860\"" May 13 00:15:45.068000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.8:22-10.0.0.1:38146 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:45.069165 systemd[1]: Started sshd@7-10.0.0.8:22-10.0.0.1:38146.service. May 13 00:15:45.072565 kernel: audit: type=1130 audit(1747095345.068:294): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.8:22-10.0.0.1:38146 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:45.088081 systemd[1]: run-containerd-runc-k8s.io-462eb01974f067c6db86868ea8191a18420bc7e2bd1b84e6118ac041b73a7860-runc.cFdNYb.mount: Deactivated successfully. May 13 00:15:45.159000 audit[2939]: USER_ACCT pid=2939 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:45.160318 sshd[2939]: Accepted publickey for core from 10.0.0.1 port 38146 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:15:45.169978 kernel: audit: type=1101 audit(1747095345.159:295): pid=2939 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:45.170069 kernel: audit: type=1103 audit(1747095345.163:296): pid=2939 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:45.170120 kernel: audit: type=1006 audit(1747095345.163:297): pid=2939 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=8 res=1 May 13 00:15:45.163000 audit[2939]: CRED_ACQ pid=2939 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:45.168878 sshd[2939]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:15:45.170603 kernel: audit: type=1300 audit(1747095345.163:297): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffcdaa33c0 a2=3 a3=1 items=0 ppid=1 pid=2939 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=8 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:45.163000 audit[2939]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffcdaa33c0 a2=3 a3=1 items=0 ppid=1 pid=2939 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=8 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:45.179190 kernel: audit: type=1327 audit(1747095345.163:297): proctitle=737368643A20636F7265205B707269765D May 13 00:15:45.163000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:15:45.203978 systemd-logind[1306]: New session 8 of user core. May 13 00:15:45.204072 systemd[1]: Started session-8.scope. May 13 00:15:45.213000 audit[2939]: USER_START pid=2939 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:45.218555 kernel: audit: type=1105 audit(1747095345.213:298): pid=2939 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:45.221790 kernel: audit: type=1103 audit(1747095345.218:299): pid=2953 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:45.218000 audit[2953]: CRED_ACQ pid=2953 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:45.374546 env[1323]: time="2025-05-13T00:15:45.374438996Z" level=info msg="StartContainer for \"462eb01974f067c6db86868ea8191a18420bc7e2bd1b84e6118ac041b73a7860\" returns successfully" May 13 00:15:45.396883 sshd[2939]: pam_unix(sshd:session): session closed for user core May 13 00:15:45.397000 audit[2939]: USER_END pid=2939 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:45.400324 systemd[1]: sshd@7-10.0.0.8:22-10.0.0.1:38146.service: Deactivated successfully. May 13 00:15:45.401113 systemd[1]: session-8.scope: Deactivated successfully. May 13 00:15:45.397000 audit[2939]: CRED_DISP pid=2939 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:45.404691 kernel: audit: type=1106 audit(1747095345.397:300): pid=2939 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:45.404776 kernel: audit: type=1104 audit(1747095345.397:301): pid=2939 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:45.404758 systemd-logind[1306]: Session 8 logged out. Waiting for processes to exit. May 13 00:15:45.400000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.8:22-10.0.0.1:38146 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:45.405683 systemd-logind[1306]: Removed session 8. May 13 00:15:45.922114 env[1323]: time="2025-05-13T00:15:45.922044335Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/calico-kubeconfig\": WRITE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 13 00:15:45.947997 env[1323]: time="2025-05-13T00:15:45.947953875Z" level=info msg="shim disconnected" id=462eb01974f067c6db86868ea8191a18420bc7e2bd1b84e6118ac041b73a7860 May 13 00:15:45.947997 env[1323]: time="2025-05-13T00:15:45.947997475Z" level=warning msg="cleaning up after shim disconnected" id=462eb01974f067c6db86868ea8191a18420bc7e2bd1b84e6118ac041b73a7860 namespace=k8s.io May 13 00:15:45.948211 env[1323]: time="2025-05-13T00:15:45.948006875Z" level=info msg="cleaning up dead shim" May 13 00:15:45.954416 env[1323]: time="2025-05-13T00:15:45.954346200Z" level=warning msg="cleanup warnings time=\"2025-05-13T00:15:45Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2993 runtime=io.containerd.runc.v2\n" May 13 00:15:45.994608 kubelet[2230]: I0513 00:15:45.994449 2230 kubelet_node_status.go:497] "Fast updating node status as it just became ready" May 13 00:15:46.016622 kubelet[2230]: I0513 00:15:46.015233 2230 topology_manager.go:215] "Topology Admit Handler" podUID="8dde30de-095c-474d-a07e-44b258f6e953" podNamespace="kube-system" podName="coredns-7db6d8ff4d-ncf86" May 13 00:15:46.016622 kubelet[2230]: I0513 00:15:46.015956 2230 topology_manager.go:215] "Topology Admit Handler" podUID="943c0c0f-6fce-4911-b87f-c827598e67dc" podNamespace="kube-system" podName="coredns-7db6d8ff4d-k77ll" May 13 00:15:46.018186 kubelet[2230]: I0513 00:15:46.017040 2230 topology_manager.go:215] "Topology Admit Handler" podUID="47f5ba25-f173-43e0-af82-849feb0085ec" podNamespace="calico-system" podName="calico-kube-controllers-66bcdc688d-zm28k" May 13 00:15:46.018186 kubelet[2230]: I0513 00:15:46.017387 2230 topology_manager.go:215] "Topology Admit Handler" podUID="4596526d-58fb-4ef8-81dc-275ff89a1ce4" podNamespace="calico-apiserver" podName="calico-apiserver-5b58ff46dd-56crg" May 13 00:15:46.018186 kubelet[2230]: I0513 00:15:46.017751 2230 topology_manager.go:215] "Topology Admit Handler" podUID="4a3e9432-3790-4f08-9c11-d537f6aab7dd" podNamespace="calico-apiserver" podName="calico-apiserver-5b58ff46dd-zk6tg" May 13 00:15:46.050578 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-462eb01974f067c6db86868ea8191a18420bc7e2bd1b84e6118ac041b73a7860-rootfs.mount: Deactivated successfully. May 13 00:15:46.130567 kubelet[2230]: I0513 00:15:46.130502 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx7fm\" (UniqueName: \"kubernetes.io/projected/47f5ba25-f173-43e0-af82-849feb0085ec-kube-api-access-kx7fm\") pod \"calico-kube-controllers-66bcdc688d-zm28k\" (UID: \"47f5ba25-f173-43e0-af82-849feb0085ec\") " pod="calico-system/calico-kube-controllers-66bcdc688d-zm28k" May 13 00:15:46.130567 kubelet[2230]: I0513 00:15:46.130567 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47f5ba25-f173-43e0-af82-849feb0085ec-tigera-ca-bundle\") pod \"calico-kube-controllers-66bcdc688d-zm28k\" (UID: \"47f5ba25-f173-43e0-af82-849feb0085ec\") " pod="calico-system/calico-kube-controllers-66bcdc688d-zm28k" May 13 00:15:46.130744 kubelet[2230]: I0513 00:15:46.130591 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m474r\" (UniqueName: \"kubernetes.io/projected/943c0c0f-6fce-4911-b87f-c827598e67dc-kube-api-access-m474r\") pod \"coredns-7db6d8ff4d-k77ll\" (UID: \"943c0c0f-6fce-4911-b87f-c827598e67dc\") " pod="kube-system/coredns-7db6d8ff4d-k77ll" May 13 00:15:46.130744 kubelet[2230]: I0513 00:15:46.130610 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnccv\" (UniqueName: \"kubernetes.io/projected/4a3e9432-3790-4f08-9c11-d537f6aab7dd-kube-api-access-qnccv\") pod \"calico-apiserver-5b58ff46dd-zk6tg\" (UID: \"4a3e9432-3790-4f08-9c11-d537f6aab7dd\") " pod="calico-apiserver/calico-apiserver-5b58ff46dd-zk6tg" May 13 00:15:46.130744 kubelet[2230]: I0513 00:15:46.130637 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzwft\" (UniqueName: \"kubernetes.io/projected/4596526d-58fb-4ef8-81dc-275ff89a1ce4-kube-api-access-mzwft\") pod \"calico-apiserver-5b58ff46dd-56crg\" (UID: \"4596526d-58fb-4ef8-81dc-275ff89a1ce4\") " pod="calico-apiserver/calico-apiserver-5b58ff46dd-56crg" May 13 00:15:46.130744 kubelet[2230]: I0513 00:15:46.130657 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/943c0c0f-6fce-4911-b87f-c827598e67dc-config-volume\") pod \"coredns-7db6d8ff4d-k77ll\" (UID: \"943c0c0f-6fce-4911-b87f-c827598e67dc\") " pod="kube-system/coredns-7db6d8ff4d-k77ll" May 13 00:15:46.130744 kubelet[2230]: I0513 00:15:46.130676 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/4596526d-58fb-4ef8-81dc-275ff89a1ce4-calico-apiserver-certs\") pod \"calico-apiserver-5b58ff46dd-56crg\" (UID: \"4596526d-58fb-4ef8-81dc-275ff89a1ce4\") " pod="calico-apiserver/calico-apiserver-5b58ff46dd-56crg" May 13 00:15:46.130890 kubelet[2230]: I0513 00:15:46.130734 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8dde30de-095c-474d-a07e-44b258f6e953-config-volume\") pod \"coredns-7db6d8ff4d-ncf86\" (UID: \"8dde30de-095c-474d-a07e-44b258f6e953\") " pod="kube-system/coredns-7db6d8ff4d-ncf86" May 13 00:15:46.130890 kubelet[2230]: I0513 00:15:46.130754 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w5vv\" (UniqueName: \"kubernetes.io/projected/8dde30de-095c-474d-a07e-44b258f6e953-kube-api-access-9w5vv\") pod \"coredns-7db6d8ff4d-ncf86\" (UID: \"8dde30de-095c-474d-a07e-44b258f6e953\") " pod="kube-system/coredns-7db6d8ff4d-ncf86" May 13 00:15:46.130890 kubelet[2230]: I0513 00:15:46.130772 2230 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/4a3e9432-3790-4f08-9c11-d537f6aab7dd-calico-apiserver-certs\") pod \"calico-apiserver-5b58ff46dd-zk6tg\" (UID: \"4a3e9432-3790-4f08-9c11-d537f6aab7dd\") " pod="calico-apiserver/calico-apiserver-5b58ff46dd-zk6tg" May 13 00:15:46.132266 env[1323]: time="2025-05-13T00:15:46.132211010Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-srb86,Uid:838e7283-9b17-4532-a9f4-4b17f2b5ca3b,Namespace:calico-system,Attempt:0,}" May 13 00:15:46.211833 kubelet[2230]: E0513 00:15:46.211653 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:46.215936 env[1323]: time="2025-05-13T00:15:46.215778630Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\"" May 13 00:15:46.330236 kubelet[2230]: E0513 00:15:46.329920 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:46.331767 env[1323]: time="2025-05-13T00:15:46.330721312Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-66bcdc688d-zm28k,Uid:47f5ba25-f173-43e0-af82-849feb0085ec,Namespace:calico-system,Attempt:0,}" May 13 00:15:46.331767 env[1323]: time="2025-05-13T00:15:46.330743192Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-ncf86,Uid:8dde30de-095c-474d-a07e-44b258f6e953,Namespace:kube-system,Attempt:0,}" May 13 00:15:46.332496 kubelet[2230]: E0513 00:15:46.332241 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:46.333842 env[1323]: time="2025-05-13T00:15:46.333643554Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-k77ll,Uid:943c0c0f-6fce-4911-b87f-c827598e67dc,Namespace:kube-system,Attempt:0,}" May 13 00:15:46.336292 env[1323]: time="2025-05-13T00:15:46.336263236Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5b58ff46dd-zk6tg,Uid:4a3e9432-3790-4f08-9c11-d537f6aab7dd,Namespace:calico-apiserver,Attempt:0,}" May 13 00:15:46.337800 env[1323]: time="2025-05-13T00:15:46.337769077Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5b58ff46dd-56crg,Uid:4596526d-58fb-4ef8-81dc-275ff89a1ce4,Namespace:calico-apiserver,Attempt:0,}" May 13 00:15:46.359922 env[1323]: time="2025-05-13T00:15:46.359848773Z" level=error msg="Failed to destroy network for sandbox \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.360275 env[1323]: time="2025-05-13T00:15:46.360234533Z" level=error msg="encountered an error cleaning up failed sandbox \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.360341 env[1323]: time="2025-05-13T00:15:46.360286973Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-srb86,Uid:838e7283-9b17-4532-a9f4-4b17f2b5ca3b,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.361952 kubelet[2230]: E0513 00:15:46.361494 2230 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.361952 kubelet[2230]: E0513 00:15:46.361579 2230 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-srb86" May 13 00:15:46.361952 kubelet[2230]: E0513 00:15:46.361601 2230 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-srb86" May 13 00:15:46.362180 kubelet[2230]: E0513 00:15:46.361656 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-srb86_calico-system(838e7283-9b17-4532-a9f4-4b17f2b5ca3b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-srb86_calico-system(838e7283-9b17-4532-a9f4-4b17f2b5ca3b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-srb86" podUID="838e7283-9b17-4532-a9f4-4b17f2b5ca3b" May 13 00:15:46.437174 env[1323]: time="2025-05-13T00:15:46.437108189Z" level=error msg="Failed to destroy network for sandbox \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.437570 env[1323]: time="2025-05-13T00:15:46.437519949Z" level=error msg="encountered an error cleaning up failed sandbox \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.437643 env[1323]: time="2025-05-13T00:15:46.437615869Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-ncf86,Uid:8dde30de-095c-474d-a07e-44b258f6e953,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.438256 kubelet[2230]: E0513 00:15:46.437862 2230 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.438256 kubelet[2230]: E0513 00:15:46.437939 2230 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-ncf86" May 13 00:15:46.438256 kubelet[2230]: E0513 00:15:46.437959 2230 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-ncf86" May 13 00:15:46.439734 kubelet[2230]: E0513 00:15:46.438010 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-ncf86_kube-system(8dde30de-095c-474d-a07e-44b258f6e953)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-ncf86_kube-system(8dde30de-095c-474d-a07e-44b258f6e953)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-ncf86" podUID="8dde30de-095c-474d-a07e-44b258f6e953" May 13 00:15:46.457420 env[1323]: time="2025-05-13T00:15:46.457344443Z" level=error msg="Failed to destroy network for sandbox \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.459224 env[1323]: time="2025-05-13T00:15:46.459178204Z" level=error msg="encountered an error cleaning up failed sandbox \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.459308 env[1323]: time="2025-05-13T00:15:46.459237844Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-66bcdc688d-zm28k,Uid:47f5ba25-f173-43e0-af82-849feb0085ec,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.459729 kubelet[2230]: E0513 00:15:46.459555 2230 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.459729 kubelet[2230]: E0513 00:15:46.459611 2230 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-66bcdc688d-zm28k" May 13 00:15:46.459729 kubelet[2230]: E0513 00:15:46.459636 2230 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-66bcdc688d-zm28k" May 13 00:15:46.459870 kubelet[2230]: E0513 00:15:46.459678 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-66bcdc688d-zm28k_calico-system(47f5ba25-f173-43e0-af82-849feb0085ec)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-66bcdc688d-zm28k_calico-system(47f5ba25-f173-43e0-af82-849feb0085ec)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-66bcdc688d-zm28k" podUID="47f5ba25-f173-43e0-af82-849feb0085ec" May 13 00:15:46.462850 env[1323]: time="2025-05-13T00:15:46.462755807Z" level=error msg="Failed to destroy network for sandbox \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.463945 env[1323]: time="2025-05-13T00:15:46.463897128Z" level=error msg="encountered an error cleaning up failed sandbox \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.464026 env[1323]: time="2025-05-13T00:15:46.463952208Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-k77ll,Uid:943c0c0f-6fce-4911-b87f-c827598e67dc,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.464194 kubelet[2230]: E0513 00:15:46.464148 2230 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.464260 kubelet[2230]: E0513 00:15:46.464211 2230 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-k77ll" May 13 00:15:46.464260 kubelet[2230]: E0513 00:15:46.464230 2230 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-k77ll" May 13 00:15:46.464324 kubelet[2230]: E0513 00:15:46.464269 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-k77ll_kube-system(943c0c0f-6fce-4911-b87f-c827598e67dc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-k77ll_kube-system(943c0c0f-6fce-4911-b87f-c827598e67dc)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-k77ll" podUID="943c0c0f-6fce-4911-b87f-c827598e67dc" May 13 00:15:46.474116 env[1323]: time="2025-05-13T00:15:46.474061935Z" level=error msg="Failed to destroy network for sandbox \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.474482 env[1323]: time="2025-05-13T00:15:46.474447775Z" level=error msg="encountered an error cleaning up failed sandbox \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.474545 env[1323]: time="2025-05-13T00:15:46.474505855Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5b58ff46dd-zk6tg,Uid:4a3e9432-3790-4f08-9c11-d537f6aab7dd,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.474757 kubelet[2230]: E0513 00:15:46.474720 2230 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.474818 kubelet[2230]: E0513 00:15:46.474779 2230 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5b58ff46dd-zk6tg" May 13 00:15:46.474818 kubelet[2230]: E0513 00:15:46.474800 2230 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5b58ff46dd-zk6tg" May 13 00:15:46.474883 kubelet[2230]: E0513 00:15:46.474838 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-5b58ff46dd-zk6tg_calico-apiserver(4a3e9432-3790-4f08-9c11-d537f6aab7dd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-5b58ff46dd-zk6tg_calico-apiserver(4a3e9432-3790-4f08-9c11-d537f6aab7dd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-5b58ff46dd-zk6tg" podUID="4a3e9432-3790-4f08-9c11-d537f6aab7dd" May 13 00:15:46.474938 env[1323]: time="2025-05-13T00:15:46.474570895Z" level=error msg="Failed to destroy network for sandbox \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.476254 env[1323]: time="2025-05-13T00:15:46.475176856Z" level=error msg="encountered an error cleaning up failed sandbox \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.476254 env[1323]: time="2025-05-13T00:15:46.475224296Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5b58ff46dd-56crg,Uid:4596526d-58fb-4ef8-81dc-275ff89a1ce4,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.476438 kubelet[2230]: E0513 00:15:46.475351 2230 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:46.476438 kubelet[2230]: E0513 00:15:46.475396 2230 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5b58ff46dd-56crg" May 13 00:15:46.476438 kubelet[2230]: E0513 00:15:46.475471 2230 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5b58ff46dd-56crg" May 13 00:15:46.476527 kubelet[2230]: E0513 00:15:46.475523 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-5b58ff46dd-56crg_calico-apiserver(4596526d-58fb-4ef8-81dc-275ff89a1ce4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-5b58ff46dd-56crg_calico-apiserver(4596526d-58fb-4ef8-81dc-275ff89a1ce4)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-5b58ff46dd-56crg" podUID="4596526d-58fb-4ef8-81dc-275ff89a1ce4" May 13 00:15:47.055347 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e-shm.mount: Deactivated successfully. May 13 00:15:47.219796 kubelet[2230]: I0513 00:15:47.219751 2230 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" May 13 00:15:47.220885 kubelet[2230]: I0513 00:15:47.220582 2230 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" May 13 00:15:47.220992 env[1323]: time="2025-05-13T00:15:47.220954701Z" level=info msg="StopPodSandbox for \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\"" May 13 00:15:47.221232 env[1323]: time="2025-05-13T00:15:47.220457221Z" level=info msg="StopPodSandbox for \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\"" May 13 00:15:47.223008 kubelet[2230]: I0513 00:15:47.222633 2230 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" May 13 00:15:47.224047 kubelet[2230]: I0513 00:15:47.224019 2230 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" May 13 00:15:47.224560 env[1323]: time="2025-05-13T00:15:47.224494144Z" level=info msg="StopPodSandbox for \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\"" May 13 00:15:47.224642 env[1323]: time="2025-05-13T00:15:47.224598184Z" level=info msg="StopPodSandbox for \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\"" May 13 00:15:47.245565 kubelet[2230]: I0513 00:15:47.239064 2230 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" May 13 00:15:47.246724 kubelet[2230]: I0513 00:15:47.246694 2230 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" May 13 00:15:47.247997 env[1323]: time="2025-05-13T00:15:47.247956759Z" level=info msg="StopPodSandbox for \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\"" May 13 00:15:47.248078 env[1323]: time="2025-05-13T00:15:47.247997759Z" level=info msg="StopPodSandbox for \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\"" May 13 00:15:47.287866 env[1323]: time="2025-05-13T00:15:47.287810946Z" level=error msg="StopPodSandbox for \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\" failed" error="failed to destroy network for sandbox \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:47.290320 kubelet[2230]: E0513 00:15:47.290245 2230 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" May 13 00:15:47.290469 kubelet[2230]: E0513 00:15:47.290337 2230 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e"} May 13 00:15:47.290469 kubelet[2230]: E0513 00:15:47.290427 2230 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"838e7283-9b17-4532-a9f4-4b17f2b5ca3b\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 13 00:15:47.290469 kubelet[2230]: E0513 00:15:47.290457 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"838e7283-9b17-4532-a9f4-4b17f2b5ca3b\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-srb86" podUID="838e7283-9b17-4532-a9f4-4b17f2b5ca3b" May 13 00:15:47.319489 env[1323]: time="2025-05-13T00:15:47.319355207Z" level=error msg="StopPodSandbox for \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\" failed" error="failed to destroy network for sandbox \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:47.319814 kubelet[2230]: E0513 00:15:47.319624 2230 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" May 13 00:15:47.319814 kubelet[2230]: E0513 00:15:47.319671 2230 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc"} May 13 00:15:47.319814 kubelet[2230]: E0513 00:15:47.319703 2230 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"4596526d-58fb-4ef8-81dc-275ff89a1ce4\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 13 00:15:47.319814 kubelet[2230]: E0513 00:15:47.319739 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"4596526d-58fb-4ef8-81dc-275ff89a1ce4\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-5b58ff46dd-56crg" podUID="4596526d-58fb-4ef8-81dc-275ff89a1ce4" May 13 00:15:47.323236 env[1323]: time="2025-05-13T00:15:47.323177890Z" level=error msg="StopPodSandbox for \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\" failed" error="failed to destroy network for sandbox \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:47.323598 kubelet[2230]: E0513 00:15:47.323553 2230 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" May 13 00:15:47.323682 kubelet[2230]: E0513 00:15:47.323612 2230 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543"} May 13 00:15:47.323682 kubelet[2230]: E0513 00:15:47.323650 2230 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"47f5ba25-f173-43e0-af82-849feb0085ec\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 13 00:15:47.323787 kubelet[2230]: E0513 00:15:47.323692 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"47f5ba25-f173-43e0-af82-849feb0085ec\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-66bcdc688d-zm28k" podUID="47f5ba25-f173-43e0-af82-849feb0085ec" May 13 00:15:47.330465 env[1323]: time="2025-05-13T00:15:47.330416335Z" level=error msg="StopPodSandbox for \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\" failed" error="failed to destroy network for sandbox \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:47.330834 kubelet[2230]: E0513 00:15:47.330794 2230 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" May 13 00:15:47.330921 kubelet[2230]: E0513 00:15:47.330847 2230 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2"} May 13 00:15:47.330921 kubelet[2230]: E0513 00:15:47.330879 2230 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"8dde30de-095c-474d-a07e-44b258f6e953\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 13 00:15:47.330921 kubelet[2230]: E0513 00:15:47.330899 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"8dde30de-095c-474d-a07e-44b258f6e953\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-ncf86" podUID="8dde30de-095c-474d-a07e-44b258f6e953" May 13 00:15:47.356072 env[1323]: time="2025-05-13T00:15:47.356018952Z" level=error msg="StopPodSandbox for \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\" failed" error="failed to destroy network for sandbox \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:47.356476 kubelet[2230]: E0513 00:15:47.356433 2230 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" May 13 00:15:47.356571 kubelet[2230]: E0513 00:15:47.356496 2230 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec"} May 13 00:15:47.356571 kubelet[2230]: E0513 00:15:47.356550 2230 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"4a3e9432-3790-4f08-9c11-d537f6aab7dd\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 13 00:15:47.356664 kubelet[2230]: E0513 00:15:47.356574 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"4a3e9432-3790-4f08-9c11-d537f6aab7dd\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-5b58ff46dd-zk6tg" podUID="4a3e9432-3790-4f08-9c11-d537f6aab7dd" May 13 00:15:47.404862 env[1323]: time="2025-05-13T00:15:47.404800105Z" level=error msg="StopPodSandbox for \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\" failed" error="failed to destroy network for sandbox \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 13 00:15:47.405142 kubelet[2230]: E0513 00:15:47.405077 2230 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" May 13 00:15:47.405213 kubelet[2230]: E0513 00:15:47.405145 2230 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f"} May 13 00:15:47.405213 kubelet[2230]: E0513 00:15:47.405178 2230 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"943c0c0f-6fce-4911-b87f-c827598e67dc\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 13 00:15:47.405213 kubelet[2230]: E0513 00:15:47.405199 2230 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"943c0c0f-6fce-4911-b87f-c827598e67dc\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-k77ll" podUID="943c0c0f-6fce-4911-b87f-c827598e67dc" May 13 00:15:50.399838 systemd[1]: Started sshd@8-10.0.0.8:22-10.0.0.1:38154.service. May 13 00:15:50.399000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.8:22-10.0.0.1:38154 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:50.402518 kernel: kauditd_printk_skb: 1 callbacks suppressed May 13 00:15:50.402672 kernel: audit: type=1130 audit(1747095350.399:303): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.8:22-10.0.0.1:38154 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:50.514524 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1392225538.mount: Deactivated successfully. May 13 00:15:50.587946 env[1323]: time="2025-05-13T00:15:50.584160811Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/node:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:50.587946 env[1323]: time="2025-05-13T00:15:50.587693213Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:cdcce3ec4624a24c28cdc07b0ee29ddf6703628edee7452a3f8a8b4816bfd057,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:50.591798 env[1323]: time="2025-05-13T00:15:50.591743776Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/node:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:50.593584 env[1323]: time="2025-05-13T00:15:50.593528137Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:15:50.593702 env[1323]: time="2025-05-13T00:15:50.593667377Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\" returns image reference \"sha256:cdcce3ec4624a24c28cdc07b0ee29ddf6703628edee7452a3f8a8b4816bfd057\"" May 13 00:15:50.612681 env[1323]: time="2025-05-13T00:15:50.612624867Z" level=info msg="CreateContainer within sandbox \"37519f84ba282eff3dde55230441a8e793daa5adcee09704f8ec4e26b62acb33\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" May 13 00:15:50.625183 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount460999320.mount: Deactivated successfully. May 13 00:15:50.629119 env[1323]: time="2025-05-13T00:15:50.629072036Z" level=info msg="CreateContainer within sandbox \"37519f84ba282eff3dde55230441a8e793daa5adcee09704f8ec4e26b62acb33\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"ca6ad9dd0ac7ae391156ed411c1cc63b4f1588009d4dd203ed8b0529b83ed992\"" May 13 00:15:50.629639 env[1323]: time="2025-05-13T00:15:50.629613077Z" level=info msg="StartContainer for \"ca6ad9dd0ac7ae391156ed411c1cc63b4f1588009d4dd203ed8b0529b83ed992\"" May 13 00:15:50.631000 audit[3386]: USER_ACCT pid=3386 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:50.635552 kernel: audit: type=1101 audit(1747095350.631:304): pid=3386 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:50.635654 sshd[3386]: Accepted publickey for core from 10.0.0.1 port 38154 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:15:50.636000 audit[3386]: CRED_ACQ pid=3386 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:50.640522 sshd[3386]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:15:50.642332 kernel: audit: type=1103 audit(1747095350.636:305): pid=3386 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:50.642413 kernel: audit: type=1006 audit(1747095350.636:306): pid=3386 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=9 res=1 May 13 00:15:50.642439 kernel: audit: type=1300 audit(1747095350.636:306): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffffba93d00 a2=3 a3=1 items=0 ppid=1 pid=3386 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=9 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:50.636000 audit[3386]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffffba93d00 a2=3 a3=1 items=0 ppid=1 pid=3386 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=9 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:50.636000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:15:50.647080 kernel: audit: type=1327 audit(1747095350.636:306): proctitle=737368643A20636F7265205B707269765D May 13 00:15:50.659345 systemd-logind[1306]: New session 9 of user core. May 13 00:15:50.660129 systemd[1]: Started session-9.scope. May 13 00:15:50.680000 audit[3386]: USER_START pid=3386 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:50.682000 audit[3414]: CRED_ACQ pid=3414 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:50.692011 kernel: audit: type=1105 audit(1747095350.680:307): pid=3386 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:50.692101 kernel: audit: type=1103 audit(1747095350.682:308): pid=3414 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:50.800070 env[1323]: time="2025-05-13T00:15:50.800014771Z" level=info msg="StartContainer for \"ca6ad9dd0ac7ae391156ed411c1cc63b4f1588009d4dd203ed8b0529b83ed992\" returns successfully" May 13 00:15:50.910025 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. May 13 00:15:50.910169 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. May 13 00:15:50.917746 sshd[3386]: pam_unix(sshd:session): session closed for user core May 13 00:15:50.918000 audit[3386]: USER_END pid=3386 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:50.921115 systemd[1]: sshd@8-10.0.0.8:22-10.0.0.1:38154.service: Deactivated successfully. May 13 00:15:50.921948 systemd[1]: session-9.scope: Deactivated successfully. May 13 00:15:50.918000 audit[3386]: CRED_DISP pid=3386 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:50.926392 kernel: audit: type=1106 audit(1747095350.918:309): pid=3386 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:50.926475 kernel: audit: type=1104 audit(1747095350.918:310): pid=3386 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:50.926475 systemd-logind[1306]: Session 9 logged out. Waiting for processes to exit. May 13 00:15:50.920000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.8:22-10.0.0.1:38154 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:50.927247 systemd-logind[1306]: Removed session 9. May 13 00:15:51.259300 kubelet[2230]: E0513 00:15:51.258875 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:52.236000 audit[3536]: AVC avc: denied { write } for pid=3536 comm="tee" name="fd" dev="proc" ino=20485 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 May 13 00:15:52.236000 audit[3536]: SYSCALL arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=fffff10d8a2a a2=241 a3=1b6 items=1 ppid=3506 pid=3536 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:52.236000 audit: CWD cwd="/etc/service/enabled/confd/log" May 13 00:15:52.236000 audit: PATH item=0 name="/dev/fd/63" inode=17408 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 13 00:15:52.236000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 May 13 00:15:52.243000 audit[3547]: AVC avc: denied { write } for pid=3547 comm="tee" name="fd" dev="proc" ino=18007 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 May 13 00:15:52.243000 audit[3547]: SYSCALL arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=ffffffd6aa2a a2=241 a3=1b6 items=1 ppid=3498 pid=3547 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:52.243000 audit: CWD cwd="/etc/service/enabled/bird6/log" May 13 00:15:52.243000 audit: PATH item=0 name="/dev/fd/63" inode=18002 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 13 00:15:52.243000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 May 13 00:15:52.260911 kubelet[2230]: E0513 00:15:52.260880 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:52.271000 audit[3556]: AVC avc: denied { write } for pid=3556 comm="tee" name="fd" dev="proc" ino=19944 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 May 13 00:15:52.271000 audit[3556]: SYSCALL arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=ffffd8b53a2b a2=241 a3=1b6 items=1 ppid=3500 pid=3556 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:52.271000 audit: CWD cwd="/etc/service/enabled/bird/log" May 13 00:15:52.271000 audit: PATH item=0 name="/dev/fd/63" inode=19929 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 13 00:15:52.271000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 May 13 00:15:52.273000 audit[3566]: AVC avc: denied { write } for pid=3566 comm="tee" name="fd" dev="proc" ino=19948 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 May 13 00:15:52.273000 audit[3566]: SYSCALL arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=ffffd744da1a a2=241 a3=1b6 items=1 ppid=3504 pid=3566 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:52.273000 audit: CWD cwd="/etc/service/enabled/allocate-tunnel-addrs/log" May 13 00:15:52.273000 audit: PATH item=0 name="/dev/fd/63" inode=19940 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 13 00:15:52.273000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 May 13 00:15:52.277000 audit[3572]: AVC avc: denied { write } for pid=3572 comm="tee" name="fd" dev="proc" ino=19955 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 May 13 00:15:52.277000 audit[3572]: SYSCALL arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=ffffd1cfaa1b a2=241 a3=1b6 items=1 ppid=3502 pid=3572 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:52.277000 audit: CWD cwd="/etc/service/enabled/node-status-reporter/log" May 13 00:15:52.277000 audit: PATH item=0 name="/dev/fd/63" inode=19941 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 13 00:15:52.277000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 May 13 00:15:52.279000 audit[3568]: AVC avc: denied { write } for pid=3568 comm="tee" name="fd" dev="proc" ino=20492 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 May 13 00:15:52.284000 audit[3571]: AVC avc: denied { write } for pid=3571 comm="tee" name="fd" dev="proc" ino=18014 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 May 13 00:15:52.279000 audit[3568]: SYSCALL arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=ffffe3506a2c a2=241 a3=1b6 items=1 ppid=3510 pid=3568 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:52.279000 audit: CWD cwd="/etc/service/enabled/cni/log" May 13 00:15:52.279000 audit: PATH item=0 name="/dev/fd/63" inode=19106 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 13 00:15:52.279000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 May 13 00:15:52.284000 audit[3571]: SYSCALL arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=fffff9a81a2a a2=241 a3=1b6 items=1 ppid=3515 pid=3571 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:52.284000 audit: CWD cwd="/etc/service/enabled/felix/log" May 13 00:15:52.284000 audit: PATH item=0 name="/dev/fd/63" inode=20489 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 13 00:15:52.284000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 May 13 00:15:52.298477 systemd[1]: run-containerd-runc-k8s.io-ca6ad9dd0ac7ae391156ed411c1cc63b4f1588009d4dd203ed8b0529b83ed992-runc.N3lFaM.mount: Deactivated successfully. May 13 00:15:55.920000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.8:22-10.0.0.1:49638 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:55.920703 systemd[1]: Started sshd@9-10.0.0.8:22-10.0.0.1:49638.service. May 13 00:15:55.921948 kernel: kauditd_printk_skb: 36 callbacks suppressed May 13 00:15:55.925396 kernel: audit: type=1130 audit(1747095355.920:319): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.8:22-10.0.0.1:49638 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:55.975000 audit[3680]: USER_ACCT pid=3680 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:55.976026 sshd[3680]: Accepted publickey for core from 10.0.0.1 port 49638 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:15:55.976000 audit[3680]: CRED_ACQ pid=3680 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:55.979467 sshd[3680]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:15:55.982188 kernel: audit: type=1101 audit(1747095355.975:320): pid=3680 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:55.982238 kernel: audit: type=1103 audit(1747095355.976:321): pid=3680 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:55.984313 kernel: audit: type=1006 audit(1747095355.976:322): pid=3680 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=10 res=1 May 13 00:15:55.976000 audit[3680]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffff3409b60 a2=3 a3=1 items=0 ppid=1 pid=3680 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:55.988083 kernel: audit: type=1300 audit(1747095355.976:322): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffff3409b60 a2=3 a3=1 items=0 ppid=1 pid=3680 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:55.988163 kernel: audit: type=1327 audit(1747095355.976:322): proctitle=737368643A20636F7265205B707269765D May 13 00:15:55.976000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:15:55.992291 systemd[1]: Started session-10.scope. May 13 00:15:55.992532 systemd-logind[1306]: New session 10 of user core. May 13 00:15:55.998000 audit[3680]: USER_START pid=3680 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.000000 audit[3683]: CRED_ACQ pid=3683 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.006121 kernel: audit: type=1105 audit(1747095355.998:323): pid=3680 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.006207 kernel: audit: type=1103 audit(1747095356.000:324): pid=3683 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.144598 sshd[3680]: pam_unix(sshd:session): session closed for user core May 13 00:15:56.145000 audit[3680]: USER_END pid=3680 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.146598 systemd[1]: Started sshd@10-10.0.0.8:22-10.0.0.1:49642.service. May 13 00:15:56.148031 systemd[1]: sshd@9-10.0.0.8:22-10.0.0.1:49638.service: Deactivated successfully. May 13 00:15:56.149376 systemd-logind[1306]: Session 10 logged out. Waiting for processes to exit. May 13 00:15:56.149443 systemd[1]: session-10.scope: Deactivated successfully. May 13 00:15:56.145000 audit[3680]: CRED_DISP pid=3680 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.150545 systemd-logind[1306]: Removed session 10. May 13 00:15:56.153181 kernel: audit: type=1106 audit(1747095356.145:325): pid=3680 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.153264 kernel: audit: type=1104 audit(1747095356.145:326): pid=3680 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.146000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.8:22-10.0.0.1:49642 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:56.147000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.8:22-10.0.0.1:49638 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:56.182000 audit[3693]: USER_ACCT pid=3693 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.183217 sshd[3693]: Accepted publickey for core from 10.0.0.1 port 49642 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:15:56.184000 audit[3693]: CRED_ACQ pid=3693 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.184000 audit[3693]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffff2f3cee0 a2=3 a3=1 items=0 ppid=1 pid=3693 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=11 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:56.184000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:15:56.185114 sshd[3693]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:15:56.189278 systemd-logind[1306]: New session 11 of user core. May 13 00:15:56.189773 systemd[1]: Started session-11.scope. May 13 00:15:56.193000 audit[3693]: USER_START pid=3693 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.194000 audit[3698]: CRED_ACQ pid=3698 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.361754 sshd[3693]: pam_unix(sshd:session): session closed for user core May 13 00:15:56.362466 systemd[1]: Started sshd@11-10.0.0.8:22-10.0.0.1:49652.service. May 13 00:15:56.362000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.8:22-10.0.0.1:49652 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:56.362000 audit[3693]: USER_END pid=3693 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.362000 audit[3693]: CRED_DISP pid=3693 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.364712 systemd-logind[1306]: Session 11 logged out. Waiting for processes to exit. May 13 00:15:56.364873 systemd[1]: sshd@10-10.0.0.8:22-10.0.0.1:49642.service: Deactivated successfully. May 13 00:15:56.364000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.8:22-10.0.0.1:49642 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:56.365632 systemd[1]: session-11.scope: Deactivated successfully. May 13 00:15:56.366005 systemd-logind[1306]: Removed session 11. May 13 00:15:56.396000 audit[3708]: USER_ACCT pid=3708 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.396932 sshd[3708]: Accepted publickey for core from 10.0.0.1 port 49652 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:15:56.397000 audit[3708]: CRED_ACQ pid=3708 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.397000 audit[3708]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffd6f14660 a2=3 a3=1 items=0 ppid=1 pid=3708 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=12 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:56.397000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:15:56.398194 sshd[3708]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:15:56.413993 systemd-logind[1306]: New session 12 of user core. May 13 00:15:56.415476 systemd[1]: Started session-12.scope. May 13 00:15:56.420000 audit[3708]: USER_START pid=3708 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.421000 audit[3713]: CRED_ACQ pid=3713 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.531431 sshd[3708]: pam_unix(sshd:session): session closed for user core May 13 00:15:56.533000 audit[3708]: USER_END pid=3708 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.533000 audit[3708]: CRED_DISP pid=3708 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:15:56.535244 systemd[1]: sshd@11-10.0.0.8:22-10.0.0.1:49652.service: Deactivated successfully. May 13 00:15:56.535000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.8:22-10.0.0.1:49652 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:15:56.536462 systemd-logind[1306]: Session 12 logged out. Waiting for processes to exit. May 13 00:15:56.536529 systemd[1]: session-12.scope: Deactivated successfully. May 13 00:15:56.537903 systemd-logind[1306]: Removed session 12. May 13 00:15:58.130183 env[1323]: time="2025-05-13T00:15:58.129844748Z" level=info msg="StopPodSandbox for \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\"" May 13 00:15:58.130183 env[1323]: time="2025-05-13T00:15:58.130048028Z" level=info msg="StopPodSandbox for \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\"" May 13 00:15:58.235928 kubelet[2230]: I0513 00:15:58.235718 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-bsdv5" podStartSLOduration=8.498915312 podStartE2EDuration="20.235698263s" podCreationTimestamp="2025-05-13 00:15:38 +0000 UTC" firstStartedPulling="2025-05-13 00:15:38.858094626 +0000 UTC m=+22.836317374" lastFinishedPulling="2025-05-13 00:15:50.594877617 +0000 UTC m=+34.573100325" observedRunningTime="2025-05-13 00:15:51.275935185 +0000 UTC m=+35.254157893" watchObservedRunningTime="2025-05-13 00:15:58.235698263 +0000 UTC m=+42.213921011" May 13 00:15:58.412208 env[1323]: 2025-05-13 00:15:58.239 [INFO][3802] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" May 13 00:15:58.412208 env[1323]: 2025-05-13 00:15:58.240 [INFO][3802] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" iface="eth0" netns="/var/run/netns/cni-0e8cd5e8-a71b-1136-8d31-74c329c1e3df" May 13 00:15:58.412208 env[1323]: 2025-05-13 00:15:58.241 [INFO][3802] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" iface="eth0" netns="/var/run/netns/cni-0e8cd5e8-a71b-1136-8d31-74c329c1e3df" May 13 00:15:58.412208 env[1323]: 2025-05-13 00:15:58.241 [INFO][3802] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" iface="eth0" netns="/var/run/netns/cni-0e8cd5e8-a71b-1136-8d31-74c329c1e3df" May 13 00:15:58.412208 env[1323]: 2025-05-13 00:15:58.241 [INFO][3802] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" May 13 00:15:58.412208 env[1323]: 2025-05-13 00:15:58.242 [INFO][3802] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" May 13 00:15:58.412208 env[1323]: 2025-05-13 00:15:58.393 [INFO][3821] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" HandleID="k8s-pod-network.316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" Workload="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:15:58.412208 env[1323]: 2025-05-13 00:15:58.393 [INFO][3821] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:15:58.412208 env[1323]: 2025-05-13 00:15:58.394 [INFO][3821] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:15:58.412208 env[1323]: 2025-05-13 00:15:58.407 [WARNING][3821] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" HandleID="k8s-pod-network.316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" Workload="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:15:58.412208 env[1323]: 2025-05-13 00:15:58.407 [INFO][3821] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" HandleID="k8s-pod-network.316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" Workload="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:15:58.412208 env[1323]: 2025-05-13 00:15:58.408 [INFO][3821] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:15:58.412208 env[1323]: 2025-05-13 00:15:58.410 [INFO][3802] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" May 13 00:15:58.416940 systemd[1]: run-netns-cni\x2d0e8cd5e8\x2da71b\x2d1136\x2d8d31\x2d74c329c1e3df.mount: Deactivated successfully. May 13 00:15:58.417343 env[1323]: time="2025-05-13T00:15:58.417284563Z" level=info msg="TearDown network for sandbox \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\" successfully" May 13 00:15:58.417455 env[1323]: time="2025-05-13T00:15:58.417435683Z" level=info msg="StopPodSandbox for \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\" returns successfully" May 13 00:15:58.418563 kubelet[2230]: E0513 00:15:58.417847 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:58.418683 env[1323]: time="2025-05-13T00:15:58.418661163Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-ncf86,Uid:8dde30de-095c-474d-a07e-44b258f6e953,Namespace:kube-system,Attempt:1,}" May 13 00:15:58.427056 env[1323]: 2025-05-13 00:15:58.236 [INFO][3803] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" May 13 00:15:58.427056 env[1323]: 2025-05-13 00:15:58.239 [INFO][3803] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" iface="eth0" netns="/var/run/netns/cni-05aa9825-a9d3-a8bc-6b63-9e646bec784e" May 13 00:15:58.427056 env[1323]: 2025-05-13 00:15:58.239 [INFO][3803] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" iface="eth0" netns="/var/run/netns/cni-05aa9825-a9d3-a8bc-6b63-9e646bec784e" May 13 00:15:58.427056 env[1323]: 2025-05-13 00:15:58.240 [INFO][3803] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" iface="eth0" netns="/var/run/netns/cni-05aa9825-a9d3-a8bc-6b63-9e646bec784e" May 13 00:15:58.427056 env[1323]: 2025-05-13 00:15:58.240 [INFO][3803] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" May 13 00:15:58.427056 env[1323]: 2025-05-13 00:15:58.240 [INFO][3803] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" May 13 00:15:58.427056 env[1323]: 2025-05-13 00:15:58.393 [INFO][3820] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" HandleID="k8s-pod-network.d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:15:58.427056 env[1323]: 2025-05-13 00:15:58.394 [INFO][3820] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:15:58.427056 env[1323]: 2025-05-13 00:15:58.408 [INFO][3820] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:15:58.427056 env[1323]: 2025-05-13 00:15:58.421 [WARNING][3820] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" HandleID="k8s-pod-network.d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:15:58.427056 env[1323]: 2025-05-13 00:15:58.422 [INFO][3820] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" HandleID="k8s-pod-network.d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:15:58.427056 env[1323]: 2025-05-13 00:15:58.423 [INFO][3820] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:15:58.427056 env[1323]: 2025-05-13 00:15:58.425 [INFO][3803] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" May 13 00:15:58.430623 env[1323]: time="2025-05-13T00:15:58.430248527Z" level=info msg="TearDown network for sandbox \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\" successfully" May 13 00:15:58.430623 env[1323]: time="2025-05-13T00:15:58.430281967Z" level=info msg="StopPodSandbox for \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\" returns successfully" May 13 00:15:58.429418 systemd[1]: run-netns-cni\x2d05aa9825\x2da9d3\x2da8bc\x2d6b63\x2d9e646bec784e.mount: Deactivated successfully. May 13 00:15:58.431157 env[1323]: time="2025-05-13T00:15:58.431122648Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5b58ff46dd-56crg,Uid:4596526d-58fb-4ef8-81dc-275ff89a1ce4,Namespace:calico-apiserver,Attempt:1,}" May 13 00:15:58.581444 systemd-networkd[1102]: cali6bab1e72142: Link UP May 13 00:15:58.583783 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready May 13 00:15:58.583814 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cali6bab1e72142: link becomes ready May 13 00:15:58.583916 systemd-networkd[1102]: cali6bab1e72142: Gained carrier May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.476 [INFO][3836] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.492 [INFO][3836] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0 coredns-7db6d8ff4d- kube-system 8dde30de-095c-474d-a07e-44b258f6e953 844 0 2025-05-13 00:15:31 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7db6d8ff4d projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7db6d8ff4d-ncf86 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali6bab1e72142 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" Namespace="kube-system" Pod="coredns-7db6d8ff4d-ncf86" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--ncf86-" May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.492 [INFO][3836] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" Namespace="kube-system" Pod="coredns-7db6d8ff4d-ncf86" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.522 [INFO][3863] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" HandleID="k8s-pod-network.bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" Workload="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.534 [INFO][3863] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" HandleID="k8s-pod-network.bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" Workload="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000365240), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7db6d8ff4d-ncf86", "timestamp":"2025-05-13 00:15:58.522265638 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.534 [INFO][3863] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.534 [INFO][3863] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.534 [INFO][3863] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.536 [INFO][3863] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" host="localhost" May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.541 [INFO][3863] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.546 [INFO][3863] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.548 [INFO][3863] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.552 [INFO][3863] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.552 [INFO][3863] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" host="localhost" May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.554 [INFO][3863] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.558 [INFO][3863] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" host="localhost" May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.563 [INFO][3863] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" host="localhost" May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.563 [INFO][3863] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" host="localhost" May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.563 [INFO][3863] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:15:58.598481 env[1323]: 2025-05-13 00:15:58.563 [INFO][3863] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" HandleID="k8s-pod-network.bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" Workload="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:15:58.599172 env[1323]: 2025-05-13 00:15:58.570 [INFO][3836] cni-plugin/k8s.go 386: Populated endpoint ContainerID="bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" Namespace="kube-system" Pod="coredns-7db6d8ff4d-ncf86" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"8dde30de-095c-474d-a07e-44b258f6e953", ResourceVersion:"844", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7db6d8ff4d-ncf86", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6bab1e72142", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:15:58.599172 env[1323]: 2025-05-13 00:15:58.570 [INFO][3836] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.129/32] ContainerID="bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" Namespace="kube-system" Pod="coredns-7db6d8ff4d-ncf86" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:15:58.599172 env[1323]: 2025-05-13 00:15:58.570 [INFO][3836] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6bab1e72142 ContainerID="bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" Namespace="kube-system" Pod="coredns-7db6d8ff4d-ncf86" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:15:58.599172 env[1323]: 2025-05-13 00:15:58.583 [INFO][3836] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" Namespace="kube-system" Pod="coredns-7db6d8ff4d-ncf86" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:15:58.599172 env[1323]: 2025-05-13 00:15:58.584 [INFO][3836] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" Namespace="kube-system" Pod="coredns-7db6d8ff4d-ncf86" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"8dde30de-095c-474d-a07e-44b258f6e953", ResourceVersion:"844", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af", Pod:"coredns-7db6d8ff4d-ncf86", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6bab1e72142", MAC:"76:5e:5d:23:f1:75", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:15:58.599172 env[1323]: 2025-05-13 00:15:58.593 [INFO][3836] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af" Namespace="kube-system" Pod="coredns-7db6d8ff4d-ncf86" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:15:58.605449 systemd-networkd[1102]: calie3fc7efc9ef: Link UP May 13 00:15:58.606941 systemd-networkd[1102]: calie3fc7efc9ef: Gained carrier May 13 00:15:58.607604 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): calie3fc7efc9ef: link becomes ready May 13 00:15:58.628124 env[1323]: time="2025-05-13T00:15:58.626485272Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 13 00:15:58.628124 env[1323]: time="2025-05-13T00:15:58.626529992Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 13 00:15:58.628124 env[1323]: time="2025-05-13T00:15:58.626593632Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 13 00:15:58.628124 env[1323]: time="2025-05-13T00:15:58.626758872Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af pid=3908 runtime=io.containerd.runc.v2 May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.476 [INFO][3846] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.494 [INFO][3846] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0 calico-apiserver-5b58ff46dd- calico-apiserver 4596526d-58fb-4ef8-81dc-275ff89a1ce4 843 0 2025-05-13 00:15:37 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:5b58ff46dd projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-5b58ff46dd-56crg eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calie3fc7efc9ef [] []}} ContainerID="30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-56crg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-" May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.494 [INFO][3846] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-56crg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.538 [INFO][3865] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" HandleID="k8s-pod-network.30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.552 [INFO][3865] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" HandleID="k8s-pod-network.30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000305900), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-5b58ff46dd-56crg", "timestamp":"2025-05-13 00:15:58.538415883 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.552 [INFO][3865] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.563 [INFO][3865] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.563 [INFO][3865] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.566 [INFO][3865] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" host="localhost" May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.571 [INFO][3865] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.576 [INFO][3865] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.578 [INFO][3865] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.581 [INFO][3865] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.582 [INFO][3865] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" host="localhost" May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.585 [INFO][3865] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7 May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.589 [INFO][3865] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" host="localhost" May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.594 [INFO][3865] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" host="localhost" May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.595 [INFO][3865] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" host="localhost" May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.595 [INFO][3865] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:15:58.628124 env[1323]: 2025-05-13 00:15:58.595 [INFO][3865] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" HandleID="k8s-pod-network.30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:15:58.629369 env[1323]: 2025-05-13 00:15:58.598 [INFO][3846] cni-plugin/k8s.go 386: Populated endpoint ContainerID="30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-56crg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0", GenerateName:"calico-apiserver-5b58ff46dd-", Namespace:"calico-apiserver", SelfLink:"", UID:"4596526d-58fb-4ef8-81dc-275ff89a1ce4", ResourceVersion:"843", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5b58ff46dd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-5b58ff46dd-56crg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie3fc7efc9ef", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:15:58.629369 env[1323]: 2025-05-13 00:15:58.598 [INFO][3846] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.130/32] ContainerID="30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-56crg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:15:58.629369 env[1323]: 2025-05-13 00:15:58.598 [INFO][3846] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie3fc7efc9ef ContainerID="30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-56crg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:15:58.629369 env[1323]: 2025-05-13 00:15:58.607 [INFO][3846] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-56crg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:15:58.629369 env[1323]: 2025-05-13 00:15:58.607 [INFO][3846] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-56crg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0", GenerateName:"calico-apiserver-5b58ff46dd-", Namespace:"calico-apiserver", SelfLink:"", UID:"4596526d-58fb-4ef8-81dc-275ff89a1ce4", ResourceVersion:"843", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5b58ff46dd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7", Pod:"calico-apiserver-5b58ff46dd-56crg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie3fc7efc9ef", MAC:"f6:38:87:98:68:75", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:15:58.629369 env[1323]: 2025-05-13 00:15:58.621 [INFO][3846] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-56crg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:15:58.654664 env[1323]: time="2025-05-13T00:15:58.654566641Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 13 00:15:58.654664 env[1323]: time="2025-05-13T00:15:58.654613201Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 13 00:15:58.654664 env[1323]: time="2025-05-13T00:15:58.654623441Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 13 00:15:58.661607 env[1323]: time="2025-05-13T00:15:58.655934202Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7 pid=3956 runtime=io.containerd.runc.v2 May 13 00:15:58.675627 systemd-resolved[1238]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 13 00:15:58.708911 env[1323]: time="2025-05-13T00:15:58.708868219Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-ncf86,Uid:8dde30de-095c-474d-a07e-44b258f6e953,Namespace:kube-system,Attempt:1,} returns sandbox id \"bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af\"" May 13 00:15:58.709965 kubelet[2230]: E0513 00:15:58.709936 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:58.714081 env[1323]: time="2025-05-13T00:15:58.714041021Z" level=info msg="CreateContainer within sandbox \"bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 13 00:15:58.719289 systemd-resolved[1238]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 13 00:15:58.753494 env[1323]: time="2025-05-13T00:15:58.753433034Z" level=info msg="CreateContainer within sandbox \"bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"9af0371193151b9df95ca27b5f972edb31c9d32d48478f643760f963ec256993\"" May 13 00:15:58.755021 env[1323]: time="2025-05-13T00:15:58.754087274Z" level=info msg="StartContainer for \"9af0371193151b9df95ca27b5f972edb31c9d32d48478f643760f963ec256993\"" May 13 00:15:58.763916 env[1323]: time="2025-05-13T00:15:58.763869078Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5b58ff46dd-56crg,Uid:4596526d-58fb-4ef8-81dc-275ff89a1ce4,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7\"" May 13 00:15:58.765685 env[1323]: time="2025-05-13T00:15:58.765639598Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" May 13 00:15:58.814498 env[1323]: time="2025-05-13T00:15:58.814314334Z" level=info msg="StartContainer for \"9af0371193151b9df95ca27b5f972edb31c9d32d48478f643760f963ec256993\" returns successfully" May 13 00:15:59.278680 kubelet[2230]: E0513 00:15:59.278037 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:15:59.289066 kubelet[2230]: I0513 00:15:59.288966 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-ncf86" podStartSLOduration=28.288949765 podStartE2EDuration="28.288949765s" podCreationTimestamp="2025-05-13 00:15:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-13 00:15:59.288883085 +0000 UTC m=+43.267105793" watchObservedRunningTime="2025-05-13 00:15:59.288949765 +0000 UTC m=+43.267172513" May 13 00:15:59.306000 audit[4052]: NETFILTER_CFG table=filter:95 family=2 entries=18 op=nft_register_rule pid=4052 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:15:59.306000 audit[4052]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=6652 a0=3 a1=ffffda331950 a2=0 a3=1 items=0 ppid=2393 pid=4052 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:59.306000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:59.315000 audit[4052]: NETFILTER_CFG table=nat:96 family=2 entries=12 op=nft_register_rule pid=4052 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:15:59.315000 audit[4052]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2700 a0=3 a1=ffffda331950 a2=0 a3=1 items=0 ppid=2393 pid=4052 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:59.315000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:59.334000 audit[4054]: NETFILTER_CFG table=filter:97 family=2 entries=15 op=nft_register_rule pid=4054 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:15:59.334000 audit[4054]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=4420 a0=3 a1=ffffe4686c20 a2=0 a3=1 items=0 ppid=2393 pid=4054 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:59.334000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:15:59.348000 audit[4054]: NETFILTER_CFG table=nat:98 family=2 entries=33 op=nft_register_chain pid=4054 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:15:59.348000 audit[4054]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=13428 a0=3 a1=ffffe4686c20 a2=0 a3=1 items=0 ppid=2393 pid=4054 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:15:59.348000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:00.131194 env[1323]: time="2025-05-13T00:16:00.130929824Z" level=info msg="StopPodSandbox for \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\"" May 13 00:16:00.272413 env[1323]: 2025-05-13 00:16:00.233 [INFO][4093] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" May 13 00:16:00.272413 env[1323]: 2025-05-13 00:16:00.233 [INFO][4093] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" iface="eth0" netns="/var/run/netns/cni-6fe42350-b749-2209-79eb-f601018535ae" May 13 00:16:00.272413 env[1323]: 2025-05-13 00:16:00.234 [INFO][4093] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" iface="eth0" netns="/var/run/netns/cni-6fe42350-b749-2209-79eb-f601018535ae" May 13 00:16:00.272413 env[1323]: 2025-05-13 00:16:00.234 [INFO][4093] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" iface="eth0" netns="/var/run/netns/cni-6fe42350-b749-2209-79eb-f601018535ae" May 13 00:16:00.272413 env[1323]: 2025-05-13 00:16:00.234 [INFO][4093] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" May 13 00:16:00.272413 env[1323]: 2025-05-13 00:16:00.234 [INFO][4093] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" May 13 00:16:00.272413 env[1323]: 2025-05-13 00:16:00.255 [INFO][4101] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" HandleID="k8s-pod-network.ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" Workload="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:00.272413 env[1323]: 2025-05-13 00:16:00.255 [INFO][4101] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:00.272413 env[1323]: 2025-05-13 00:16:00.255 [INFO][4101] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:00.272413 env[1323]: 2025-05-13 00:16:00.266 [WARNING][4101] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" HandleID="k8s-pod-network.ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" Workload="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:00.272413 env[1323]: 2025-05-13 00:16:00.266 [INFO][4101] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" HandleID="k8s-pod-network.ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" Workload="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:00.272413 env[1323]: 2025-05-13 00:16:00.267 [INFO][4101] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:00.272413 env[1323]: 2025-05-13 00:16:00.270 [INFO][4093] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" May 13 00:16:00.272934 env[1323]: time="2025-05-13T00:16:00.272623585Z" level=info msg="TearDown network for sandbox \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\" successfully" May 13 00:16:00.272934 env[1323]: time="2025-05-13T00:16:00.272661105Z" level=info msg="StopPodSandbox for \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\" returns successfully" May 13 00:16:00.274766 systemd[1]: run-netns-cni\x2d6fe42350\x2db749\x2d2209\x2d79eb\x2df601018535ae.mount: Deactivated successfully. May 13 00:16:00.275789 env[1323]: time="2025-05-13T00:16:00.275744026Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-srb86,Uid:838e7283-9b17-4532-a9f4-4b17f2b5ca3b,Namespace:calico-system,Attempt:1,}" May 13 00:16:00.280367 kubelet[2230]: E0513 00:16:00.280324 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:16:00.362641 systemd-networkd[1102]: calie3fc7efc9ef: Gained IPv6LL May 13 00:16:00.418827 systemd-networkd[1102]: calia028079e884: Link UP May 13 00:16:00.421683 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready May 13 00:16:00.421720 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): calia028079e884: link becomes ready May 13 00:16:00.421576 systemd-networkd[1102]: calia028079e884: Gained carrier May 13 00:16:00.428806 systemd-networkd[1102]: cali6bab1e72142: Gained IPv6LL May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.337 [INFO][4111] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.351 [INFO][4111] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--srb86-eth0 csi-node-driver- calico-system 838e7283-9b17-4532-a9f4-4b17f2b5ca3b 886 0 2025-05-13 00:15:38 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:55b7b4b9d k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-srb86 eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calia028079e884 [] []}} ContainerID="ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" Namespace="calico-system" Pod="csi-node-driver-srb86" WorkloadEndpoint="localhost-k8s-csi--node--driver--srb86-" May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.351 [INFO][4111] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" Namespace="calico-system" Pod="csi-node-driver-srb86" WorkloadEndpoint="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.377 [INFO][4127] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" HandleID="k8s-pod-network.ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" Workload="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.389 [INFO][4127] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" HandleID="k8s-pod-network.ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" Workload="localhost-k8s-csi--node--driver--srb86-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002786f0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-srb86", "timestamp":"2025-05-13 00:16:00.377040456 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.389 [INFO][4127] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.389 [INFO][4127] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.389 [INFO][4127] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.391 [INFO][4127] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" host="localhost" May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.395 [INFO][4127] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.400 [INFO][4127] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.402 [INFO][4127] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.404 [INFO][4127] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.404 [INFO][4127] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" host="localhost" May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.405 [INFO][4127] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.409 [INFO][4127] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" host="localhost" May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.414 [INFO][4127] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" host="localhost" May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.414 [INFO][4127] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" host="localhost" May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.415 [INFO][4127] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:00.436006 env[1323]: 2025-05-13 00:16:00.415 [INFO][4127] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" HandleID="k8s-pod-network.ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" Workload="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:00.436637 env[1323]: 2025-05-13 00:16:00.417 [INFO][4111] cni-plugin/k8s.go 386: Populated endpoint ContainerID="ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" Namespace="calico-system" Pod="csi-node-driver-srb86" WorkloadEndpoint="localhost-k8s-csi--node--driver--srb86-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--srb86-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"838e7283-9b17-4532-a9f4-4b17f2b5ca3b", ResourceVersion:"886", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b7b4b9d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-srb86", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calia028079e884", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:00.436637 env[1323]: 2025-05-13 00:16:00.417 [INFO][4111] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.131/32] ContainerID="ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" Namespace="calico-system" Pod="csi-node-driver-srb86" WorkloadEndpoint="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:00.436637 env[1323]: 2025-05-13 00:16:00.417 [INFO][4111] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia028079e884 ContainerID="ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" Namespace="calico-system" Pod="csi-node-driver-srb86" WorkloadEndpoint="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:00.436637 env[1323]: 2025-05-13 00:16:00.419 [INFO][4111] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" Namespace="calico-system" Pod="csi-node-driver-srb86" WorkloadEndpoint="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:00.436637 env[1323]: 2025-05-13 00:16:00.421 [INFO][4111] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" Namespace="calico-system" Pod="csi-node-driver-srb86" WorkloadEndpoint="localhost-k8s-csi--node--driver--srb86-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--srb86-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"838e7283-9b17-4532-a9f4-4b17f2b5ca3b", ResourceVersion:"886", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b7b4b9d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff", Pod:"csi-node-driver-srb86", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calia028079e884", MAC:"6e:a3:08:51:04:f2", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:00.436637 env[1323]: 2025-05-13 00:16:00.433 [INFO][4111] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff" Namespace="calico-system" Pod="csi-node-driver-srb86" WorkloadEndpoint="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:00.453288 env[1323]: time="2025-05-13T00:16:00.453209198Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 13 00:16:00.453288 env[1323]: time="2025-05-13T00:16:00.453253438Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 13 00:16:00.453504 env[1323]: time="2025-05-13T00:16:00.453264638Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 13 00:16:00.453504 env[1323]: time="2025-05-13T00:16:00.453458118Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff pid=4155 runtime=io.containerd.runc.v2 May 13 00:16:00.505615 systemd-resolved[1238]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 13 00:16:00.514816 env[1323]: time="2025-05-13T00:16:00.514768016Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-srb86,Uid:838e7283-9b17-4532-a9f4-4b17f2b5ca3b,Namespace:calico-system,Attempt:1,} returns sandbox id \"ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff\"" May 13 00:16:00.685571 env[1323]: time="2025-05-13T00:16:00.684390745Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/apiserver:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:00.687167 env[1323]: time="2025-05-13T00:16:00.687132666Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:eca64fb9fcc40e83ed2310ac1fab340ba460a939c54e10dc0b7428f02b9b6253,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:00.687769 env[1323]: time="2025-05-13T00:16:00.687745026Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/apiserver:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:00.689518 env[1323]: time="2025-05-13T00:16:00.689482906Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:00.690038 env[1323]: time="2025-05-13T00:16:00.690006466Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:eca64fb9fcc40e83ed2310ac1fab340ba460a939c54e10dc0b7428f02b9b6253\"" May 13 00:16:00.691784 env[1323]: time="2025-05-13T00:16:00.691755787Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\"" May 13 00:16:00.692678 env[1323]: time="2025-05-13T00:16:00.692648067Z" level=info msg="CreateContainer within sandbox \"30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 13 00:16:00.703623 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3466575194.mount: Deactivated successfully. May 13 00:16:00.705678 env[1323]: time="2025-05-13T00:16:00.705630631Z" level=info msg="CreateContainer within sandbox \"30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"bd5e8c0830cd6ca53c98fb729473377addaedc9c9e7a14cc7bad9a5d0b817fd1\"" May 13 00:16:00.706335 env[1323]: time="2025-05-13T00:16:00.706282911Z" level=info msg="StartContainer for \"bd5e8c0830cd6ca53c98fb729473377addaedc9c9e7a14cc7bad9a5d0b817fd1\"" May 13 00:16:00.807918 env[1323]: time="2025-05-13T00:16:00.807864661Z" level=info msg="StartContainer for \"bd5e8c0830cd6ca53c98fb729473377addaedc9c9e7a14cc7bad9a5d0b817fd1\" returns successfully" May 13 00:16:01.130109 env[1323]: time="2025-05-13T00:16:01.130071992Z" level=info msg="StopPodSandbox for \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\"" May 13 00:16:01.130375 env[1323]: time="2025-05-13T00:16:01.130117232Z" level=info msg="StopPodSandbox for \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\"" May 13 00:16:01.223430 env[1323]: 2025-05-13 00:16:01.181 [INFO][4283] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" May 13 00:16:01.223430 env[1323]: 2025-05-13 00:16:01.181 [INFO][4283] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" iface="eth0" netns="/var/run/netns/cni-9e0b2d9d-95b1-1a08-ec78-9e99b68f736a" May 13 00:16:01.223430 env[1323]: 2025-05-13 00:16:01.182 [INFO][4283] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" iface="eth0" netns="/var/run/netns/cni-9e0b2d9d-95b1-1a08-ec78-9e99b68f736a" May 13 00:16:01.223430 env[1323]: 2025-05-13 00:16:01.182 [INFO][4283] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" iface="eth0" netns="/var/run/netns/cni-9e0b2d9d-95b1-1a08-ec78-9e99b68f736a" May 13 00:16:01.223430 env[1323]: 2025-05-13 00:16:01.182 [INFO][4283] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" May 13 00:16:01.223430 env[1323]: 2025-05-13 00:16:01.182 [INFO][4283] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" May 13 00:16:01.223430 env[1323]: 2025-05-13 00:16:01.203 [INFO][4301] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" HandleID="k8s-pod-network.922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:01.223430 env[1323]: 2025-05-13 00:16:01.203 [INFO][4301] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:01.223430 env[1323]: 2025-05-13 00:16:01.204 [INFO][4301] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:01.223430 env[1323]: 2025-05-13 00:16:01.212 [WARNING][4301] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" HandleID="k8s-pod-network.922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:01.223430 env[1323]: 2025-05-13 00:16:01.212 [INFO][4301] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" HandleID="k8s-pod-network.922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:01.223430 env[1323]: 2025-05-13 00:16:01.214 [INFO][4301] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:01.223430 env[1323]: 2025-05-13 00:16:01.217 [INFO][4283] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" May 13 00:16:01.224138 env[1323]: time="2025-05-13T00:16:01.223615778Z" level=info msg="TearDown network for sandbox \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\" successfully" May 13 00:16:01.224138 env[1323]: time="2025-05-13T00:16:01.223648058Z" level=info msg="StopPodSandbox for \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\" returns successfully" May 13 00:16:01.225079 env[1323]: time="2025-05-13T00:16:01.225044738Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5b58ff46dd-zk6tg,Uid:4a3e9432-3790-4f08-9c11-d537f6aab7dd,Namespace:calico-apiserver,Attempt:1,}" May 13 00:16:01.233570 env[1323]: 2025-05-13 00:16:01.180 [INFO][4284] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" May 13 00:16:01.233570 env[1323]: 2025-05-13 00:16:01.180 [INFO][4284] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" iface="eth0" netns="/var/run/netns/cni-429289c8-6b8e-743c-1deb-de4f6d8c0c04" May 13 00:16:01.233570 env[1323]: 2025-05-13 00:16:01.181 [INFO][4284] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" iface="eth0" netns="/var/run/netns/cni-429289c8-6b8e-743c-1deb-de4f6d8c0c04" May 13 00:16:01.233570 env[1323]: 2025-05-13 00:16:01.181 [INFO][4284] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" iface="eth0" netns="/var/run/netns/cni-429289c8-6b8e-743c-1deb-de4f6d8c0c04" May 13 00:16:01.233570 env[1323]: 2025-05-13 00:16:01.181 [INFO][4284] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" May 13 00:16:01.233570 env[1323]: 2025-05-13 00:16:01.181 [INFO][4284] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" May 13 00:16:01.233570 env[1323]: 2025-05-13 00:16:01.207 [INFO][4299] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" HandleID="k8s-pod-network.663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" Workload="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:01.233570 env[1323]: 2025-05-13 00:16:01.208 [INFO][4299] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:01.233570 env[1323]: 2025-05-13 00:16:01.214 [INFO][4299] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:01.233570 env[1323]: 2025-05-13 00:16:01.227 [WARNING][4299] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" HandleID="k8s-pod-network.663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" Workload="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:01.233570 env[1323]: 2025-05-13 00:16:01.227 [INFO][4299] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" HandleID="k8s-pod-network.663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" Workload="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:01.233570 env[1323]: 2025-05-13 00:16:01.229 [INFO][4299] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:01.233570 env[1323]: 2025-05-13 00:16:01.231 [INFO][4284] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" May 13 00:16:01.234020 env[1323]: time="2025-05-13T00:16:01.233716460Z" level=info msg="TearDown network for sandbox \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\" successfully" May 13 00:16:01.234020 env[1323]: time="2025-05-13T00:16:01.233749460Z" level=info msg="StopPodSandbox for \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\" returns successfully" May 13 00:16:01.234068 kubelet[2230]: E0513 00:16:01.234026 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:16:01.234712 env[1323]: time="2025-05-13T00:16:01.234681461Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-k77ll,Uid:943c0c0f-6fce-4911-b87f-c827598e67dc,Namespace:kube-system,Attempt:1,}" May 13 00:16:01.285640 kubelet[2230]: E0513 00:16:01.285611 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:16:01.325000 audit[4349]: NETFILTER_CFG table=filter:99 family=2 entries=12 op=nft_register_rule pid=4349 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:01.329282 kernel: kauditd_printk_skb: 35 callbacks suppressed May 13 00:16:01.329398 kernel: audit: type=1325 audit(1747095361.325:350): table=filter:99 family=2 entries=12 op=nft_register_rule pid=4349 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:01.329425 kernel: audit: type=1300 audit(1747095361.325:350): arch=c00000b7 syscall=211 success=yes exit=4420 a0=3 a1=ffffe2b46fd0 a2=0 a3=1 items=0 ppid=2393 pid=4349 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:01.325000 audit[4349]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=4420 a0=3 a1=ffffe2b46fd0 a2=0 a3=1 items=0 ppid=2393 pid=4349 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:01.325000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:01.335182 kernel: audit: type=1327 audit(1747095361.325:350): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:01.342000 audit[4349]: NETFILTER_CFG table=nat:100 family=2 entries=18 op=nft_register_rule pid=4349 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:01.342000 audit[4349]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=5004 a0=3 a1=ffffe2b46fd0 a2=0 a3=1 items=0 ppid=2393 pid=4349 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:01.348797 kernel: audit: type=1325 audit(1747095361.342:351): table=nat:100 family=2 entries=18 op=nft_register_rule pid=4349 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:01.348889 kernel: audit: type=1300 audit(1747095361.342:351): arch=c00000b7 syscall=211 success=yes exit=5004 a0=3 a1=ffffe2b46fd0 a2=0 a3=1 items=0 ppid=2393 pid=4349 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:01.348913 kernel: audit: type=1327 audit(1747095361.342:351): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:01.342000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:01.423397 systemd-networkd[1102]: cali99ccf0c6009: Link UP May 13 00:16:01.428978 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready May 13 00:16:01.429650 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cali99ccf0c6009: link becomes ready May 13 00:16:01.429118 systemd-networkd[1102]: cali99ccf0c6009: Gained carrier May 13 00:16:01.446198 kubelet[2230]: I0513 00:16:01.445731 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-5b58ff46dd-56crg" podStartSLOduration=22.519659929 podStartE2EDuration="24.445711358s" podCreationTimestamp="2025-05-13 00:15:37 +0000 UTC" firstStartedPulling="2025-05-13 00:15:58.765095958 +0000 UTC m=+42.743318666" lastFinishedPulling="2025-05-13 00:16:00.691147347 +0000 UTC m=+44.669370095" observedRunningTime="2025-05-13 00:16:01.314914162 +0000 UTC m=+45.293136910" watchObservedRunningTime="2025-05-13 00:16:01.445711358 +0000 UTC m=+45.423934066" May 13 00:16:01.458587 systemd[1]: run-netns-cni\x2d429289c8\x2d6b8e\x2d743c\x2d1deb\x2dde4f6d8c0c04.mount: Deactivated successfully. May 13 00:16:01.458728 systemd[1]: run-netns-cni\x2d9e0b2d9d\x2d95b1\x2d1a08\x2dec78\x2d9e99b68f736a.mount: Deactivated successfully. May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.281 [INFO][4330] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.322 [INFO][4330] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0 coredns-7db6d8ff4d- kube-system 943c0c0f-6fce-4911-b87f-c827598e67dc 898 0 2025-05-13 00:15:31 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7db6d8ff4d projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7db6d8ff4d-k77ll eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali99ccf0c6009 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" Namespace="kube-system" Pod="coredns-7db6d8ff4d-k77ll" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--k77ll-" May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.322 [INFO][4330] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" Namespace="kube-system" Pod="coredns-7db6d8ff4d-k77ll" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.374 [INFO][4354] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" HandleID="k8s-pod-network.97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" Workload="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.385 [INFO][4354] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" HandleID="k8s-pod-network.97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" Workload="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000522ca0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7db6d8ff4d-k77ll", "timestamp":"2025-05-13 00:16:01.374102539 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.385 [INFO][4354] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.385 [INFO][4354] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.385 [INFO][4354] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.387 [INFO][4354] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" host="localhost" May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.392 [INFO][4354] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.396 [INFO][4354] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.398 [INFO][4354] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.403 [INFO][4354] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.403 [INFO][4354] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" host="localhost" May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.405 [INFO][4354] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646 May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.409 [INFO][4354] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" host="localhost" May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.415 [INFO][4354] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" host="localhost" May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.415 [INFO][4354] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" host="localhost" May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.415 [INFO][4354] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:01.462759 env[1323]: 2025-05-13 00:16:01.415 [INFO][4354] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" HandleID="k8s-pod-network.97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" Workload="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:01.465391 env[1323]: 2025-05-13 00:16:01.420 [INFO][4330] cni-plugin/k8s.go 386: Populated endpoint ContainerID="97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" Namespace="kube-system" Pod="coredns-7db6d8ff4d-k77ll" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"943c0c0f-6fce-4911-b87f-c827598e67dc", ResourceVersion:"898", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7db6d8ff4d-k77ll", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali99ccf0c6009", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:01.465391 env[1323]: 2025-05-13 00:16:01.421 [INFO][4330] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.132/32] ContainerID="97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" Namespace="kube-system" Pod="coredns-7db6d8ff4d-k77ll" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:01.465391 env[1323]: 2025-05-13 00:16:01.421 [INFO][4330] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali99ccf0c6009 ContainerID="97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" Namespace="kube-system" Pod="coredns-7db6d8ff4d-k77ll" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:01.465391 env[1323]: 2025-05-13 00:16:01.429 [INFO][4330] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" Namespace="kube-system" Pod="coredns-7db6d8ff4d-k77ll" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:01.465391 env[1323]: 2025-05-13 00:16:01.430 [INFO][4330] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" Namespace="kube-system" Pod="coredns-7db6d8ff4d-k77ll" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"943c0c0f-6fce-4911-b87f-c827598e67dc", ResourceVersion:"898", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646", Pod:"coredns-7db6d8ff4d-k77ll", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali99ccf0c6009", MAC:"6e:6a:03:2a:90:d7", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:01.465391 env[1323]: 2025-05-13 00:16:01.442 [INFO][4330] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646" Namespace="kube-system" Pod="coredns-7db6d8ff4d-k77ll" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:01.474648 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cali0bed5f3decf: link becomes ready May 13 00:16:01.473801 systemd-networkd[1102]: cali0bed5f3decf: Link UP May 13 00:16:01.473999 systemd-networkd[1102]: cali0bed5f3decf: Gained carrier May 13 00:16:01.488996 env[1323]: time="2025-05-13T00:16:01.488923490Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 13 00:16:01.488996 env[1323]: time="2025-05-13T00:16:01.488965730Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 13 00:16:01.489159 env[1323]: time="2025-05-13T00:16:01.488979570Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 13 00:16:01.489225 env[1323]: time="2025-05-13T00:16:01.489173010Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646 pid=4390 runtime=io.containerd.runc.v2 May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.268 [INFO][4317] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.299 [INFO][4317] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0 calico-apiserver-5b58ff46dd- calico-apiserver 4a3e9432-3790-4f08-9c11-d537f6aab7dd 899 0 2025-05-13 00:15:37 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:5b58ff46dd projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-5b58ff46dd-zk6tg eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali0bed5f3decf [] []}} ContainerID="54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-zk6tg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-" May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.299 [INFO][4317] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-zk6tg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.376 [INFO][4351] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" HandleID="k8s-pod-network.54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.392 [INFO][4351] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" HandleID="k8s-pod-network.54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002fc580), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-5b58ff46dd-zk6tg", "timestamp":"2025-05-13 00:16:01.376891899 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.392 [INFO][4351] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.415 [INFO][4351] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.415 [INFO][4351] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.417 [INFO][4351] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" host="localhost" May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.422 [INFO][4351] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.429 [INFO][4351] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.438 [INFO][4351] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.440 [INFO][4351] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.440 [INFO][4351] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" host="localhost" May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.443 [INFO][4351] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6 May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.447 [INFO][4351] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" host="localhost" May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.460 [INFO][4351] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" host="localhost" May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.460 [INFO][4351] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" host="localhost" May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.460 [INFO][4351] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:01.497917 env[1323]: 2025-05-13 00:16:01.460 [INFO][4351] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" HandleID="k8s-pod-network.54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:01.498520 env[1323]: 2025-05-13 00:16:01.465 [INFO][4317] cni-plugin/k8s.go 386: Populated endpoint ContainerID="54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-zk6tg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0", GenerateName:"calico-apiserver-5b58ff46dd-", Namespace:"calico-apiserver", SelfLink:"", UID:"4a3e9432-3790-4f08-9c11-d537f6aab7dd", ResourceVersion:"899", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5b58ff46dd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-5b58ff46dd-zk6tg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali0bed5f3decf", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:01.498520 env[1323]: 2025-05-13 00:16:01.465 [INFO][4317] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.133/32] ContainerID="54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-zk6tg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:01.498520 env[1323]: 2025-05-13 00:16:01.465 [INFO][4317] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali0bed5f3decf ContainerID="54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-zk6tg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:01.498520 env[1323]: 2025-05-13 00:16:01.473 [INFO][4317] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-zk6tg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:01.498520 env[1323]: 2025-05-13 00:16:01.473 [INFO][4317] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-zk6tg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0", GenerateName:"calico-apiserver-5b58ff46dd-", Namespace:"calico-apiserver", SelfLink:"", UID:"4a3e9432-3790-4f08-9c11-d537f6aab7dd", ResourceVersion:"899", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5b58ff46dd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6", Pod:"calico-apiserver-5b58ff46dd-zk6tg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali0bed5f3decf", MAC:"ae:1f:14:cf:8e:b5", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:01.498520 env[1323]: 2025-05-13 00:16:01.492 [INFO][4317] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6" Namespace="calico-apiserver" Pod="calico-apiserver-5b58ff46dd-zk6tg" WorkloadEndpoint="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:01.510436 systemd[1]: run-containerd-runc-k8s.io-97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646-runc.1TKlV2.mount: Deactivated successfully. May 13 00:16:01.534000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.8:22-10.0.0.1:49658 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:01.535148 systemd[1]: Started sshd@12-10.0.0.8:22-10.0.0.1:49658.service. May 13 00:16:01.547591 kernel: audit: type=1130 audit(1747095361.534:352): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.8:22-10.0.0.1:49658 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:01.549913 env[1323]: time="2025-05-13T00:16:01.542478345Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 13 00:16:01.549913 env[1323]: time="2025-05-13T00:16:01.542529785Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 13 00:16:01.549913 env[1323]: time="2025-05-13T00:16:01.542553905Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 13 00:16:01.549913 env[1323]: time="2025-05-13T00:16:01.542724425Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6 pid=4435 runtime=io.containerd.runc.v2 May 13 00:16:01.550485 systemd-resolved[1238]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 13 00:16:01.584000 audit[4429]: USER_ACCT pid=4429 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:01.589104 sshd[4429]: Accepted publickey for core from 10.0.0.1 port 49658 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:16:01.589640 kernel: audit: type=1101 audit(1747095361.584:353): pid=4429 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:01.590000 audit[4429]: CRED_ACQ pid=4429 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:01.596244 kernel: audit: type=1103 audit(1747095361.590:354): pid=4429 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:01.596305 kernel: audit: type=1006 audit(1747095361.593:355): pid=4429 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=13 res=1 May 13 00:16:01.593000 audit[4429]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffd21f2d60 a2=3 a3=1 items=0 ppid=1 pid=4429 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=13 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:01.593000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:16:01.597074 sshd[4429]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:16:01.597618 env[1323]: time="2025-05-13T00:16:01.597572600Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-k77ll,Uid:943c0c0f-6fce-4911-b87f-c827598e67dc,Namespace:kube-system,Attempt:1,} returns sandbox id \"97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646\"" May 13 00:16:01.598961 kubelet[2230]: E0513 00:16:01.598364 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:16:01.604665 systemd[1]: Started session-13.scope. May 13 00:16:01.605373 systemd-logind[1306]: New session 13 of user core. May 13 00:16:01.606789 systemd-resolved[1238]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 13 00:16:01.611705 env[1323]: time="2025-05-13T00:16:01.611659763Z" level=info msg="CreateContainer within sandbox \"97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 13 00:16:01.614000 audit[4429]: USER_START pid=4429 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:01.616000 audit[4472]: CRED_ACQ pid=4472 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:01.627714 env[1323]: time="2025-05-13T00:16:01.627670128Z" level=info msg="CreateContainer within sandbox \"97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"cb6fbb3994ed7eb87668802eed9a7b635c1b98b70c43a1c2b3226135ff9df0bd\"" May 13 00:16:01.628191 env[1323]: time="2025-05-13T00:16:01.628118168Z" level=info msg="StartContainer for \"cb6fbb3994ed7eb87668802eed9a7b635c1b98b70c43a1c2b3226135ff9df0bd\"" May 13 00:16:01.638392 env[1323]: time="2025-05-13T00:16:01.638335211Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5b58ff46dd-zk6tg,Uid:4a3e9432-3790-4f08-9c11-d537f6aab7dd,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6\"" May 13 00:16:01.642421 env[1323]: time="2025-05-13T00:16:01.642372972Z" level=info msg="CreateContainer within sandbox \"54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 13 00:16:01.666076 env[1323]: time="2025-05-13T00:16:01.666023818Z" level=info msg="CreateContainer within sandbox \"54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"5546b688856bba71ac60ffa012553c0011d9756627e66ee14a4637d5b56453d8\"" May 13 00:16:01.668156 env[1323]: time="2025-05-13T00:16:01.668123179Z" level=info msg="StartContainer for \"5546b688856bba71ac60ffa012553c0011d9756627e66ee14a4637d5b56453d8\"" May 13 00:16:01.740113 env[1323]: time="2025-05-13T00:16:01.740011758Z" level=info msg="StartContainer for \"cb6fbb3994ed7eb87668802eed9a7b635c1b98b70c43a1c2b3226135ff9df0bd\" returns successfully" May 13 00:16:01.790336 env[1323]: time="2025-05-13T00:16:01.790276452Z" level=info msg="StartContainer for \"5546b688856bba71ac60ffa012553c0011d9756627e66ee14a4637d5b56453d8\" returns successfully" May 13 00:16:01.913229 sshd[4429]: pam_unix(sshd:session): session closed for user core May 13 00:16:01.913000 audit[4429]: USER_END pid=4429 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:01.913000 audit[4429]: CRED_DISP pid=4429 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:01.916461 systemd[1]: sshd@12-10.0.0.8:22-10.0.0.1:49658.service: Deactivated successfully. May 13 00:16:01.916000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.8:22-10.0.0.1:49658 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:01.917232 systemd-logind[1306]: Session 13 logged out. Waiting for processes to exit. May 13 00:16:01.917962 systemd[1]: session-13.scope: Deactivated successfully. May 13 00:16:01.918516 systemd-logind[1306]: Removed session 13. May 13 00:16:02.131109 env[1323]: time="2025-05-13T00:16:02.131072303Z" level=info msg="StopPodSandbox for \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\"" May 13 00:16:02.142853 env[1323]: time="2025-05-13T00:16:02.142760066Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/csi:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:02.173225 env[1323]: time="2025-05-13T00:16:02.173184753Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:15faf29e8b518d846c91c15785ff89e783d356ea0f2b22826f47a556ea32645b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:02.186582 env[1323]: time="2025-05-13T00:16:02.186530477Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/csi:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:02.203955 env[1323]: time="2025-05-13T00:16:02.203910521Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:02.204240 env[1323]: time="2025-05-13T00:16:02.204209401Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\" returns image reference \"sha256:15faf29e8b518d846c91c15785ff89e783d356ea0f2b22826f47a556ea32645b\"" May 13 00:16:02.209062 env[1323]: time="2025-05-13T00:16:02.209021363Z" level=info msg="CreateContainer within sandbox \"ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" May 13 00:16:02.288785 kubelet[2230]: E0513 00:16:02.288752 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:16:02.292132 kubelet[2230]: I0513 00:16:02.292098 2230 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 13 00:16:02.346892 systemd-networkd[1102]: calia028079e884: Gained IPv6LL May 13 00:16:02.364947 env[1323]: 2025-05-13 00:16:02.211 [INFO][4603] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" May 13 00:16:02.364947 env[1323]: 2025-05-13 00:16:02.211 [INFO][4603] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" iface="eth0" netns="/var/run/netns/cni-1f04b35d-9603-7d3f-a3a7-12470ac43be2" May 13 00:16:02.364947 env[1323]: 2025-05-13 00:16:02.212 [INFO][4603] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" iface="eth0" netns="/var/run/netns/cni-1f04b35d-9603-7d3f-a3a7-12470ac43be2" May 13 00:16:02.364947 env[1323]: 2025-05-13 00:16:02.212 [INFO][4603] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" iface="eth0" netns="/var/run/netns/cni-1f04b35d-9603-7d3f-a3a7-12470ac43be2" May 13 00:16:02.364947 env[1323]: 2025-05-13 00:16:02.212 [INFO][4603] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" May 13 00:16:02.364947 env[1323]: 2025-05-13 00:16:02.212 [INFO][4603] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" May 13 00:16:02.364947 env[1323]: 2025-05-13 00:16:02.258 [INFO][4613] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" HandleID="k8s-pod-network.37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" Workload="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:02.364947 env[1323]: 2025-05-13 00:16:02.258 [INFO][4613] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:02.364947 env[1323]: 2025-05-13 00:16:02.258 [INFO][4613] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:02.364947 env[1323]: 2025-05-13 00:16:02.351 [WARNING][4613] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" HandleID="k8s-pod-network.37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" Workload="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:02.364947 env[1323]: 2025-05-13 00:16:02.352 [INFO][4613] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" HandleID="k8s-pod-network.37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" Workload="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:02.364947 env[1323]: 2025-05-13 00:16:02.361 [INFO][4613] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:02.364947 env[1323]: 2025-05-13 00:16:02.363 [INFO][4603] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" May 13 00:16:02.365583 env[1323]: time="2025-05-13T00:16:02.365068083Z" level=info msg="TearDown network for sandbox \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\" successfully" May 13 00:16:02.365583 env[1323]: time="2025-05-13T00:16:02.365099283Z" level=info msg="StopPodSandbox for \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\" returns successfully" May 13 00:16:02.365713 env[1323]: time="2025-05-13T00:16:02.365678883Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-66bcdc688d-zm28k,Uid:47f5ba25-f173-43e0-af82-849feb0085ec,Namespace:calico-system,Attempt:1,}" May 13 00:16:02.427000 audit[4623]: NETFILTER_CFG table=filter:101 family=2 entries=12 op=nft_register_rule pid=4623 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:02.427000 audit[4623]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=4420 a0=3 a1=fffff7eb80a0 a2=0 a3=1 items=0 ppid=2393 pid=4623 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:02.427000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:02.436000 audit[4623]: NETFILTER_CFG table=nat:102 family=2 entries=42 op=nft_register_rule pid=4623 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:02.436000 audit[4623]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=13428 a0=3 a1=fffff7eb80a0 a2=0 a3=1 items=0 ppid=2393 pid=4623 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:02.436000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:02.458074 systemd[1]: run-netns-cni\x2d1f04b35d\x2d9603\x2d7d3f\x2da3a7\x2d12470ac43be2.mount: Deactivated successfully. May 13 00:16:02.519967 kubelet[2230]: I0513 00:16:02.519908 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-k77ll" podStartSLOduration=31.519891962 podStartE2EDuration="31.519891962s" podCreationTimestamp="2025-05-13 00:15:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-13 00:16:02.380080366 +0000 UTC m=+46.358303114" watchObservedRunningTime="2025-05-13 00:16:02.519891962 +0000 UTC m=+46.498114710" May 13 00:16:02.521202 kubelet[2230]: I0513 00:16:02.521160 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-5b58ff46dd-zk6tg" podStartSLOduration=25.521149642 podStartE2EDuration="25.521149642s" podCreationTimestamp="2025-05-13 00:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-13 00:16:02.519859122 +0000 UTC m=+46.498081910" watchObservedRunningTime="2025-05-13 00:16:02.521149642 +0000 UTC m=+46.499372350" May 13 00:16:02.531000 audit[4625]: NETFILTER_CFG table=filter:103 family=2 entries=12 op=nft_register_rule pid=4625 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:02.531000 audit[4625]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=4420 a0=3 a1=ffffd0998400 a2=0 a3=1 items=0 ppid=2393 pid=4625 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:02.531000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:02.535000 audit[4625]: NETFILTER_CFG table=nat:104 family=2 entries=18 op=nft_register_rule pid=4625 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:02.535000 audit[4625]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=5004 a0=3 a1=ffffd0998400 a2=0 a3=1 items=0 ppid=2393 pid=4625 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:02.535000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:02.627992 env[1323]: time="2025-05-13T00:16:02.627938510Z" level=info msg="CreateContainer within sandbox \"ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"66ee9b82d53a177c83fc0e5def9be4c933430a1d822b5832d452936cb2492247\"" May 13 00:16:02.628644 env[1323]: time="2025-05-13T00:16:02.628618550Z" level=info msg="StartContainer for \"66ee9b82d53a177c83fc0e5def9be4c933430a1d822b5832d452936cb2492247\"" May 13 00:16:02.766425 env[1323]: time="2025-05-13T00:16:02.766310465Z" level=info msg="StartContainer for \"66ee9b82d53a177c83fc0e5def9be4c933430a1d822b5832d452936cb2492247\" returns successfully" May 13 00:16:02.768669 env[1323]: time="2025-05-13T00:16:02.768632586Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\"" May 13 00:16:02.830263 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready May 13 00:16:02.830395 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): calic35c4880c53: link becomes ready May 13 00:16:02.828647 systemd-networkd[1102]: calic35c4880c53: Link UP May 13 00:16:02.829247 systemd-networkd[1102]: calic35c4880c53: Gained carrier May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.697 [INFO][4648] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.718 [INFO][4648] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0 calico-kube-controllers-66bcdc688d- calico-system 47f5ba25-f173-43e0-af82-849feb0085ec 921 0 2025-05-13 00:15:38 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:66bcdc688d projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-66bcdc688d-zm28k eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calic35c4880c53 [] []}} ContainerID="d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" Namespace="calico-system" Pod="calico-kube-controllers-66bcdc688d-zm28k" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-" May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.718 [INFO][4648] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" Namespace="calico-system" Pod="calico-kube-controllers-66bcdc688d-zm28k" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.778 [INFO][4668] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" HandleID="k8s-pod-network.d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" Workload="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.790 [INFO][4668] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" HandleID="k8s-pod-network.d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" Workload="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004c3d0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-66bcdc688d-zm28k", "timestamp":"2025-05-13 00:16:02.778328348 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.790 [INFO][4668] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.790 [INFO][4668] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.790 [INFO][4668] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.792 [INFO][4668] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" host="localhost" May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.798 [INFO][4668] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.804 [INFO][4668] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.807 [INFO][4668] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.810 [INFO][4668] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.810 [INFO][4668] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" host="localhost" May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.812 [INFO][4668] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42 May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.816 [INFO][4668] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" host="localhost" May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.823 [INFO][4668] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" host="localhost" May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.823 [INFO][4668] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" host="localhost" May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.824 [INFO][4668] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:02.871673 env[1323]: 2025-05-13 00:16:02.824 [INFO][4668] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" HandleID="k8s-pod-network.d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" Workload="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:02.872348 env[1323]: 2025-05-13 00:16:02.825 [INFO][4648] cni-plugin/k8s.go 386: Populated endpoint ContainerID="d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" Namespace="calico-system" Pod="calico-kube-controllers-66bcdc688d-zm28k" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0", GenerateName:"calico-kube-controllers-66bcdc688d-", Namespace:"calico-system", SelfLink:"", UID:"47f5ba25-f173-43e0-af82-849feb0085ec", ResourceVersion:"921", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"66bcdc688d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-66bcdc688d-zm28k", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calic35c4880c53", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:02.872348 env[1323]: 2025-05-13 00:16:02.826 [INFO][4648] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.134/32] ContainerID="d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" Namespace="calico-system" Pod="calico-kube-controllers-66bcdc688d-zm28k" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:02.872348 env[1323]: 2025-05-13 00:16:02.826 [INFO][4648] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic35c4880c53 ContainerID="d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" Namespace="calico-system" Pod="calico-kube-controllers-66bcdc688d-zm28k" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:02.872348 env[1323]: 2025-05-13 00:16:02.829 [INFO][4648] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" Namespace="calico-system" Pod="calico-kube-controllers-66bcdc688d-zm28k" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:02.872348 env[1323]: 2025-05-13 00:16:02.831 [INFO][4648] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" Namespace="calico-system" Pod="calico-kube-controllers-66bcdc688d-zm28k" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0", GenerateName:"calico-kube-controllers-66bcdc688d-", Namespace:"calico-system", SelfLink:"", UID:"47f5ba25-f173-43e0-af82-849feb0085ec", ResourceVersion:"921", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"66bcdc688d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42", Pod:"calico-kube-controllers-66bcdc688d-zm28k", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calic35c4880c53", MAC:"9a:7a:bb:d1:df:33", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:02.872348 env[1323]: 2025-05-13 00:16:02.855 [INFO][4648] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42" Namespace="calico-system" Pod="calico-kube-controllers-66bcdc688d-zm28k" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:02.893673 env[1323]: time="2025-05-13T00:16:02.893585698Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 13 00:16:02.893673 env[1323]: time="2025-05-13T00:16:02.893627818Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 13 00:16:02.893673 env[1323]: time="2025-05-13T00:16:02.893638578Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 13 00:16:02.893864 env[1323]: time="2025-05-13T00:16:02.893766178Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42 pid=4700 runtime=io.containerd.runc.v2 May 13 00:16:02.944647 systemd-resolved[1238]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 13 00:16:02.969012 env[1323]: time="2025-05-13T00:16:02.968963157Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-66bcdc688d-zm28k,Uid:47f5ba25-f173-43e0-af82-849feb0085ec,Namespace:calico-system,Attempt:1,} returns sandbox id \"d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42\"" May 13 00:16:03.296759 kubelet[2230]: E0513 00:16:03.296245 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:16:03.306691 systemd-networkd[1102]: cali0bed5f3decf: Gained IPv6LL May 13 00:16:03.434686 systemd-networkd[1102]: cali99ccf0c6009: Gained IPv6LL May 13 00:16:03.450000 audit[4758]: NETFILTER_CFG table=filter:105 family=2 entries=11 op=nft_register_rule pid=4758 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:03.450000 audit[4758]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=3676 a0=3 a1=fffff6c6afe0 a2=0 a3=1 items=0 ppid=2393 pid=4758 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:03.450000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:03.476000 audit[4758]: NETFILTER_CFG table=nat:106 family=2 entries=61 op=nft_register_chain pid=4758 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:03.476000 audit[4758]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=22668 a0=3 a1=fffff6c6afe0 a2=0 a3=1 items=0 ppid=2393 pid=4758 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:03.476000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:03.899669 env[1323]: time="2025-05-13T00:16:03.899626100Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:03.901312 env[1323]: time="2025-05-13T00:16:03.901273141Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:a91b1f00752edc175f270a01b33683fa80818734aa2274388785eaf3364315dc,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:03.902625 env[1323]: time="2025-05-13T00:16:03.902592861Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:03.904398 env[1323]: time="2025-05-13T00:16:03.904370901Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:03.904921 env[1323]: time="2025-05-13T00:16:03.904893342Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" returns image reference \"sha256:a91b1f00752edc175f270a01b33683fa80818734aa2274388785eaf3364315dc\"" May 13 00:16:03.906864 env[1323]: time="2025-05-13T00:16:03.906828942Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\"" May 13 00:16:03.908719 env[1323]: time="2025-05-13T00:16:03.908689702Z" level=info msg="CreateContainer within sandbox \"ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" May 13 00:16:03.924610 env[1323]: time="2025-05-13T00:16:03.924575666Z" level=info msg="CreateContainer within sandbox \"ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"fa7c5b30a7ed34bbbcfceaec0e478f2b10450b388c6816bce333af26b4595981\"" May 13 00:16:03.925165 env[1323]: time="2025-05-13T00:16:03.925142706Z" level=info msg="StartContainer for \"fa7c5b30a7ed34bbbcfceaec0e478f2b10450b388c6816bce333af26b4595981\"" May 13 00:16:04.001446 env[1323]: time="2025-05-13T00:16:04.001397765Z" level=info msg="StartContainer for \"fa7c5b30a7ed34bbbcfceaec0e478f2b10450b388c6816bce333af26b4595981\" returns successfully" May 13 00:16:04.074654 systemd-networkd[1102]: calic35c4880c53: Gained IPv6LL May 13 00:16:04.202777 kubelet[2230]: I0513 00:16:04.202675 2230 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 May 13 00:16:04.204665 kubelet[2230]: I0513 00:16:04.204638 2230 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock May 13 00:16:04.300606 kubelet[2230]: E0513 00:16:04.300576 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:16:04.313444 kubelet[2230]: I0513 00:16:04.313382 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-srb86" podStartSLOduration=22.923221629 podStartE2EDuration="26.313305755s" podCreationTimestamp="2025-05-13 00:15:38 +0000 UTC" firstStartedPulling="2025-05-13 00:16:00.516056496 +0000 UTC m=+44.494279244" lastFinishedPulling="2025-05-13 00:16:03.906140542 +0000 UTC m=+47.884363370" observedRunningTime="2025-05-13 00:16:04.313188155 +0000 UTC m=+48.291410903" watchObservedRunningTime="2025-05-13 00:16:04.313305755 +0000 UTC m=+48.291528503" May 13 00:16:04.457081 systemd[1]: run-containerd-runc-k8s.io-fa7c5b30a7ed34bbbcfceaec0e478f2b10450b388c6816bce333af26b4595981-runc.AykeDh.mount: Deactivated successfully. May 13 00:16:05.577638 env[1323]: time="2025-05-13T00:16:05.577596151Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/kube-controllers:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:05.578904 env[1323]: time="2025-05-13T00:16:05.578870591Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ec7c64189a2fd01b24b044fea1840d441e9884a0df32c2e9d6982cfbbea1f814,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:05.580822 env[1323]: time="2025-05-13T00:16:05.580786391Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/kube-controllers:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:05.582045 env[1323]: time="2025-05-13T00:16:05.582020432Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 13 00:16:05.582522 env[1323]: time="2025-05-13T00:16:05.582493392Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" returns image reference \"sha256:ec7c64189a2fd01b24b044fea1840d441e9884a0df32c2e9d6982cfbbea1f814\"" May 13 00:16:05.595147 env[1323]: time="2025-05-13T00:16:05.595107634Z" level=info msg="CreateContainer within sandbox \"d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" May 13 00:16:05.603957 env[1323]: time="2025-05-13T00:16:05.603921316Z" level=info msg="CreateContainer within sandbox \"d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"916672c254d91b42e4bf7e4c911fb0dfeec25345fa0fe604ad9d8326d70997ef\"" May 13 00:16:05.605807 env[1323]: time="2025-05-13T00:16:05.605774917Z" level=info msg="StartContainer for \"916672c254d91b42e4bf7e4c911fb0dfeec25345fa0fe604ad9d8326d70997ef\"" May 13 00:16:05.677146 env[1323]: time="2025-05-13T00:16:05.677095892Z" level=info msg="StartContainer for \"916672c254d91b42e4bf7e4c911fb0dfeec25345fa0fe604ad9d8326d70997ef\" returns successfully" May 13 00:16:06.317052 kubelet[2230]: I0513 00:16:06.316997 2230 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-66bcdc688d-zm28k" podStartSLOduration=25.703771867 podStartE2EDuration="28.316973622s" podCreationTimestamp="2025-05-13 00:15:38 +0000 UTC" firstStartedPulling="2025-05-13 00:16:02.970300117 +0000 UTC m=+46.948522865" lastFinishedPulling="2025-05-13 00:16:05.583501912 +0000 UTC m=+49.561724620" observedRunningTime="2025-05-13 00:16:06.316556782 +0000 UTC m=+50.294779530" watchObservedRunningTime="2025-05-13 00:16:06.316973622 +0000 UTC m=+50.295196370" May 13 00:16:06.916000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.8:22-10.0.0.1:53072 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:06.916646 systemd[1]: Started sshd@13-10.0.0.8:22-10.0.0.1:53072.service. May 13 00:16:06.920855 kernel: kauditd_printk_skb: 25 callbacks suppressed May 13 00:16:06.920965 kernel: audit: type=1130 audit(1747095366.916:367): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.8:22-10.0.0.1:53072 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:06.956000 audit[4930]: USER_ACCT pid=4930 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:06.957236 sshd[4930]: Accepted publickey for core from 10.0.0.1 port 53072 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:16:06.959412 sshd[4930]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:16:06.958000 audit[4930]: CRED_ACQ pid=4930 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:06.964056 kernel: audit: type=1101 audit(1747095366.956:368): pid=4930 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:06.964141 kernel: audit: type=1103 audit(1747095366.958:369): pid=4930 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:06.964168 kernel: audit: type=1006 audit(1747095366.958:370): pid=4930 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=14 res=1 May 13 00:16:06.963814 systemd-logind[1306]: New session 14 of user core. May 13 00:16:06.964524 systemd[1]: Started session-14.scope. May 13 00:16:06.958000 audit[4930]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffffe876680 a2=3 a3=1 items=0 ppid=1 pid=4930 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:06.970731 kernel: audit: type=1300 audit(1747095366.958:370): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffffe876680 a2=3 a3=1 items=0 ppid=1 pid=4930 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:06.970809 kernel: audit: type=1327 audit(1747095366.958:370): proctitle=737368643A20636F7265205B707269765D May 13 00:16:06.958000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:16:06.968000 audit[4930]: USER_START pid=4930 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:06.975299 kernel: audit: type=1105 audit(1747095366.968:371): pid=4930 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:06.975394 kernel: audit: type=1103 audit(1747095366.969:372): pid=4933 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:06.969000 audit[4933]: CRED_ACQ pid=4933 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:07.126885 sshd[4930]: pam_unix(sshd:session): session closed for user core May 13 00:16:07.127000 audit[4930]: USER_END pid=4930 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:07.130506 systemd[1]: sshd@13-10.0.0.8:22-10.0.0.1:53072.service: Deactivated successfully. May 13 00:16:07.131498 systemd-logind[1306]: Session 14 logged out. Waiting for processes to exit. May 13 00:16:07.131575 systemd[1]: session-14.scope: Deactivated successfully. May 13 00:16:07.127000 audit[4930]: CRED_DISP pid=4930 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:07.132520 systemd-logind[1306]: Removed session 14. May 13 00:16:07.135936 kernel: audit: type=1106 audit(1747095367.127:373): pid=4930 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:07.136032 kernel: audit: type=1104 audit(1747095367.127:374): pid=4930 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:07.127000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.8:22-10.0.0.1:53072 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:08.244353 kubelet[2230]: I0513 00:16:08.244316 2230 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 13 00:16:08.245396 kubelet[2230]: E0513 00:16:08.245372 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:16:08.280000 audit[4980]: NETFILTER_CFG table=filter:107 family=2 entries=9 op=nft_register_rule pid=4980 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:08.280000 audit[4980]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2932 a0=3 a1=ffffcf240820 a2=0 a3=1 items=0 ppid=2393 pid=4980 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.280000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:08.289000 audit[4980]: NETFILTER_CFG table=nat:108 family=2 entries=27 op=nft_register_chain pid=4980 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:08.289000 audit[4980]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=9348 a0=3 a1=ffffcf240820 a2=0 a3=1 items=0 ppid=2393 pid=4980 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.289000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:08.308677 kubelet[2230]: E0513 00:16:08.308531 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:16:08.782000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit: BPF prog-id=10 op=LOAD May 13 00:16:08.782000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=fffff56a6c68 a2=98 a3=fffff56a6c58 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.782000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.782000 audit: BPF prog-id=10 op=UNLOAD May 13 00:16:08.782000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit: BPF prog-id=11 op=LOAD May 13 00:16:08.782000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=5 a1=fffff56a68f8 a2=74 a3=95 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.782000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.782000 audit: BPF prog-id=11 op=UNLOAD May 13 00:16:08.782000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.782000 audit: BPF prog-id=12 op=LOAD May 13 00:16:08.782000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=5 a1=fffff56a6958 a2=94 a3=2 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.782000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.782000 audit: BPF prog-id=12 op=UNLOAD May 13 00:16:08.873000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.873000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.873000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.873000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.873000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.873000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.873000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.873000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.873000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.873000 audit: BPF prog-id=13 op=LOAD May 13 00:16:08.873000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=5 a1=fffff56a6918 a2=40 a3=fffff56a6948 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.873000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.874000 audit: BPF prog-id=13 op=UNLOAD May 13 00:16:08.874000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.874000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=0 a1=fffff56a6a30 a2=50 a3=0 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.874000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=fffff56a6988 a2=28 a3=fffff56a6ab8 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff56a69b8 a2=28 a3=fffff56a6ae8 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff56a6868 a2=28 a3=fffff56a6998 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=fffff56a69d8 a2=28 a3=fffff56a6b08 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=fffff56a69b8 a2=28 a3=fffff56a6ae8 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=fffff56a69a8 a2=28 a3=fffff56a6ad8 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=fffff56a69d8 a2=28 a3=fffff56a6b08 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff56a69b8 a2=28 a3=fffff56a6ae8 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff56a69d8 a2=28 a3=fffff56a6b08 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff56a69a8 a2=28 a3=fffff56a6ad8 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=fffff56a6a28 a2=28 a3=fffff56a6b68 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=5 a0=0 a1=fffff56a6760 a2=50 a3=0 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit: BPF prog-id=14 op=LOAD May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=6 a0=5 a1=fffff56a6768 a2=94 a3=5 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit: BPF prog-id=14 op=UNLOAD May 13 00:16:08.883000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=5 a0=0 a1=fffff56a6870 a2=50 a3=0 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=0 a0=16 a1=fffff56a69b8 a2=4 a3=3 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.883000 audit[5014]: AVC avc: denied { confidentiality } for pid=5014 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 May 13 00:16:08.883000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=5 a1=fffff56a6998 a2=94 a3=6 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.883000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.884000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { confidentiality } for pid=5014 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 May 13 00:16:08.884000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=5 a1=fffff56a6168 a2=94 a3=83 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.884000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.884000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { perfmon } for pid=5014 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { bpf } for pid=5014 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.884000 audit[5014]: AVC avc: denied { confidentiality } for pid=5014 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 May 13 00:16:08.884000 audit[5014]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=5 a1=fffff56a6168 a2=94 a3=83 items=0 ppid=4997 pid=5014 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.884000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 13 00:16:08.897000 audit[5017]: AVC avc: denied { bpf } for pid=5017 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { bpf } for pid=5017 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { bpf } for pid=5017 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { bpf } for pid=5017 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit: BPF prog-id=15 op=LOAD May 13 00:16:08.897000 audit[5017]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=ffffe2e0cae8 a2=98 a3=ffffe2e0cad8 items=0 ppid=4997 pid=5017 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.897000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F May 13 00:16:08.897000 audit: BPF prog-id=15 op=UNLOAD May 13 00:16:08.897000 audit[5017]: AVC avc: denied { bpf } for pid=5017 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { bpf } for pid=5017 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { bpf } for pid=5017 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit[5017]: AVC avc: denied { bpf } for pid=5017 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.897000 audit: BPF prog-id=16 op=LOAD May 13 00:16:08.897000 audit[5017]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=ffffe2e0c998 a2=74 a3=95 items=0 ppid=4997 pid=5017 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.897000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F May 13 00:16:08.898000 audit: BPF prog-id=16 op=UNLOAD May 13 00:16:08.898000 audit[5017]: AVC avc: denied { bpf } for pid=5017 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.898000 audit[5017]: AVC avc: denied { bpf } for pid=5017 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.898000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.898000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.898000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.898000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.898000 audit[5017]: AVC avc: denied { perfmon } for pid=5017 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.898000 audit[5017]: AVC avc: denied { bpf } for pid=5017 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.898000 audit[5017]: AVC avc: denied { bpf } for pid=5017 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:08.898000 audit: BPF prog-id=17 op=LOAD May 13 00:16:08.898000 audit[5017]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=ffffe2e0c9c8 a2=40 a3=ffffe2e0c9f8 items=0 ppid=4997 pid=5017 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:08.898000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F May 13 00:16:08.898000 audit: BPF prog-id=17 op=UNLOAD May 13 00:16:08.976434 systemd-networkd[1102]: vxlan.calico: Link UP May 13 00:16:08.976442 systemd-networkd[1102]: vxlan.calico: Gained carrier May 13 00:16:09.003000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.003000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.003000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.003000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.003000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.003000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.003000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.003000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.003000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.003000 audit: BPF prog-id=18 op=LOAD May 13 00:16:09.003000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=fffff42b18e8 a2=98 a3=fffff42b18d8 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.003000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.003000 audit: BPF prog-id=18 op=UNLOAD May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit: BPF prog-id=19 op=LOAD May 13 00:16:09.004000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=fffff42b15c8 a2=74 a3=95 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.004000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.004000 audit: BPF prog-id=19 op=UNLOAD May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit: BPF prog-id=20 op=LOAD May 13 00:16:09.004000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=fffff42b1628 a2=94 a3=2 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.004000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.004000 audit: BPF prog-id=20 op=UNLOAD May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=12 a1=fffff42b1658 a2=28 a3=fffff42b1788 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.004000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff42b1688 a2=28 a3=fffff42b17b8 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.004000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff42b1538 a2=28 a3=fffff42b1668 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.004000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=12 a1=fffff42b16a8 a2=28 a3=fffff42b17d8 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.004000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=12 a1=fffff42b1688 a2=28 a3=fffff42b17b8 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.004000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=12 a1=fffff42b1678 a2=28 a3=fffff42b17a8 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.004000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=12 a1=fffff42b16a8 a2=28 a3=fffff42b17d8 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.004000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff42b1688 a2=28 a3=fffff42b17b8 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.004000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff42b16a8 a2=28 a3=fffff42b17d8 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.004000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff42b1678 a2=28 a3=fffff42b17a8 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.004000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=12 a1=fffff42b16f8 a2=28 a3=fffff42b1838 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.004000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.004000 audit: BPF prog-id=21 op=LOAD May 13 00:16:09.004000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=6 a0=5 a1=fffff42b1518 a2=40 a3=fffff42b1548 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.004000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.004000 audit: BPF prog-id=21 op=UNLOAD May 13 00:16:09.005000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=0 a1=fffff42b1540 a2=50 a3=0 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.005000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=6 a0=0 a1=fffff42b1540 a2=50 a3=0 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.005000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit: BPF prog-id=22 op=LOAD May 13 00:16:09.005000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=6 a0=5 a1=fffff42b0ca8 a2=94 a3=2 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.005000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.005000 audit: BPF prog-id=22 op=UNLOAD May 13 00:16:09.005000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { perfmon } for pid=5047 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit[5047]: AVC avc: denied { bpf } for pid=5047 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.005000 audit: BPF prog-id=23 op=LOAD May 13 00:16:09.005000 audit[5047]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=6 a0=5 a1=fffff42b0e38 a2=94 a3=30 items=0 ppid=4997 pid=5047 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.005000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit: BPF prog-id=24 op=LOAD May 13 00:16:09.017000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=fffff1f830e8 a2=98 a3=fffff1f830d8 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.017000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.017000 audit: BPF prog-id=24 op=UNLOAD May 13 00:16:09.017000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.017000 audit: BPF prog-id=25 op=LOAD May 13 00:16:09.017000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=5 a1=fffff1f82d78 a2=74 a3=95 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.017000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.018000 audit: BPF prog-id=25 op=UNLOAD May 13 00:16:09.018000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.018000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.018000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.018000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.018000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.018000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.018000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.018000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.018000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.018000 audit: BPF prog-id=26 op=LOAD May 13 00:16:09.018000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=5 a1=fffff1f82dd8 a2=94 a3=2 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.018000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.018000 audit: BPF prog-id=26 op=UNLOAD May 13 00:16:09.112000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.112000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.112000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.112000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.112000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.112000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.112000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.112000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.112000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.112000 audit: BPF prog-id=27 op=LOAD May 13 00:16:09.112000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=5 a1=fffff1f82d98 a2=40 a3=fffff1f82dc8 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.112000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.112000 audit: BPF prog-id=27 op=UNLOAD May 13 00:16:09.112000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.112000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=0 a1=fffff1f82eb0 a2=50 a3=0 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.112000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.124000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.124000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=fffff1f82e08 a2=28 a3=fffff1f82f38 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.124000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.124000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.124000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff1f82e38 a2=28 a3=fffff1f82f68 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.124000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.124000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.124000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff1f82ce8 a2=28 a3=fffff1f82e18 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.124000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.124000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.124000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=fffff1f82e58 a2=28 a3=fffff1f82f88 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.124000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.124000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.124000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=fffff1f82e38 a2=28 a3=fffff1f82f68 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.124000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.124000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.124000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=fffff1f82e28 a2=28 a3=fffff1f82f58 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.124000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.124000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.124000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=fffff1f82e58 a2=28 a3=fffff1f82f88 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.124000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.124000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.124000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff1f82e38 a2=28 a3=fffff1f82f68 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.124000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.124000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.124000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff1f82e58 a2=28 a3=fffff1f82f88 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.124000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.124000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.124000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff1f82e28 a2=28 a3=fffff1f82f58 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.124000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.124000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.124000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=fffff1f82ea8 a2=28 a3=fffff1f82fe8 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.124000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=5 a0=0 a1=fffff1f82be0 a2=50 a3=0 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.125000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit: BPF prog-id=28 op=LOAD May 13 00:16:09.125000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=6 a0=5 a1=fffff1f82be8 a2=94 a3=5 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.125000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.125000 audit: BPF prog-id=28 op=UNLOAD May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=5 a0=0 a1=fffff1f82cf0 a2=50 a3=0 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.125000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=0 a0=16 a1=fffff1f82e38 a2=4 a3=3 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.125000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { confidentiality } for pid=5056 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 May 13 00:16:09.125000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=5 a1=fffff1f82e18 a2=94 a3=6 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.125000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { confidentiality } for pid=5056 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 May 13 00:16:09.125000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=5 a1=fffff1f825e8 a2=94 a3=83 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.125000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { perfmon } for pid=5056 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.125000 audit[5056]: AVC avc: denied { confidentiality } for pid=5056 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 May 13 00:16:09.125000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=5 a1=fffff1f825e8 a2=94 a3=83 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.125000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.126000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.126000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=0 a0=f a1=fffff1f84028 a2=10 a3=fffff1f84118 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.126000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.126000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.126000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=0 a0=f a1=fffff1f83ee8 a2=10 a3=fffff1f83fd8 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.126000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.126000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.126000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=0 a0=f a1=fffff1f83e58 a2=10 a3=fffff1f83fd8 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.126000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.126000 audit[5056]: AVC avc: denied { bpf } for pid=5056 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 13 00:16:09.126000 audit[5056]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=0 a0=f a1=fffff1f83e58 a2=10 a3=fffff1f83fd8 items=0 ppid=4997 pid=5056 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.126000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 13 00:16:09.134000 audit: BPF prog-id=23 op=UNLOAD May 13 00:16:09.204000 audit[5107]: NETFILTER_CFG table=nat:109 family=2 entries=15 op=nft_register_chain pid=5107 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" May 13 00:16:09.204000 audit[5107]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=5084 a0=3 a1=ffffd0677fc0 a2=0 a3=ffff9d902fa8 items=0 ppid=4997 pid=5107 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.204000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 May 13 00:16:09.207000 audit[5111]: NETFILTER_CFG table=mangle:110 family=2 entries=16 op=nft_register_chain pid=5111 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" May 13 00:16:09.207000 audit[5111]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=6868 a0=3 a1=ffffdb5bf450 a2=0 a3=ffffa3825fa8 items=0 ppid=4997 pid=5111 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.207000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 May 13 00:16:09.211000 audit[5106]: NETFILTER_CFG table=raw:111 family=2 entries=21 op=nft_register_chain pid=5106 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" May 13 00:16:09.211000 audit[5106]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=8452 a0=3 a1=ffffd9296cc0 a2=0 a3=ffffa695cfa8 items=0 ppid=4997 pid=5106 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.211000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 May 13 00:16:09.215000 audit[5110]: NETFILTER_CFG table=filter:112 family=2 entries=209 op=nft_register_chain pid=5110 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" May 13 00:16:09.215000 audit[5110]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=122920 a0=3 a1=fffffb901700 a2=0 a3=ffffa1ee5fa8 items=0 ppid=4997 pid=5110 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:09.215000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 May 13 00:16:10.858646 systemd-networkd[1102]: vxlan.calico: Gained IPv6LL May 13 00:16:12.130000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.8:22-10.0.0.1:53078 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:12.130422 systemd[1]: Started sshd@14-10.0.0.8:22-10.0.0.1:53078.service. May 13 00:16:12.134269 kernel: kauditd_printk_skb: 487 callbacks suppressed May 13 00:16:12.134358 kernel: audit: type=1130 audit(1747095372.130:473): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.8:22-10.0.0.1:53078 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:12.172000 audit[5146]: USER_ACCT pid=5146 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:12.173354 sshd[5146]: Accepted publickey for core from 10.0.0.1 port 53078 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:16:12.175355 sshd[5146]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:16:12.176658 kernel: audit: type=1101 audit(1747095372.172:474): pid=5146 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:12.176741 kernel: audit: type=1103 audit(1747095372.174:475): pid=5146 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:12.174000 audit[5146]: CRED_ACQ pid=5146 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:12.182082 kernel: audit: type=1006 audit(1747095372.174:476): pid=5146 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=15 res=1 May 13 00:16:12.182157 kernel: audit: type=1300 audit(1747095372.174:476): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffebd31d50 a2=3 a3=1 items=0 ppid=1 pid=5146 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:12.174000 audit[5146]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffebd31d50 a2=3 a3=1 items=0 ppid=1 pid=5146 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:12.182025 systemd-logind[1306]: New session 15 of user core. May 13 00:16:12.182407 systemd[1]: Started session-15.scope. May 13 00:16:12.174000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:16:12.188184 kernel: audit: type=1327 audit(1747095372.174:476): proctitle=737368643A20636F7265205B707269765D May 13 00:16:12.186000 audit[5146]: USER_START pid=5146 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:12.192236 kernel: audit: type=1105 audit(1747095372.186:477): pid=5146 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:12.188000 audit[5149]: CRED_ACQ pid=5149 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:12.195328 kernel: audit: type=1103 audit(1747095372.188:478): pid=5149 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:12.347013 sshd[5146]: pam_unix(sshd:session): session closed for user core May 13 00:16:12.347000 audit[5146]: USER_END pid=5146 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:12.350024 systemd[1]: sshd@14-10.0.0.8:22-10.0.0.1:53078.service: Deactivated successfully. May 13 00:16:12.350888 systemd[1]: session-15.scope: Deactivated successfully. May 13 00:16:12.348000 audit[5146]: CRED_DISP pid=5146 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:12.356233 kernel: audit: type=1106 audit(1747095372.347:479): pid=5146 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:12.356315 kernel: audit: type=1104 audit(1747095372.348:480): pid=5146 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:12.349000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.8:22-10.0.0.1:53078 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:12.356439 systemd-logind[1306]: Session 15 logged out. Waiting for processes to exit. May 13 00:16:12.357294 systemd-logind[1306]: Removed session 15. May 13 00:16:16.108128 env[1323]: time="2025-05-13T00:16:16.107725073Z" level=info msg="StopPodSandbox for \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\"" May 13 00:16:16.200245 env[1323]: 2025-05-13 00:16:16.149 [WARNING][5183] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"943c0c0f-6fce-4911-b87f-c827598e67dc", ResourceVersion:"930", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646", Pod:"coredns-7db6d8ff4d-k77ll", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali99ccf0c6009", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:16.200245 env[1323]: 2025-05-13 00:16:16.150 [INFO][5183] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" May 13 00:16:16.200245 env[1323]: 2025-05-13 00:16:16.150 [INFO][5183] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" iface="eth0" netns="" May 13 00:16:16.200245 env[1323]: 2025-05-13 00:16:16.150 [INFO][5183] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" May 13 00:16:16.200245 env[1323]: 2025-05-13 00:16:16.150 [INFO][5183] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" May 13 00:16:16.200245 env[1323]: 2025-05-13 00:16:16.182 [INFO][5193] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" HandleID="k8s-pod-network.663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" Workload="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:16.200245 env[1323]: 2025-05-13 00:16:16.183 [INFO][5193] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:16.200245 env[1323]: 2025-05-13 00:16:16.183 [INFO][5193] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:16.200245 env[1323]: 2025-05-13 00:16:16.193 [WARNING][5193] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" HandleID="k8s-pod-network.663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" Workload="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:16.200245 env[1323]: 2025-05-13 00:16:16.193 [INFO][5193] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" HandleID="k8s-pod-network.663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" Workload="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:16.200245 env[1323]: 2025-05-13 00:16:16.195 [INFO][5193] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:16.200245 env[1323]: 2025-05-13 00:16:16.197 [INFO][5183] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" May 13 00:16:16.200850 env[1323]: time="2025-05-13T00:16:16.200813642Z" level=info msg="TearDown network for sandbox \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\" successfully" May 13 00:16:16.200921 env[1323]: time="2025-05-13T00:16:16.200904802Z" level=info msg="StopPodSandbox for \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\" returns successfully" May 13 00:16:16.204190 env[1323]: time="2025-05-13T00:16:16.204154443Z" level=info msg="RemovePodSandbox for \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\"" May 13 00:16:16.204295 env[1323]: time="2025-05-13T00:16:16.204200203Z" level=info msg="Forcibly stopping sandbox \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\"" May 13 00:16:16.291672 env[1323]: 2025-05-13 00:16:16.243 [WARNING][5216] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"943c0c0f-6fce-4911-b87f-c827598e67dc", ResourceVersion:"930", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"97a23e8f07d0fe2ce2d16c37ccf4bf016785133f090831885f3acf0ce387c646", Pod:"coredns-7db6d8ff4d-k77ll", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali99ccf0c6009", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:16.291672 env[1323]: 2025-05-13 00:16:16.244 [INFO][5216] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" May 13 00:16:16.291672 env[1323]: 2025-05-13 00:16:16.244 [INFO][5216] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" iface="eth0" netns="" May 13 00:16:16.291672 env[1323]: 2025-05-13 00:16:16.244 [INFO][5216] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" May 13 00:16:16.291672 env[1323]: 2025-05-13 00:16:16.244 [INFO][5216] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" May 13 00:16:16.291672 env[1323]: 2025-05-13 00:16:16.269 [INFO][5224] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" HandleID="k8s-pod-network.663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" Workload="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:16.291672 env[1323]: 2025-05-13 00:16:16.269 [INFO][5224] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:16.291672 env[1323]: 2025-05-13 00:16:16.270 [INFO][5224] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:16.291672 env[1323]: 2025-05-13 00:16:16.282 [WARNING][5224] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" HandleID="k8s-pod-network.663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" Workload="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:16.291672 env[1323]: 2025-05-13 00:16:16.282 [INFO][5224] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" HandleID="k8s-pod-network.663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" Workload="localhost-k8s-coredns--7db6d8ff4d--k77ll-eth0" May 13 00:16:16.291672 env[1323]: 2025-05-13 00:16:16.286 [INFO][5224] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:16.291672 env[1323]: 2025-05-13 00:16:16.289 [INFO][5216] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f" May 13 00:16:16.292099 env[1323]: time="2025-05-13T00:16:16.291698092Z" level=info msg="TearDown network for sandbox \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\" successfully" May 13 00:16:16.303568 env[1323]: time="2025-05-13T00:16:16.300356693Z" level=info msg="RemovePodSandbox \"663d7ab19ecf90d371ba661c3119faafc5bff145f25210039cbd3e282fb2ec0f\" returns successfully" May 13 00:16:16.303568 env[1323]: time="2025-05-13T00:16:16.300941133Z" level=info msg="StopPodSandbox for \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\"" May 13 00:16:16.382269 env[1323]: 2025-05-13 00:16:16.337 [WARNING][5247] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"8dde30de-095c-474d-a07e-44b258f6e953", ResourceVersion:"876", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af", Pod:"coredns-7db6d8ff4d-ncf86", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6bab1e72142", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:16.382269 env[1323]: 2025-05-13 00:16:16.338 [INFO][5247] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" May 13 00:16:16.382269 env[1323]: 2025-05-13 00:16:16.338 [INFO][5247] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" iface="eth0" netns="" May 13 00:16:16.382269 env[1323]: 2025-05-13 00:16:16.338 [INFO][5247] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" May 13 00:16:16.382269 env[1323]: 2025-05-13 00:16:16.338 [INFO][5247] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" May 13 00:16:16.382269 env[1323]: 2025-05-13 00:16:16.363 [INFO][5263] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" HandleID="k8s-pod-network.316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" Workload="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:16:16.382269 env[1323]: 2025-05-13 00:16:16.363 [INFO][5263] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:16.382269 env[1323]: 2025-05-13 00:16:16.363 [INFO][5263] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:16.382269 env[1323]: 2025-05-13 00:16:16.372 [WARNING][5263] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" HandleID="k8s-pod-network.316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" Workload="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:16:16.382269 env[1323]: 2025-05-13 00:16:16.372 [INFO][5263] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" HandleID="k8s-pod-network.316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" Workload="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:16:16.382269 env[1323]: 2025-05-13 00:16:16.377 [INFO][5263] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:16.382269 env[1323]: 2025-05-13 00:16:16.381 [INFO][5247] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" May 13 00:16:16.382837 env[1323]: time="2025-05-13T00:16:16.382314901Z" level=info msg="TearDown network for sandbox \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\" successfully" May 13 00:16:16.382837 env[1323]: time="2025-05-13T00:16:16.382352021Z" level=info msg="StopPodSandbox for \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\" returns successfully" May 13 00:16:16.383367 env[1323]: time="2025-05-13T00:16:16.383342381Z" level=info msg="RemovePodSandbox for \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\"" May 13 00:16:16.383500 env[1323]: time="2025-05-13T00:16:16.383461061Z" level=info msg="Forcibly stopping sandbox \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\"" May 13 00:16:16.453314 env[1323]: 2025-05-13 00:16:16.419 [WARNING][5299] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"8dde30de-095c-474d-a07e-44b258f6e953", ResourceVersion:"876", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"bcbd8a0939a5e54268e8b534c872213ad385b30b799db7063ee2ee37173e05af", Pod:"coredns-7db6d8ff4d-ncf86", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6bab1e72142", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:16.453314 env[1323]: 2025-05-13 00:16:16.419 [INFO][5299] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" May 13 00:16:16.453314 env[1323]: 2025-05-13 00:16:16.419 [INFO][5299] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" iface="eth0" netns="" May 13 00:16:16.453314 env[1323]: 2025-05-13 00:16:16.419 [INFO][5299] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" May 13 00:16:16.453314 env[1323]: 2025-05-13 00:16:16.419 [INFO][5299] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" May 13 00:16:16.453314 env[1323]: 2025-05-13 00:16:16.438 [INFO][5308] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" HandleID="k8s-pod-network.316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" Workload="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:16:16.453314 env[1323]: 2025-05-13 00:16:16.438 [INFO][5308] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:16.453314 env[1323]: 2025-05-13 00:16:16.438 [INFO][5308] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:16.453314 env[1323]: 2025-05-13 00:16:16.448 [WARNING][5308] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" HandleID="k8s-pod-network.316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" Workload="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:16:16.453314 env[1323]: 2025-05-13 00:16:16.448 [INFO][5308] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" HandleID="k8s-pod-network.316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" Workload="localhost-k8s-coredns--7db6d8ff4d--ncf86-eth0" May 13 00:16:16.453314 env[1323]: 2025-05-13 00:16:16.450 [INFO][5308] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:16.453314 env[1323]: 2025-05-13 00:16:16.451 [INFO][5299] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2" May 13 00:16:16.453769 env[1323]: time="2025-05-13T00:16:16.453348189Z" level=info msg="TearDown network for sandbox \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\" successfully" May 13 00:16:16.456174 env[1323]: time="2025-05-13T00:16:16.456139789Z" level=info msg="RemovePodSandbox \"316570aa1c6a1732be95c15fb79864ddb704456658d4a61a78a8893a48f003a2\" returns successfully" May 13 00:16:16.456658 env[1323]: time="2025-05-13T00:16:16.456632309Z" level=info msg="StopPodSandbox for \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\"" May 13 00:16:16.534086 env[1323]: 2025-05-13 00:16:16.494 [WARNING][5330] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--srb86-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"838e7283-9b17-4532-a9f4-4b17f2b5ca3b", ResourceVersion:"959", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b7b4b9d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff", Pod:"csi-node-driver-srb86", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calia028079e884", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:16.534086 env[1323]: 2025-05-13 00:16:16.495 [INFO][5330] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" May 13 00:16:16.534086 env[1323]: 2025-05-13 00:16:16.495 [INFO][5330] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" iface="eth0" netns="" May 13 00:16:16.534086 env[1323]: 2025-05-13 00:16:16.495 [INFO][5330] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" May 13 00:16:16.534086 env[1323]: 2025-05-13 00:16:16.495 [INFO][5330] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" May 13 00:16:16.534086 env[1323]: 2025-05-13 00:16:16.515 [INFO][5338] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" HandleID="k8s-pod-network.ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" Workload="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:16.534086 env[1323]: 2025-05-13 00:16:16.515 [INFO][5338] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:16.534086 env[1323]: 2025-05-13 00:16:16.515 [INFO][5338] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:16.534086 env[1323]: 2025-05-13 00:16:16.528 [WARNING][5338] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" HandleID="k8s-pod-network.ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" Workload="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:16.534086 env[1323]: 2025-05-13 00:16:16.528 [INFO][5338] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" HandleID="k8s-pod-network.ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" Workload="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:16.534086 env[1323]: 2025-05-13 00:16:16.530 [INFO][5338] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:16.534086 env[1323]: 2025-05-13 00:16:16.532 [INFO][5330] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" May 13 00:16:16.534655 env[1323]: time="2025-05-13T00:16:16.534605357Z" level=info msg="TearDown network for sandbox \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\" successfully" May 13 00:16:16.534739 env[1323]: time="2025-05-13T00:16:16.534723117Z" level=info msg="StopPodSandbox for \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\" returns successfully" May 13 00:16:16.535310 env[1323]: time="2025-05-13T00:16:16.535282277Z" level=info msg="RemovePodSandbox for \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\"" May 13 00:16:16.535437 env[1323]: time="2025-05-13T00:16:16.535399237Z" level=info msg="Forcibly stopping sandbox \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\"" May 13 00:16:16.619858 env[1323]: 2025-05-13 00:16:16.570 [WARNING][5360] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--srb86-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"838e7283-9b17-4532-a9f4-4b17f2b5ca3b", ResourceVersion:"959", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b7b4b9d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"ec015621a8cfb1e270c83a9d8fea5884286349bc20c2dc43cc19eddf9bc526ff", Pod:"csi-node-driver-srb86", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calia028079e884", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:16.619858 env[1323]: 2025-05-13 00:16:16.570 [INFO][5360] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" May 13 00:16:16.619858 env[1323]: 2025-05-13 00:16:16.570 [INFO][5360] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" iface="eth0" netns="" May 13 00:16:16.619858 env[1323]: 2025-05-13 00:16:16.570 [INFO][5360] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" May 13 00:16:16.619858 env[1323]: 2025-05-13 00:16:16.570 [INFO][5360] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" May 13 00:16:16.619858 env[1323]: 2025-05-13 00:16:16.603 [INFO][5368] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" HandleID="k8s-pod-network.ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" Workload="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:16.619858 env[1323]: 2025-05-13 00:16:16.603 [INFO][5368] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:16.619858 env[1323]: 2025-05-13 00:16:16.604 [INFO][5368] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:16.619858 env[1323]: 2025-05-13 00:16:16.614 [WARNING][5368] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" HandleID="k8s-pod-network.ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" Workload="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:16.619858 env[1323]: 2025-05-13 00:16:16.614 [INFO][5368] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" HandleID="k8s-pod-network.ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" Workload="localhost-k8s-csi--node--driver--srb86-eth0" May 13 00:16:16.619858 env[1323]: 2025-05-13 00:16:16.616 [INFO][5368] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:16.619858 env[1323]: 2025-05-13 00:16:16.618 [INFO][5360] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e" May 13 00:16:16.620305 env[1323]: time="2025-05-13T00:16:16.619885926Z" level=info msg="TearDown network for sandbox \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\" successfully" May 13 00:16:16.624153 env[1323]: time="2025-05-13T00:16:16.624102606Z" level=info msg="RemovePodSandbox \"ee0c0b59929110b7d9c26a4a85533626afd4104cfdd99a1b234655b90724471e\" returns successfully" May 13 00:16:16.624642 env[1323]: time="2025-05-13T00:16:16.624616046Z" level=info msg="StopPodSandbox for \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\"" May 13 00:16:16.703481 env[1323]: 2025-05-13 00:16:16.665 [WARNING][5392] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0", GenerateName:"calico-apiserver-5b58ff46dd-", Namespace:"calico-apiserver", SelfLink:"", UID:"4a3e9432-3790-4f08-9c11-d537f6aab7dd", ResourceVersion:"944", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5b58ff46dd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6", Pod:"calico-apiserver-5b58ff46dd-zk6tg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali0bed5f3decf", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:16.703481 env[1323]: 2025-05-13 00:16:16.665 [INFO][5392] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" May 13 00:16:16.703481 env[1323]: 2025-05-13 00:16:16.665 [INFO][5392] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" iface="eth0" netns="" May 13 00:16:16.703481 env[1323]: 2025-05-13 00:16:16.665 [INFO][5392] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" May 13 00:16:16.703481 env[1323]: 2025-05-13 00:16:16.665 [INFO][5392] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" May 13 00:16:16.703481 env[1323]: 2025-05-13 00:16:16.686 [INFO][5400] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" HandleID="k8s-pod-network.922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:16.703481 env[1323]: 2025-05-13 00:16:16.686 [INFO][5400] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:16.703481 env[1323]: 2025-05-13 00:16:16.686 [INFO][5400] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:16.703481 env[1323]: 2025-05-13 00:16:16.697 [WARNING][5400] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" HandleID="k8s-pod-network.922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:16.703481 env[1323]: 2025-05-13 00:16:16.697 [INFO][5400] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" HandleID="k8s-pod-network.922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:16.703481 env[1323]: 2025-05-13 00:16:16.698 [INFO][5400] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:16.703481 env[1323]: 2025-05-13 00:16:16.700 [INFO][5392] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" May 13 00:16:16.703481 env[1323]: time="2025-05-13T00:16:16.702054374Z" level=info msg="TearDown network for sandbox \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\" successfully" May 13 00:16:16.703481 env[1323]: time="2025-05-13T00:16:16.702085494Z" level=info msg="StopPodSandbox for \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\" returns successfully" May 13 00:16:16.703481 env[1323]: time="2025-05-13T00:16:16.703009334Z" level=info msg="RemovePodSandbox for \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\"" May 13 00:16:16.703481 env[1323]: time="2025-05-13T00:16:16.703075054Z" level=info msg="Forcibly stopping sandbox \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\"" May 13 00:16:16.777950 env[1323]: 2025-05-13 00:16:16.743 [WARNING][5423] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0", GenerateName:"calico-apiserver-5b58ff46dd-", Namespace:"calico-apiserver", SelfLink:"", UID:"4a3e9432-3790-4f08-9c11-d537f6aab7dd", ResourceVersion:"944", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5b58ff46dd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"54cdb7ac526c17a5cb1a9365cc608d5c72e073184491729f88eac9d4398afdb6", Pod:"calico-apiserver-5b58ff46dd-zk6tg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali0bed5f3decf", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:16.777950 env[1323]: 2025-05-13 00:16:16.743 [INFO][5423] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" May 13 00:16:16.777950 env[1323]: 2025-05-13 00:16:16.743 [INFO][5423] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" iface="eth0" netns="" May 13 00:16:16.777950 env[1323]: 2025-05-13 00:16:16.743 [INFO][5423] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" May 13 00:16:16.777950 env[1323]: 2025-05-13 00:16:16.743 [INFO][5423] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" May 13 00:16:16.777950 env[1323]: 2025-05-13 00:16:16.764 [INFO][5431] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" HandleID="k8s-pod-network.922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:16.777950 env[1323]: 2025-05-13 00:16:16.764 [INFO][5431] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:16.777950 env[1323]: 2025-05-13 00:16:16.764 [INFO][5431] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:16.777950 env[1323]: 2025-05-13 00:16:16.772 [WARNING][5431] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" HandleID="k8s-pod-network.922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:16.777950 env[1323]: 2025-05-13 00:16:16.772 [INFO][5431] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" HandleID="k8s-pod-network.922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--zk6tg-eth0" May 13 00:16:16.777950 env[1323]: 2025-05-13 00:16:16.774 [INFO][5431] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:16.777950 env[1323]: 2025-05-13 00:16:16.775 [INFO][5423] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec" May 13 00:16:16.778591 env[1323]: time="2025-05-13T00:16:16.778515662Z" level=info msg="TearDown network for sandbox \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\" successfully" May 13 00:16:16.781628 env[1323]: time="2025-05-13T00:16:16.781590583Z" level=info msg="RemovePodSandbox \"922c20f2a30e01cc32a7c03af20a90d65e91463dac651b837faea5790494aeec\" returns successfully" May 13 00:16:16.782443 env[1323]: time="2025-05-13T00:16:16.782304943Z" level=info msg="StopPodSandbox for \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\"" May 13 00:16:16.854445 env[1323]: 2025-05-13 00:16:16.820 [WARNING][5454] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0", GenerateName:"calico-apiserver-5b58ff46dd-", Namespace:"calico-apiserver", SelfLink:"", UID:"4596526d-58fb-4ef8-81dc-275ff89a1ce4", ResourceVersion:"902", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5b58ff46dd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7", Pod:"calico-apiserver-5b58ff46dd-56crg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie3fc7efc9ef", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:16.854445 env[1323]: 2025-05-13 00:16:16.820 [INFO][5454] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" May 13 00:16:16.854445 env[1323]: 2025-05-13 00:16:16.820 [INFO][5454] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" iface="eth0" netns="" May 13 00:16:16.854445 env[1323]: 2025-05-13 00:16:16.820 [INFO][5454] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" May 13 00:16:16.854445 env[1323]: 2025-05-13 00:16:16.820 [INFO][5454] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" May 13 00:16:16.854445 env[1323]: 2025-05-13 00:16:16.840 [INFO][5463] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" HandleID="k8s-pod-network.d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:16:16.854445 env[1323]: 2025-05-13 00:16:16.840 [INFO][5463] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:16.854445 env[1323]: 2025-05-13 00:16:16.840 [INFO][5463] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:16.854445 env[1323]: 2025-05-13 00:16:16.849 [WARNING][5463] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" HandleID="k8s-pod-network.d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:16:16.854445 env[1323]: 2025-05-13 00:16:16.849 [INFO][5463] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" HandleID="k8s-pod-network.d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:16:16.854445 env[1323]: 2025-05-13 00:16:16.851 [INFO][5463] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:16.854445 env[1323]: 2025-05-13 00:16:16.853 [INFO][5454] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" May 13 00:16:16.854882 env[1323]: time="2025-05-13T00:16:16.854494430Z" level=info msg="TearDown network for sandbox \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\" successfully" May 13 00:16:16.854882 env[1323]: time="2025-05-13T00:16:16.854527550Z" level=info msg="StopPodSandbox for \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\" returns successfully" May 13 00:16:16.855069 env[1323]: time="2025-05-13T00:16:16.855014510Z" level=info msg="RemovePodSandbox for \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\"" May 13 00:16:16.855114 env[1323]: time="2025-05-13T00:16:16.855057710Z" level=info msg="Forcibly stopping sandbox \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\"" May 13 00:16:16.926067 env[1323]: 2025-05-13 00:16:16.892 [WARNING][5486] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0", GenerateName:"calico-apiserver-5b58ff46dd-", Namespace:"calico-apiserver", SelfLink:"", UID:"4596526d-58fb-4ef8-81dc-275ff89a1ce4", ResourceVersion:"902", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5b58ff46dd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"30601a66f5e8df7003c49b7a1e8861ea93d9fe5bdfb8982055cfc877de4e2ac7", Pod:"calico-apiserver-5b58ff46dd-56crg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie3fc7efc9ef", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:16.926067 env[1323]: 2025-05-13 00:16:16.892 [INFO][5486] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" May 13 00:16:16.926067 env[1323]: 2025-05-13 00:16:16.892 [INFO][5486] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" iface="eth0" netns="" May 13 00:16:16.926067 env[1323]: 2025-05-13 00:16:16.892 [INFO][5486] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" May 13 00:16:16.926067 env[1323]: 2025-05-13 00:16:16.892 [INFO][5486] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" May 13 00:16:16.926067 env[1323]: 2025-05-13 00:16:16.912 [INFO][5494] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" HandleID="k8s-pod-network.d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:16:16.926067 env[1323]: 2025-05-13 00:16:16.912 [INFO][5494] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:16.926067 env[1323]: 2025-05-13 00:16:16.912 [INFO][5494] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:16.926067 env[1323]: 2025-05-13 00:16:16.921 [WARNING][5494] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" HandleID="k8s-pod-network.d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:16:16.926067 env[1323]: 2025-05-13 00:16:16.921 [INFO][5494] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" HandleID="k8s-pod-network.d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" Workload="localhost-k8s-calico--apiserver--5b58ff46dd--56crg-eth0" May 13 00:16:16.926067 env[1323]: 2025-05-13 00:16:16.922 [INFO][5494] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:16.926067 env[1323]: 2025-05-13 00:16:16.924 [INFO][5486] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc" May 13 00:16:16.926519 env[1323]: time="2025-05-13T00:16:16.926099157Z" level=info msg="TearDown network for sandbox \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\" successfully" May 13 00:16:16.929349 env[1323]: time="2025-05-13T00:16:16.929304918Z" level=info msg="RemovePodSandbox \"d130efc27bc7f2a48d3f67e40ad98b20f650ce068831fe05cbac337b117b2dbc\" returns successfully" May 13 00:16:16.929962 env[1323]: time="2025-05-13T00:16:16.929931638Z" level=info msg="StopPodSandbox for \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\"" May 13 00:16:17.003177 env[1323]: 2025-05-13 00:16:16.965 [WARNING][5517] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0", GenerateName:"calico-kube-controllers-66bcdc688d-", Namespace:"calico-system", SelfLink:"", UID:"47f5ba25-f173-43e0-af82-849feb0085ec", ResourceVersion:"973", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"66bcdc688d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42", Pod:"calico-kube-controllers-66bcdc688d-zm28k", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calic35c4880c53", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:17.003177 env[1323]: 2025-05-13 00:16:16.965 [INFO][5517] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" May 13 00:16:17.003177 env[1323]: 2025-05-13 00:16:16.965 [INFO][5517] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" iface="eth0" netns="" May 13 00:16:17.003177 env[1323]: 2025-05-13 00:16:16.965 [INFO][5517] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" May 13 00:16:17.003177 env[1323]: 2025-05-13 00:16:16.965 [INFO][5517] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" May 13 00:16:17.003177 env[1323]: 2025-05-13 00:16:16.986 [INFO][5525] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" HandleID="k8s-pod-network.37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" Workload="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:17.003177 env[1323]: 2025-05-13 00:16:16.986 [INFO][5525] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:17.003177 env[1323]: 2025-05-13 00:16:16.986 [INFO][5525] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:17.003177 env[1323]: 2025-05-13 00:16:16.997 [WARNING][5525] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" HandleID="k8s-pod-network.37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" Workload="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:17.003177 env[1323]: 2025-05-13 00:16:16.997 [INFO][5525] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" HandleID="k8s-pod-network.37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" Workload="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:17.003177 env[1323]: 2025-05-13 00:16:16.998 [INFO][5525] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:17.003177 env[1323]: 2025-05-13 00:16:17.000 [INFO][5517] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" May 13 00:16:17.003177 env[1323]: time="2025-05-13T00:16:17.001772245Z" level=info msg="TearDown network for sandbox \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\" successfully" May 13 00:16:17.003177 env[1323]: time="2025-05-13T00:16:17.001806685Z" level=info msg="StopPodSandbox for \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\" returns successfully" May 13 00:16:17.003177 env[1323]: time="2025-05-13T00:16:17.002274485Z" level=info msg="RemovePodSandbox for \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\"" May 13 00:16:17.003177 env[1323]: time="2025-05-13T00:16:17.002308725Z" level=info msg="Forcibly stopping sandbox \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\"" May 13 00:16:17.075850 env[1323]: 2025-05-13 00:16:17.038 [WARNING][5548] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0", GenerateName:"calico-kube-controllers-66bcdc688d-", Namespace:"calico-system", SelfLink:"", UID:"47f5ba25-f173-43e0-af82-849feb0085ec", ResourceVersion:"973", Generation:0, CreationTimestamp:time.Date(2025, time.May, 13, 0, 15, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"66bcdc688d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"d2d10e4e739816a60cfccffa065e23b506f9b57318b289dd0806f34d9c705b42", Pod:"calico-kube-controllers-66bcdc688d-zm28k", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calic35c4880c53", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 13 00:16:17.075850 env[1323]: 2025-05-13 00:16:17.038 [INFO][5548] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" May 13 00:16:17.075850 env[1323]: 2025-05-13 00:16:17.038 [INFO][5548] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" iface="eth0" netns="" May 13 00:16:17.075850 env[1323]: 2025-05-13 00:16:17.038 [INFO][5548] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" May 13 00:16:17.075850 env[1323]: 2025-05-13 00:16:17.038 [INFO][5548] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" May 13 00:16:17.075850 env[1323]: 2025-05-13 00:16:17.061 [INFO][5557] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" HandleID="k8s-pod-network.37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" Workload="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:17.075850 env[1323]: 2025-05-13 00:16:17.061 [INFO][5557] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 13 00:16:17.075850 env[1323]: 2025-05-13 00:16:17.062 [INFO][5557] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 13 00:16:17.075850 env[1323]: 2025-05-13 00:16:17.070 [WARNING][5557] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" HandleID="k8s-pod-network.37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" Workload="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:17.075850 env[1323]: 2025-05-13 00:16:17.070 [INFO][5557] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" HandleID="k8s-pod-network.37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" Workload="localhost-k8s-calico--kube--controllers--66bcdc688d--zm28k-eth0" May 13 00:16:17.075850 env[1323]: 2025-05-13 00:16:17.072 [INFO][5557] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 13 00:16:17.075850 env[1323]: 2025-05-13 00:16:17.074 [INFO][5548] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543" May 13 00:16:17.076268 env[1323]: time="2025-05-13T00:16:17.075870093Z" level=info msg="TearDown network for sandbox \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\" successfully" May 13 00:16:17.078968 env[1323]: time="2025-05-13T00:16:17.078932253Z" level=info msg="RemovePodSandbox \"37c58ff9e29ab86554c4fa5a0b42a75f60cba9eda9e1eb32b4ae680a677d7543\" returns successfully" May 13 00:16:17.349000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.8:22-10.0.0.1:36892 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:17.350051 systemd[1]: Started sshd@15-10.0.0.8:22-10.0.0.1:36892.service. May 13 00:16:17.351386 kernel: kauditd_printk_skb: 1 callbacks suppressed May 13 00:16:17.351428 kernel: audit: type=1130 audit(1747095377.349:482): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.8:22-10.0.0.1:36892 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:17.385000 audit[5565]: USER_ACCT pid=5565 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.386280 sshd[5565]: Accepted publickey for core from 10.0.0.1 port 36892 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:16:17.387700 sshd[5565]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:16:17.386000 audit[5565]: CRED_ACQ pid=5565 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.393161 kernel: audit: type=1101 audit(1747095377.385:483): pid=5565 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.393253 kernel: audit: type=1103 audit(1747095377.386:484): pid=5565 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.393274 kernel: audit: type=1006 audit(1747095377.386:485): pid=5565 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=16 res=1 May 13 00:16:17.386000 audit[5565]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffc6e3ebd0 a2=3 a3=1 items=0 ppid=1 pid=5565 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:17.395577 systemd-logind[1306]: New session 16 of user core. May 13 00:16:17.396456 systemd[1]: Started session-16.scope. May 13 00:16:17.398922 kernel: audit: type=1300 audit(1747095377.386:485): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffc6e3ebd0 a2=3 a3=1 items=0 ppid=1 pid=5565 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:17.399004 kernel: audit: type=1327 audit(1747095377.386:485): proctitle=737368643A20636F7265205B707269765D May 13 00:16:17.386000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:16:17.400000 audit[5565]: USER_START pid=5565 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.401000 audit[5568]: CRED_ACQ pid=5568 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.408177 kernel: audit: type=1105 audit(1747095377.400:486): pid=5565 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.408261 kernel: audit: type=1103 audit(1747095377.401:487): pid=5568 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.535746 sshd[5565]: pam_unix(sshd:session): session closed for user core May 13 00:16:17.536000 audit[5565]: USER_END pid=5565 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.537778 systemd[1]: Started sshd@16-10.0.0.8:22-10.0.0.1:36904.service. May 13 00:16:17.538630 systemd[1]: sshd@15-10.0.0.8:22-10.0.0.1:36892.service: Deactivated successfully. May 13 00:16:17.539554 systemd[1]: session-16.scope: Deactivated successfully. May 13 00:16:17.536000 audit[5565]: CRED_DISP pid=5565 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.544826 kernel: audit: type=1106 audit(1747095377.536:488): pid=5565 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.544891 kernel: audit: type=1104 audit(1747095377.536:489): pid=5565 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.537000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.8:22-10.0.0.1:36904 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:17.538000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.8:22-10.0.0.1:36892 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:17.545003 systemd-logind[1306]: Session 16 logged out. Waiting for processes to exit. May 13 00:16:17.546253 systemd-logind[1306]: Removed session 16. May 13 00:16:17.572000 audit[5577]: USER_ACCT pid=5577 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.573353 sshd[5577]: Accepted publickey for core from 10.0.0.1 port 36904 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:16:17.573000 audit[5577]: CRED_ACQ pid=5577 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.574000 audit[5577]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffffa25a330 a2=3 a3=1 items=0 ppid=1 pid=5577 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=17 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:17.574000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:16:17.575085 sshd[5577]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:16:17.578854 systemd-logind[1306]: New session 17 of user core. May 13 00:16:17.579798 systemd[1]: Started session-17.scope. May 13 00:16:17.582000 audit[5577]: USER_START pid=5577 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.584000 audit[5582]: CRED_ACQ pid=5582 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.807409 sshd[5577]: pam_unix(sshd:session): session closed for user core May 13 00:16:17.808000 audit[5577]: USER_END pid=5577 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.809504 systemd[1]: Started sshd@17-10.0.0.8:22-10.0.0.1:36906.service. May 13 00:16:17.809000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-10.0.0.8:22-10.0.0.1:36906 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:17.809000 audit[5577]: CRED_DISP pid=5577 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.813743 systemd-logind[1306]: Session 17 logged out. Waiting for processes to exit. May 13 00:16:17.814013 systemd[1]: sshd@16-10.0.0.8:22-10.0.0.1:36904.service: Deactivated successfully. May 13 00:16:17.813000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.8:22-10.0.0.1:36904 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:17.815356 systemd[1]: session-17.scope: Deactivated successfully. May 13 00:16:17.815929 systemd-logind[1306]: Removed session 17. May 13 00:16:17.849000 audit[5589]: USER_ACCT pid=5589 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.850069 sshd[5589]: Accepted publickey for core from 10.0.0.1 port 36906 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:16:17.850000 audit[5589]: CRED_ACQ pid=5589 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.850000 audit[5589]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffff56573b0 a2=3 a3=1 items=0 ppid=1 pid=5589 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=18 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:17.850000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:16:17.851423 sshd[5589]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:16:17.855042 systemd-logind[1306]: New session 18 of user core. May 13 00:16:17.855916 systemd[1]: Started session-18.scope. May 13 00:16:17.858000 audit[5589]: USER_START pid=5589 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:17.860000 audit[5594]: CRED_ACQ pid=5594 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.384000 audit[5617]: NETFILTER_CFG table=filter:113 family=2 entries=20 op=nft_register_rule pid=5617 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:19.384000 audit[5617]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=11860 a0=3 a1=ffffecbd4350 a2=0 a3=1 items=0 ppid=2393 pid=5617 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:19.384000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:19.394000 audit[5617]: NETFILTER_CFG table=nat:114 family=2 entries=22 op=nft_register_rule pid=5617 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:19.394000 audit[5617]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=6540 a0=3 a1=ffffecbd4350 a2=0 a3=1 items=0 ppid=2393 pid=5617 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:19.394000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:19.405146 sshd[5589]: pam_unix(sshd:session): session closed for user core May 13 00:16:19.406000 audit[5589]: USER_END pid=5589 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.406000 audit[5589]: CRED_DISP pid=5589 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.407729 systemd[1]: Started sshd@18-10.0.0.8:22-10.0.0.1:36916.service. May 13 00:16:19.407000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.8:22-10.0.0.1:36916 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:19.413039 systemd[1]: sshd@17-10.0.0.8:22-10.0.0.1:36906.service: Deactivated successfully. May 13 00:16:19.412000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-10.0.0.8:22-10.0.0.1:36906 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:19.414037 systemd-logind[1306]: Session 18 logged out. Waiting for processes to exit. May 13 00:16:19.414111 systemd[1]: session-18.scope: Deactivated successfully. May 13 00:16:19.414891 systemd-logind[1306]: Removed session 18. May 13 00:16:19.415000 audit[5620]: NETFILTER_CFG table=filter:115 family=2 entries=32 op=nft_register_rule pid=5620 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:19.415000 audit[5620]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=11860 a0=3 a1=ffffdf902c40 a2=0 a3=1 items=0 ppid=2393 pid=5620 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:19.415000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:19.420000 audit[5620]: NETFILTER_CFG table=nat:116 family=2 entries=22 op=nft_register_rule pid=5620 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:19.420000 audit[5620]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=6540 a0=3 a1=ffffdf902c40 a2=0 a3=1 items=0 ppid=2393 pid=5620 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:19.420000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:19.442000 audit[5619]: USER_ACCT pid=5619 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.443398 sshd[5619]: Accepted publickey for core from 10.0.0.1 port 36916 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:16:19.444000 audit[5619]: CRED_ACQ pid=5619 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.444000 audit[5619]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffc692e350 a2=3 a3=1 items=0 ppid=1 pid=5619 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=19 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:19.444000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:16:19.444997 sshd[5619]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:16:19.448450 systemd-logind[1306]: New session 19 of user core. May 13 00:16:19.449327 systemd[1]: Started session-19.scope. May 13 00:16:19.452000 audit[5619]: USER_START pid=5619 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.454000 audit[5625]: CRED_ACQ pid=5625 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.740831 sshd[5619]: pam_unix(sshd:session): session closed for user core May 13 00:16:19.741000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.8:22-10.0.0.1:36922 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:19.742345 systemd[1]: Started sshd@19-10.0.0.8:22-10.0.0.1:36922.service. May 13 00:16:19.742000 audit[5619]: USER_END pid=5619 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.742000 audit[5619]: CRED_DISP pid=5619 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.744715 systemd-logind[1306]: Session 19 logged out. Waiting for processes to exit. May 13 00:16:19.745251 systemd[1]: sshd@18-10.0.0.8:22-10.0.0.1:36916.service: Deactivated successfully. May 13 00:16:19.744000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.8:22-10.0.0.1:36916 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:19.746174 systemd[1]: session-19.scope: Deactivated successfully. May 13 00:16:19.746733 systemd-logind[1306]: Removed session 19. May 13 00:16:19.779000 audit[5632]: USER_ACCT pid=5632 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.779969 sshd[5632]: Accepted publickey for core from 10.0.0.1 port 36922 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:16:19.780000 audit[5632]: CRED_ACQ pid=5632 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.780000 audit[5632]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffc33c43e0 a2=3 a3=1 items=0 ppid=1 pid=5632 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=20 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:19.780000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:16:19.781357 sshd[5632]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:16:19.785809 systemd[1]: Started session-20.scope. May 13 00:16:19.786152 systemd-logind[1306]: New session 20 of user core. May 13 00:16:19.790000 audit[5632]: USER_START pid=5632 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.791000 audit[5637]: CRED_ACQ pid=5637 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.920669 sshd[5632]: pam_unix(sshd:session): session closed for user core May 13 00:16:19.921000 audit[5632]: USER_END pid=5632 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.921000 audit[5632]: CRED_DISP pid=5632 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:19.923543 systemd-logind[1306]: Session 20 logged out. Waiting for processes to exit. May 13 00:16:19.923851 systemd[1]: sshd@19-10.0.0.8:22-10.0.0.1:36922.service: Deactivated successfully. May 13 00:16:19.923000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.8:22-10.0.0.1:36922 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:19.924666 systemd[1]: session-20.scope: Deactivated successfully. May 13 00:16:19.925657 systemd-logind[1306]: Removed session 20. May 13 00:16:23.979569 kernel: kauditd_printk_skb: 57 callbacks suppressed May 13 00:16:23.979707 kernel: audit: type=1325 audit(1747095383.976:531): table=filter:117 family=2 entries=20 op=nft_register_rule pid=5650 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:23.976000 audit[5650]: NETFILTER_CFG table=filter:117 family=2 entries=20 op=nft_register_rule pid=5650 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:23.976000 audit[5650]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2932 a0=3 a1=ffffc31fca40 a2=0 a3=1 items=0 ppid=2393 pid=5650 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:23.984597 kernel: audit: type=1300 audit(1747095383.976:531): arch=c00000b7 syscall=211 success=yes exit=2932 a0=3 a1=ffffc31fca40 a2=0 a3=1 items=0 ppid=2393 pid=5650 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:23.984660 kernel: audit: type=1327 audit(1747095383.976:531): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:23.976000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:23.990000 audit[5650]: NETFILTER_CFG table=nat:118 family=2 entries=106 op=nft_register_chain pid=5650 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:23.990000 audit[5650]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=49452 a0=3 a1=ffffc31fca40 a2=0 a3=1 items=0 ppid=2393 pid=5650 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:23.998238 kernel: audit: type=1325 audit(1747095383.990:532): table=nat:118 family=2 entries=106 op=nft_register_chain pid=5650 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 13 00:16:23.998294 kernel: audit: type=1300 audit(1747095383.990:532): arch=c00000b7 syscall=211 success=yes exit=49452 a0=3 a1=ffffc31fca40 a2=0 a3=1 items=0 ppid=2393 pid=5650 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:23.998331 kernel: audit: type=1327 audit(1747095383.990:532): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:23.990000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 13 00:16:24.923993 systemd[1]: Started sshd@20-10.0.0.8:22-10.0.0.1:32964.service. May 13 00:16:24.922000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.8:22-10.0.0.1:32964 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:24.927570 kernel: audit: type=1130 audit(1747095384.922:533): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.8:22-10.0.0.1:32964 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:24.957000 audit[5652]: USER_ACCT pid=5652 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:24.959415 sshd[5652]: Accepted publickey for core from 10.0.0.1 port 32964 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:16:24.961075 sshd[5652]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:16:24.959000 audit[5652]: CRED_ACQ pid=5652 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:24.966175 kernel: audit: type=1101 audit(1747095384.957:534): pid=5652 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:24.966280 kernel: audit: type=1103 audit(1747095384.959:535): pid=5652 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:24.970250 kernel: audit: type=1006 audit(1747095384.959:536): pid=5652 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=21 res=1 May 13 00:16:24.959000 audit[5652]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffebd155d0 a2=3 a3=1 items=0 ppid=1 pid=5652 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=21 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:24.959000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:16:24.970547 systemd-logind[1306]: New session 21 of user core. May 13 00:16:24.971399 systemd[1]: Started session-21.scope. May 13 00:16:24.977000 audit[5652]: USER_START pid=5652 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:24.978000 audit[5655]: CRED_ACQ pid=5655 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:25.098189 sshd[5652]: pam_unix(sshd:session): session closed for user core May 13 00:16:25.097000 audit[5652]: USER_END pid=5652 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:25.097000 audit[5652]: CRED_DISP pid=5652 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:25.100867 systemd[1]: sshd@20-10.0.0.8:22-10.0.0.1:32964.service: Deactivated successfully. May 13 00:16:25.099000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.8:22-10.0.0.1:32964 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:25.101905 systemd-logind[1306]: Session 21 logged out. Waiting for processes to exit. May 13 00:16:25.101960 systemd[1]: session-21.scope: Deactivated successfully. May 13 00:16:25.102716 systemd-logind[1306]: Removed session 21. May 13 00:16:30.101876 systemd[1]: Started sshd@21-10.0.0.8:22-10.0.0.1:32974.service. May 13 00:16:30.100000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.8:22-10.0.0.1:32974 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:30.102758 kernel: kauditd_printk_skb: 7 callbacks suppressed May 13 00:16:30.102840 kernel: audit: type=1130 audit(1747095390.100:542): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.8:22-10.0.0.1:32974 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:30.132000 audit[5694]: USER_ACCT pid=5694 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:30.134348 sshd[5694]: Accepted publickey for core from 10.0.0.1 port 32974 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:16:30.136006 sshd[5694]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:16:30.134000 audit[5694]: CRED_ACQ pid=5694 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:30.140402 kernel: audit: type=1101 audit(1747095390.132:543): pid=5694 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:30.140455 kernel: audit: type=1103 audit(1747095390.134:544): pid=5694 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:30.142483 kernel: audit: type=1006 audit(1747095390.134:545): pid=5694 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=22 res=1 May 13 00:16:30.134000 audit[5694]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffe4df5e60 a2=3 a3=1 items=0 ppid=1 pid=5694 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=22 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:30.145286 systemd[1]: Started session-22.scope. May 13 00:16:30.146008 kernel: audit: type=1300 audit(1747095390.134:545): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffe4df5e60 a2=3 a3=1 items=0 ppid=1 pid=5694 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=22 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:30.146028 systemd-logind[1306]: New session 22 of user core. May 13 00:16:30.134000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:16:30.147161 kernel: audit: type=1327 audit(1747095390.134:545): proctitle=737368643A20636F7265205B707269765D May 13 00:16:30.150000 audit[5694]: USER_START pid=5694 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:30.152000 audit[5697]: CRED_ACQ pid=5697 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:30.160530 kernel: audit: type=1105 audit(1747095390.150:546): pid=5694 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:30.160636 kernel: audit: type=1103 audit(1747095390.152:547): pid=5697 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:30.264680 sshd[5694]: pam_unix(sshd:session): session closed for user core May 13 00:16:30.264000 audit[5694]: USER_END pid=5694 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:30.267202 systemd[1]: sshd@21-10.0.0.8:22-10.0.0.1:32974.service: Deactivated successfully. May 13 00:16:30.268191 systemd-logind[1306]: Session 22 logged out. Waiting for processes to exit. May 13 00:16:30.268280 systemd[1]: session-22.scope: Deactivated successfully. May 13 00:16:30.269366 systemd-logind[1306]: Removed session 22. May 13 00:16:30.264000 audit[5694]: CRED_DISP pid=5694 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:30.272711 kernel: audit: type=1106 audit(1747095390.264:548): pid=5694 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:30.272773 kernel: audit: type=1104 audit(1747095390.264:549): pid=5694 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:30.265000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.8:22-10.0.0.1:32974 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:34.129740 kubelet[2230]: E0513 00:16:34.129706 2230 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 13 00:16:35.266000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.8:22-10.0.0.1:59178 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:35.267762 systemd[1]: Started sshd@22-10.0.0.8:22-10.0.0.1:59178.service. May 13 00:16:35.268657 kernel: kauditd_printk_skb: 1 callbacks suppressed May 13 00:16:35.268711 kernel: audit: type=1130 audit(1747095395.266:551): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.8:22-10.0.0.1:59178 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 13 00:16:35.301000 audit[5710]: USER_ACCT pid=5710 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:35.303470 sshd[5710]: Accepted publickey for core from 10.0.0.1 port 59178 ssh2: RSA SHA256:JqaCSrDIbVVQNbxsbpFjz60HxEXsX2X9A6oTs4HqYQk May 13 00:16:35.304846 sshd[5710]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 13 00:16:35.302000 audit[5710]: CRED_ACQ pid=5710 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:35.309626 kernel: audit: type=1101 audit(1747095395.301:552): pid=5710 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:35.309671 kernel: audit: type=1103 audit(1747095395.302:553): pid=5710 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:35.312528 kernel: audit: type=1006 audit(1747095395.303:554): pid=5710 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=23 res=1 May 13 00:16:35.312671 kernel: audit: type=1300 audit(1747095395.303:554): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffff37cb830 a2=3 a3=1 items=0 ppid=1 pid=5710 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=23 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:35.303000 audit[5710]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffff37cb830 a2=3 a3=1 items=0 ppid=1 pid=5710 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=23 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 13 00:16:35.312270 systemd-logind[1306]: New session 23 of user core. May 13 00:16:35.313295 systemd[1]: Started session-23.scope. May 13 00:16:35.303000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 13 00:16:35.316675 kernel: audit: type=1327 audit(1747095395.303:554): proctitle=737368643A20636F7265205B707269765D May 13 00:16:35.319000 audit[5710]: USER_START pid=5710 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:35.324550 kernel: audit: type=1105 audit(1747095395.319:555): pid=5710 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:35.323000 audit[5713]: CRED_ACQ pid=5713 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:35.328562 kernel: audit: type=1103 audit(1747095395.323:556): pid=5713 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:35.427038 sshd[5710]: pam_unix(sshd:session): session closed for user core May 13 00:16:35.426000 audit[5710]: USER_END pid=5710 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:35.429763 systemd[1]: sshd@22-10.0.0.8:22-10.0.0.1:59178.service: Deactivated successfully. May 13 00:16:35.430935 systemd-logind[1306]: Session 23 logged out. Waiting for processes to exit. May 13 00:16:35.430983 systemd[1]: session-23.scope: Deactivated successfully. May 13 00:16:35.431817 systemd-logind[1306]: Removed session 23. May 13 00:16:35.426000 audit[5710]: CRED_DISP pid=5710 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:35.434847 kernel: audit: type=1106 audit(1747095395.426:557): pid=5710 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:35.434925 kernel: audit: type=1104 audit(1747095395.426:558): pid=5710 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 13 00:16:35.428000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.8:22-10.0.0.1:59178 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'