Aug 13 00:01:19.005838 kernel: Booting Linux on physical CPU 0x0000000000 [0x410fd083] Aug 13 00:01:19.005875 kernel: Linux version 5.15.189-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP PREEMPT Tue Aug 12 22:50:30 -00 2025 Aug 13 00:01:19.005897 kernel: efi: EFI v2.70 by EDK II Aug 13 00:01:19.005913 kernel: efi: SMBIOS=0x7bed0000 SMBIOS 3.0=0x7beb0000 ACPI=0x786e0000 ACPI 2.0=0x786e0014 MEMATTR=0x7affea98 MEMRESERVE=0x716fcf98 Aug 13 00:01:19.005927 kernel: ACPI: Early table checksum verification disabled Aug 13 00:01:19.005940 kernel: ACPI: RSDP 0x00000000786E0014 000024 (v02 AMAZON) Aug 13 00:01:19.005956 kernel: ACPI: XSDT 0x00000000786D00E8 000064 (v01 AMAZON AMZNFACP 00000001 01000013) Aug 13 00:01:19.005970 kernel: ACPI: FACP 0x00000000786B0000 000114 (v06 AMAZON AMZNFACP 00000001 AMZN 00000001) Aug 13 00:01:19.005984 kernel: ACPI: DSDT 0x0000000078640000 00159D (v02 AMAZON AMZNDSDT 00000001 INTL 20160527) Aug 13 00:01:19.005997 kernel: ACPI: APIC 0x00000000786C0000 000108 (v04 AMAZON AMZNAPIC 00000001 AMZN 00000001) Aug 13 00:01:19.006015 kernel: ACPI: SPCR 0x00000000786A0000 000050 (v02 AMAZON AMZNSPCR 00000001 AMZN 00000001) Aug 13 00:01:19.006029 kernel: ACPI: GTDT 0x0000000078690000 000060 (v02 AMAZON AMZNGTDT 00000001 AMZN 00000001) Aug 13 00:01:19.006043 kernel: ACPI: MCFG 0x0000000078680000 00003C (v02 AMAZON AMZNMCFG 00000001 AMZN 00000001) Aug 13 00:01:19.006057 kernel: ACPI: SLIT 0x0000000078670000 00002D (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Aug 13 00:01:19.006074 kernel: ACPI: IORT 0x0000000078660000 000078 (v01 AMAZON AMZNIORT 00000001 AMZN 00000001) Aug 13 00:01:19.006092 kernel: ACPI: PPTT 0x0000000078650000 0000EC (v01 AMAZON AMZNPPTT 00000001 AMZN 00000001) Aug 13 00:01:19.009170 kernel: ACPI: SPCR: console: uart,mmio,0x90a0000,115200 Aug 13 00:01:19.009190 kernel: earlycon: uart0 at MMIO 0x00000000090a0000 (options '115200') Aug 13 00:01:19.009206 kernel: printk: bootconsole [uart0] enabled Aug 13 00:01:19.009221 kernel: NUMA: Failed to initialise from firmware Aug 13 00:01:19.009236 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000004b5ffffff] Aug 13 00:01:19.009251 kernel: NUMA: NODE_DATA [mem 0x4b5843900-0x4b5848fff] Aug 13 00:01:19.009266 kernel: Zone ranges: Aug 13 00:01:19.009281 kernel: DMA [mem 0x0000000040000000-0x00000000ffffffff] Aug 13 00:01:19.009295 kernel: DMA32 empty Aug 13 00:01:19.009309 kernel: Normal [mem 0x0000000100000000-0x00000004b5ffffff] Aug 13 00:01:19.009331 kernel: Movable zone start for each node Aug 13 00:01:19.009346 kernel: Early memory node ranges Aug 13 00:01:19.009360 kernel: node 0: [mem 0x0000000040000000-0x000000007862ffff] Aug 13 00:01:19.009375 kernel: node 0: [mem 0x0000000078630000-0x000000007863ffff] Aug 13 00:01:19.009389 kernel: node 0: [mem 0x0000000078640000-0x00000000786effff] Aug 13 00:01:19.009403 kernel: node 0: [mem 0x00000000786f0000-0x000000007872ffff] Aug 13 00:01:19.009418 kernel: node 0: [mem 0x0000000078730000-0x000000007bbfffff] Aug 13 00:01:19.009432 kernel: node 0: [mem 0x000000007bc00000-0x000000007bfdffff] Aug 13 00:01:19.009447 kernel: node 0: [mem 0x000000007bfe0000-0x000000007fffffff] Aug 13 00:01:19.009461 kernel: node 0: [mem 0x0000000400000000-0x00000004b5ffffff] Aug 13 00:01:19.009475 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000004b5ffffff] Aug 13 00:01:19.009490 kernel: On node 0, zone Normal: 8192 pages in unavailable ranges Aug 13 00:01:19.009508 kernel: psci: probing for conduit method from ACPI. Aug 13 00:01:19.009523 kernel: psci: PSCIv1.0 detected in firmware. Aug 13 00:01:19.009544 kernel: psci: Using standard PSCI v0.2 function IDs Aug 13 00:01:19.009560 kernel: psci: Trusted OS migration not required Aug 13 00:01:19.009575 kernel: psci: SMC Calling Convention v1.1 Aug 13 00:01:19.009594 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000001) Aug 13 00:01:19.009610 kernel: ACPI: SRAT not present Aug 13 00:01:19.009626 kernel: percpu: Embedded 30 pages/cpu s82968 r8192 d31720 u122880 Aug 13 00:01:19.009641 kernel: pcpu-alloc: s82968 r8192 d31720 u122880 alloc=30*4096 Aug 13 00:01:19.009657 kernel: pcpu-alloc: [0] 0 [0] 1 Aug 13 00:01:19.009672 kernel: Detected PIPT I-cache on CPU0 Aug 13 00:01:19.009687 kernel: CPU features: detected: GIC system register CPU interface Aug 13 00:01:19.009702 kernel: CPU features: detected: Spectre-v2 Aug 13 00:01:19.009717 kernel: CPU features: detected: Spectre-v3a Aug 13 00:01:19.009732 kernel: CPU features: detected: Spectre-BHB Aug 13 00:01:19.009747 kernel: CPU features: kernel page table isolation forced ON by KASLR Aug 13 00:01:19.009766 kernel: CPU features: detected: Kernel page table isolation (KPTI) Aug 13 00:01:19.009782 kernel: CPU features: detected: ARM erratum 1742098 Aug 13 00:01:19.009797 kernel: CPU features: detected: ARM errata 1165522, 1319367, or 1530923 Aug 13 00:01:19.009812 kernel: Built 1 zonelists, mobility grouping on. Total pages: 991872 Aug 13 00:01:19.009827 kernel: Policy zone: Normal Aug 13 00:01:19.009845 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlycon flatcar.first_boot=detected acpi=force flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=32404c0887e5b8a80b0f069916a8040bfd969c7a8f47a2db1168b24bc04220cc Aug 13 00:01:19.009861 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Aug 13 00:01:19.009877 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Aug 13 00:01:19.009892 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 00:01:19.009907 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Aug 13 00:01:19.009926 kernel: software IO TLB: mapped [mem 0x000000007c000000-0x0000000080000000] (64MB) Aug 13 00:01:19.009942 kernel: Memory: 3824460K/4030464K available (9792K kernel code, 2094K rwdata, 7588K rodata, 36416K init, 777K bss, 206004K reserved, 0K cma-reserved) Aug 13 00:01:19.009958 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Aug 13 00:01:19.009973 kernel: trace event string verifier disabled Aug 13 00:01:19.009988 kernel: rcu: Preemptible hierarchical RCU implementation. Aug 13 00:01:19.010004 kernel: rcu: RCU event tracing is enabled. Aug 13 00:01:19.010020 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Aug 13 00:01:19.010036 kernel: Trampoline variant of Tasks RCU enabled. Aug 13 00:01:19.010051 kernel: Tracing variant of Tasks RCU enabled. Aug 13 00:01:19.010066 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Aug 13 00:01:19.010081 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Aug 13 00:01:19.010096 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Aug 13 00:01:19.012858 kernel: GICv3: 96 SPIs implemented Aug 13 00:01:19.012878 kernel: GICv3: 0 Extended SPIs implemented Aug 13 00:01:19.012894 kernel: GICv3: Distributor has no Range Selector support Aug 13 00:01:19.012909 kernel: Root IRQ handler: gic_handle_irq Aug 13 00:01:19.012925 kernel: GICv3: 16 PPIs implemented Aug 13 00:01:19.012940 kernel: GICv3: CPU0: found redistributor 0 region 0:0x0000000010200000 Aug 13 00:01:19.012955 kernel: ACPI: SRAT not present Aug 13 00:01:19.012971 kernel: ITS [mem 0x10080000-0x1009ffff] Aug 13 00:01:19.012986 kernel: ITS@0x0000000010080000: allocated 8192 Devices @400090000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:01:19.013002 kernel: ITS@0x0000000010080000: allocated 8192 Interrupt Collections @4000a0000 (flat, esz 8, psz 64K, shr 1) Aug 13 00:01:19.013018 kernel: GICv3: using LPI property table @0x00000004000b0000 Aug 13 00:01:19.013041 kernel: ITS: Using hypervisor restricted LPI range [128] Aug 13 00:01:19.013057 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000004000d0000 Aug 13 00:01:19.013072 kernel: arch_timer: cp15 timer(s) running at 83.33MHz (virt). Aug 13 00:01:19.013088 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x13381ebeec, max_idle_ns: 440795203145 ns Aug 13 00:01:19.013127 kernel: sched_clock: 56 bits at 83MHz, resolution 12ns, wraps every 4398046511100ns Aug 13 00:01:19.013147 kernel: Console: colour dummy device 80x25 Aug 13 00:01:19.013164 kernel: printk: console [tty1] enabled Aug 13 00:01:19.013180 kernel: ACPI: Core revision 20210730 Aug 13 00:01:19.013196 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 166.66 BogoMIPS (lpj=83333) Aug 13 00:01:19.013212 kernel: pid_max: default: 32768 minimum: 301 Aug 13 00:01:19.013232 kernel: LSM: Security Framework initializing Aug 13 00:01:19.013248 kernel: SELinux: Initializing. Aug 13 00:01:19.013264 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Aug 13 00:01:19.013280 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Aug 13 00:01:19.013296 kernel: rcu: Hierarchical SRCU implementation. Aug 13 00:01:19.013312 kernel: Platform MSI: ITS@0x10080000 domain created Aug 13 00:01:19.013327 kernel: PCI/MSI: ITS@0x10080000 domain created Aug 13 00:01:19.013343 kernel: Remapping and enabling EFI services. Aug 13 00:01:19.013358 kernel: smp: Bringing up secondary CPUs ... Aug 13 00:01:19.013374 kernel: Detected PIPT I-cache on CPU1 Aug 13 00:01:19.013394 kernel: GICv3: CPU1: found redistributor 1 region 0:0x0000000010220000 Aug 13 00:01:19.013410 kernel: GICv3: CPU1: using allocated LPI pending table @0x00000004000e0000 Aug 13 00:01:19.013426 kernel: CPU1: Booted secondary processor 0x0000000001 [0x410fd083] Aug 13 00:01:19.013441 kernel: smp: Brought up 1 node, 2 CPUs Aug 13 00:01:19.013456 kernel: SMP: Total of 2 processors activated. Aug 13 00:01:19.013472 kernel: CPU features: detected: 32-bit EL0 Support Aug 13 00:01:19.013487 kernel: CPU features: detected: 32-bit EL1 Support Aug 13 00:01:19.013503 kernel: CPU features: detected: CRC32 instructions Aug 13 00:01:19.013518 kernel: CPU: All CPU(s) started at EL1 Aug 13 00:01:19.013538 kernel: alternatives: patching kernel code Aug 13 00:01:19.013554 kernel: devtmpfs: initialized Aug 13 00:01:19.013580 kernel: KASLR disabled due to lack of seed Aug 13 00:01:19.013601 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Aug 13 00:01:19.013617 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Aug 13 00:01:19.013633 kernel: pinctrl core: initialized pinctrl subsystem Aug 13 00:01:19.013650 kernel: SMBIOS 3.0.0 present. Aug 13 00:01:19.013666 kernel: DMI: Amazon EC2 a1.large/, BIOS 1.0 11/1/2018 Aug 13 00:01:19.013682 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Aug 13 00:01:19.013698 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Aug 13 00:01:19.013715 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Aug 13 00:01:19.013735 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Aug 13 00:01:19.013752 kernel: audit: initializing netlink subsys (disabled) Aug 13 00:01:19.013768 kernel: audit: type=2000 audit(0.293:1): state=initialized audit_enabled=0 res=1 Aug 13 00:01:19.013785 kernel: thermal_sys: Registered thermal governor 'step_wise' Aug 13 00:01:19.013801 kernel: cpuidle: using governor menu Aug 13 00:01:19.013821 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Aug 13 00:01:19.013838 kernel: ASID allocator initialised with 32768 entries Aug 13 00:01:19.013854 kernel: ACPI: bus type PCI registered Aug 13 00:01:19.013870 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Aug 13 00:01:19.013887 kernel: Serial: AMBA PL011 UART driver Aug 13 00:01:19.013903 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages Aug 13 00:01:19.013919 kernel: HugeTLB registered 32.0 MiB page size, pre-allocated 0 pages Aug 13 00:01:19.013936 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Aug 13 00:01:19.013952 kernel: HugeTLB registered 64.0 KiB page size, pre-allocated 0 pages Aug 13 00:01:19.013973 kernel: cryptd: max_cpu_qlen set to 1000 Aug 13 00:01:19.013990 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Aug 13 00:01:19.014006 kernel: ACPI: Added _OSI(Module Device) Aug 13 00:01:19.014022 kernel: ACPI: Added _OSI(Processor Device) Aug 13 00:01:19.014038 kernel: ACPI: Added _OSI(Processor Aggregator Device) Aug 13 00:01:19.014054 kernel: ACPI: Added _OSI(Linux-Dell-Video) Aug 13 00:01:19.014070 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Aug 13 00:01:19.014086 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Aug 13 00:01:19.014119 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Aug 13 00:01:19.014139 kernel: ACPI: Interpreter enabled Aug 13 00:01:19.014161 kernel: ACPI: Using GIC for interrupt routing Aug 13 00:01:19.014177 kernel: ACPI: MCFG table detected, 1 entries Aug 13 00:01:19.014193 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-0f]) Aug 13 00:01:19.014470 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:01:19.014656 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Aug 13 00:01:19.014837 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Aug 13 00:01:19.015035 kernel: acpi PNP0A08:00: ECAM area [mem 0x20000000-0x20ffffff] reserved by PNP0C02:00 Aug 13 00:01:19.015248 kernel: acpi PNP0A08:00: ECAM at [mem 0x20000000-0x20ffffff] for [bus 00-0f] Aug 13 00:01:19.015273 kernel: ACPI: Remapped I/O 0x000000001fff0000 to [io 0x0000-0xffff window] Aug 13 00:01:19.015290 kernel: acpiphp: Slot [1] registered Aug 13 00:01:19.015307 kernel: acpiphp: Slot [2] registered Aug 13 00:01:19.015323 kernel: acpiphp: Slot [3] registered Aug 13 00:01:19.015339 kernel: acpiphp: Slot [4] registered Aug 13 00:01:19.015355 kernel: acpiphp: Slot [5] registered Aug 13 00:01:19.015371 kernel: acpiphp: Slot [6] registered Aug 13 00:01:19.015387 kernel: acpiphp: Slot [7] registered Aug 13 00:01:19.015408 kernel: acpiphp: Slot [8] registered Aug 13 00:01:19.015425 kernel: acpiphp: Slot [9] registered Aug 13 00:01:19.015441 kernel: acpiphp: Slot [10] registered Aug 13 00:01:19.015457 kernel: acpiphp: Slot [11] registered Aug 13 00:01:19.015473 kernel: acpiphp: Slot [12] registered Aug 13 00:01:19.015489 kernel: acpiphp: Slot [13] registered Aug 13 00:01:19.015505 kernel: acpiphp: Slot [14] registered Aug 13 00:01:19.015520 kernel: acpiphp: Slot [15] registered Aug 13 00:01:19.015536 kernel: acpiphp: Slot [16] registered Aug 13 00:01:19.015556 kernel: acpiphp: Slot [17] registered Aug 13 00:01:19.015572 kernel: acpiphp: Slot [18] registered Aug 13 00:01:19.015588 kernel: acpiphp: Slot [19] registered Aug 13 00:01:19.015604 kernel: acpiphp: Slot [20] registered Aug 13 00:01:19.015621 kernel: acpiphp: Slot [21] registered Aug 13 00:01:19.015637 kernel: acpiphp: Slot [22] registered Aug 13 00:01:19.015653 kernel: acpiphp: Slot [23] registered Aug 13 00:01:19.015668 kernel: acpiphp: Slot [24] registered Aug 13 00:01:19.015684 kernel: acpiphp: Slot [25] registered Aug 13 00:01:19.015700 kernel: acpiphp: Slot [26] registered Aug 13 00:01:19.015720 kernel: acpiphp: Slot [27] registered Aug 13 00:01:19.015736 kernel: acpiphp: Slot [28] registered Aug 13 00:01:19.015752 kernel: acpiphp: Slot [29] registered Aug 13 00:01:19.015768 kernel: acpiphp: Slot [30] registered Aug 13 00:01:19.015784 kernel: acpiphp: Slot [31] registered Aug 13 00:01:19.015800 kernel: PCI host bridge to bus 0000:00 Aug 13 00:01:19.015992 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xffffffff window] Aug 13 00:01:19.016185 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Aug 13 00:01:19.016360 kernel: pci_bus 0000:00: root bus resource [mem 0x400000000000-0x407fffffffff window] Aug 13 00:01:19.016528 kernel: pci_bus 0000:00: root bus resource [bus 00-0f] Aug 13 00:01:19.016735 kernel: pci 0000:00:00.0: [1d0f:0200] type 00 class 0x060000 Aug 13 00:01:19.016949 kernel: pci 0000:00:01.0: [1d0f:8250] type 00 class 0x070003 Aug 13 00:01:19.027287 kernel: pci 0000:00:01.0: reg 0x10: [mem 0x80118000-0x80118fff] Aug 13 00:01:19.027536 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 Aug 13 00:01:19.027735 kernel: pci 0000:00:04.0: reg 0x10: [mem 0x80114000-0x80117fff] Aug 13 00:01:19.027923 kernel: pci 0000:00:04.0: PME# supported from D0 D1 D2 D3hot D3cold Aug 13 00:01:19.028146 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 Aug 13 00:01:19.028342 kernel: pci 0000:00:05.0: reg 0x10: [mem 0x80110000-0x80113fff] Aug 13 00:01:19.028530 kernel: pci 0000:00:05.0: reg 0x18: [mem 0x80000000-0x800fffff pref] Aug 13 00:01:19.028715 kernel: pci 0000:00:05.0: reg 0x20: [mem 0x80100000-0x8010ffff] Aug 13 00:01:19.028900 kernel: pci 0000:00:05.0: PME# supported from D0 D1 D2 D3hot D3cold Aug 13 00:01:19.029091 kernel: pci 0000:00:05.0: BAR 2: assigned [mem 0x80000000-0x800fffff pref] Aug 13 00:01:19.029310 kernel: pci 0000:00:05.0: BAR 4: assigned [mem 0x80100000-0x8010ffff] Aug 13 00:01:19.029497 kernel: pci 0000:00:04.0: BAR 0: assigned [mem 0x80110000-0x80113fff] Aug 13 00:01:19.029688 kernel: pci 0000:00:05.0: BAR 0: assigned [mem 0x80114000-0x80117fff] Aug 13 00:01:19.029883 kernel: pci 0000:00:01.0: BAR 0: assigned [mem 0x80118000-0x80118fff] Aug 13 00:01:19.030064 kernel: pci_bus 0000:00: resource 4 [mem 0x80000000-0xffffffff window] Aug 13 00:01:19.030286 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Aug 13 00:01:19.030468 kernel: pci_bus 0000:00: resource 6 [mem 0x400000000000-0x407fffffffff window] Aug 13 00:01:19.030503 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Aug 13 00:01:19.030520 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Aug 13 00:01:19.030538 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Aug 13 00:01:19.030554 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Aug 13 00:01:19.030571 kernel: iommu: Default domain type: Translated Aug 13 00:01:19.030599 kernel: iommu: DMA domain TLB invalidation policy: strict mode Aug 13 00:01:19.030617 kernel: vgaarb: loaded Aug 13 00:01:19.030634 kernel: pps_core: LinuxPPS API ver. 1 registered Aug 13 00:01:19.030655 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Aug 13 00:01:19.030672 kernel: PTP clock support registered Aug 13 00:01:19.030698 kernel: Registered efivars operations Aug 13 00:01:19.030718 kernel: clocksource: Switched to clocksource arch_sys_counter Aug 13 00:01:19.030735 kernel: VFS: Disk quotas dquot_6.6.0 Aug 13 00:01:19.030751 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Aug 13 00:01:19.030767 kernel: pnp: PnP ACPI init Aug 13 00:01:19.030989 kernel: system 00:00: [mem 0x20000000-0x2fffffff] could not be reserved Aug 13 00:01:19.031035 kernel: pnp: PnP ACPI: found 1 devices Aug 13 00:01:19.031054 kernel: NET: Registered PF_INET protocol family Aug 13 00:01:19.031083 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Aug 13 00:01:19.031114 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Aug 13 00:01:19.031136 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Aug 13 00:01:19.031153 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Aug 13 00:01:19.031170 kernel: TCP bind hash table entries: 32768 (order: 7, 524288 bytes, linear) Aug 13 00:01:19.031186 kernel: TCP: Hash tables configured (established 32768 bind 32768) Aug 13 00:01:19.031203 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Aug 13 00:01:19.031224 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Aug 13 00:01:19.031241 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Aug 13 00:01:19.031257 kernel: PCI: CLS 0 bytes, default 64 Aug 13 00:01:19.031287 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 3 counters available Aug 13 00:01:19.031304 kernel: kvm [1]: HYP mode not available Aug 13 00:01:19.031320 kernel: Initialise system trusted keyrings Aug 13 00:01:19.031337 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Aug 13 00:01:19.031354 kernel: Key type asymmetric registered Aug 13 00:01:19.031370 kernel: Asymmetric key parser 'x509' registered Aug 13 00:01:19.031391 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Aug 13 00:01:19.031408 kernel: io scheduler mq-deadline registered Aug 13 00:01:19.031435 kernel: io scheduler kyber registered Aug 13 00:01:19.031454 kernel: io scheduler bfq registered Aug 13 00:01:19.031685 kernel: pl061_gpio ARMH0061:00: PL061 GPIO chip registered Aug 13 00:01:19.031711 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Aug 13 00:01:19.031728 kernel: ACPI: button: Power Button [PWRB] Aug 13 00:01:19.031756 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0E:00/input/input1 Aug 13 00:01:19.031776 kernel: ACPI: button: Sleep Button [SLPB] Aug 13 00:01:19.031798 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Aug 13 00:01:19.031815 kernel: ACPI: \_SB_.PCI0.GSI2: Enabled at IRQ 37 Aug 13 00:01:19.032053 kernel: serial 0000:00:01.0: enabling device (0010 -> 0012) Aug 13 00:01:19.032088 kernel: printk: console [ttyS0] disabled Aug 13 00:01:19.047531 kernel: 0000:00:01.0: ttyS0 at MMIO 0x80118000 (irq = 14, base_baud = 115200) is a 16550A Aug 13 00:01:19.047561 kernel: printk: console [ttyS0] enabled Aug 13 00:01:19.047579 kernel: printk: bootconsole [uart0] disabled Aug 13 00:01:19.047596 kernel: thunder_xcv, ver 1.0 Aug 13 00:01:19.047612 kernel: thunder_bgx, ver 1.0 Aug 13 00:01:19.047637 kernel: nicpf, ver 1.0 Aug 13 00:01:19.047653 kernel: nicvf, ver 1.0 Aug 13 00:01:19.047906 kernel: rtc-efi rtc-efi.0: registered as rtc0 Aug 13 00:01:19.048084 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-08-13T00:01:18 UTC (1755043278) Aug 13 00:01:19.048128 kernel: hid: raw HID events driver (C) Jiri Kosina Aug 13 00:01:19.048147 kernel: NET: Registered PF_INET6 protocol family Aug 13 00:01:19.048163 kernel: Segment Routing with IPv6 Aug 13 00:01:19.048180 kernel: In-situ OAM (IOAM) with IPv6 Aug 13 00:01:19.048202 kernel: NET: Registered PF_PACKET protocol family Aug 13 00:01:19.048218 kernel: Key type dns_resolver registered Aug 13 00:01:19.048234 kernel: registered taskstats version 1 Aug 13 00:01:19.048250 kernel: Loading compiled-in X.509 certificates Aug 13 00:01:19.048267 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.189-flatcar: 72b807ae6dac6ab18c2f4ab9460d3472cf28c19d' Aug 13 00:01:19.048283 kernel: Key type .fscrypt registered Aug 13 00:01:19.048299 kernel: Key type fscrypt-provisioning registered Aug 13 00:01:19.048315 kernel: ima: No TPM chip found, activating TPM-bypass! Aug 13 00:01:19.048331 kernel: ima: Allocated hash algorithm: sha1 Aug 13 00:01:19.048351 kernel: ima: No architecture policies found Aug 13 00:01:19.048367 kernel: clk: Disabling unused clocks Aug 13 00:01:19.048383 kernel: Freeing unused kernel memory: 36416K Aug 13 00:01:19.048399 kernel: Run /init as init process Aug 13 00:01:19.048415 kernel: with arguments: Aug 13 00:01:19.048431 kernel: /init Aug 13 00:01:19.048447 kernel: with environment: Aug 13 00:01:19.048463 kernel: HOME=/ Aug 13 00:01:19.048478 kernel: TERM=linux Aug 13 00:01:19.048498 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Aug 13 00:01:19.048520 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Aug 13 00:01:19.048541 systemd[1]: Detected virtualization amazon. Aug 13 00:01:19.048559 systemd[1]: Detected architecture arm64. Aug 13 00:01:19.048576 systemd[1]: Running in initrd. Aug 13 00:01:19.048594 systemd[1]: No hostname configured, using default hostname. Aug 13 00:01:19.048611 systemd[1]: Hostname set to . Aug 13 00:01:19.048633 systemd[1]: Initializing machine ID from VM UUID. Aug 13 00:01:19.048651 systemd[1]: Queued start job for default target initrd.target. Aug 13 00:01:19.050171 systemd[1]: Started systemd-ask-password-console.path. Aug 13 00:01:19.050191 systemd[1]: Reached target cryptsetup.target. Aug 13 00:01:19.050210 systemd[1]: Reached target paths.target. Aug 13 00:01:19.050227 systemd[1]: Reached target slices.target. Aug 13 00:01:19.050244 systemd[1]: Reached target swap.target. Aug 13 00:01:19.050261 systemd[1]: Reached target timers.target. Aug 13 00:01:19.050287 systemd[1]: Listening on iscsid.socket. Aug 13 00:01:19.050305 systemd[1]: Listening on iscsiuio.socket. Aug 13 00:01:19.050322 systemd[1]: Listening on systemd-journald-audit.socket. Aug 13 00:01:19.050340 systemd[1]: Listening on systemd-journald-dev-log.socket. Aug 13 00:01:19.050358 systemd[1]: Listening on systemd-journald.socket. Aug 13 00:01:19.050375 systemd[1]: Listening on systemd-networkd.socket. Aug 13 00:01:19.050392 systemd[1]: Listening on systemd-udevd-control.socket. Aug 13 00:01:19.050410 systemd[1]: Listening on systemd-udevd-kernel.socket. Aug 13 00:01:19.050431 systemd[1]: Reached target sockets.target. Aug 13 00:01:19.050449 systemd[1]: Starting kmod-static-nodes.service... Aug 13 00:01:19.050466 systemd[1]: Finished network-cleanup.service. Aug 13 00:01:19.050484 systemd[1]: Starting systemd-fsck-usr.service... Aug 13 00:01:19.050502 systemd[1]: Starting systemd-journald.service... Aug 13 00:01:19.050519 systemd[1]: Starting systemd-modules-load.service... Aug 13 00:01:19.050537 systemd[1]: Starting systemd-resolved.service... Aug 13 00:01:19.050554 systemd[1]: Starting systemd-vconsole-setup.service... Aug 13 00:01:19.050572 systemd[1]: Finished kmod-static-nodes.service. Aug 13 00:01:19.050593 systemd[1]: Finished systemd-fsck-usr.service. Aug 13 00:01:19.050612 kernel: audit: type=1130 audit(1755043278.994:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.050630 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Aug 13 00:01:19.050648 systemd[1]: Finished systemd-vconsole-setup.service. Aug 13 00:01:19.050666 kernel: audit: type=1130 audit(1755043279.037:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.050684 systemd[1]: Starting dracut-cmdline-ask.service... Aug 13 00:01:19.050704 systemd-journald[310]: Journal started Aug 13 00:01:19.050798 systemd-journald[310]: Runtime Journal (/run/log/journal/ec2a58ee5e450242648ec4eb845689c5) is 8.0M, max 75.4M, 67.4M free. Aug 13 00:01:18.994000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.037000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:18.974913 systemd-modules-load[311]: Inserted module 'overlay' Aug 13 00:01:19.068900 systemd[1]: Started systemd-journald.service. Aug 13 00:01:19.068962 kernel: audit: type=1130 audit(1755043279.056:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.056000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.069801 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Aug 13 00:01:19.069000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.085990 kernel: audit: type=1130 audit(1755043279.069:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.095138 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Aug 13 00:01:19.102325 systemd-modules-load[311]: Inserted module 'br_netfilter' Aug 13 00:01:19.103132 kernel: Bridge firewalling registered Aug 13 00:01:19.110314 systemd-resolved[312]: Positive Trust Anchors: Aug 13 00:01:19.110349 systemd-resolved[312]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 00:01:19.110403 systemd-resolved[312]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Aug 13 00:01:19.141139 kernel: SCSI subsystem initialized Aug 13 00:01:19.158722 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Aug 13 00:01:19.158788 kernel: device-mapper: uevent: version 1.0.3 Aug 13 00:01:19.162223 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Aug 13 00:01:19.167615 systemd-modules-load[311]: Inserted module 'dm_multipath' Aug 13 00:01:19.168999 systemd[1]: Finished dracut-cmdline-ask.service. Aug 13 00:01:19.173000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.174927 systemd[1]: Finished systemd-modules-load.service. Aug 13 00:01:19.187589 systemd[1]: Starting dracut-cmdline.service... Aug 13 00:01:19.183000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.199824 kernel: audit: type=1130 audit(1755043279.173:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.199864 kernel: audit: type=1130 audit(1755043279.183:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.200354 systemd[1]: Starting systemd-sysctl.service... Aug 13 00:01:19.223530 systemd[1]: Finished systemd-sysctl.service. Aug 13 00:01:19.235458 kernel: audit: type=1130 audit(1755043279.224:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.224000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.235557 dracut-cmdline[330]: dracut-dracut-053 Aug 13 00:01:19.240853 dracut-cmdline[330]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlycon flatcar.first_boot=detected acpi=force flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=32404c0887e5b8a80b0f069916a8040bfd969c7a8f47a2db1168b24bc04220cc Aug 13 00:01:19.360141 kernel: Loading iSCSI transport class v2.0-870. Aug 13 00:01:19.381147 kernel: iscsi: registered transport (tcp) Aug 13 00:01:19.409024 kernel: iscsi: registered transport (qla4xxx) Aug 13 00:01:19.409121 kernel: QLogic iSCSI HBA Driver Aug 13 00:01:19.606146 kernel: random: crng init done Aug 13 00:01:19.606233 systemd-resolved[312]: Defaulting to hostname 'linux'. Aug 13 00:01:19.612841 systemd[1]: Started systemd-resolved.service. Aug 13 00:01:19.631634 kernel: audit: type=1130 audit(1755043279.613:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.613000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.614822 systemd[1]: Reached target nss-lookup.target. Aug 13 00:01:19.633397 systemd[1]: Finished dracut-cmdline.service. Aug 13 00:01:19.651212 kernel: audit: type=1130 audit(1755043279.636:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.636000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:19.646410 systemd[1]: Starting dracut-pre-udev.service... Aug 13 00:01:19.713141 kernel: raid6: neonx8 gen() 6350 MB/s Aug 13 00:01:19.731135 kernel: raid6: neonx8 xor() 4747 MB/s Aug 13 00:01:19.749134 kernel: raid6: neonx4 gen() 6473 MB/s Aug 13 00:01:19.767135 kernel: raid6: neonx4 xor() 4913 MB/s Aug 13 00:01:19.785135 kernel: raid6: neonx2 gen() 5734 MB/s Aug 13 00:01:19.803133 kernel: raid6: neonx2 xor() 4524 MB/s Aug 13 00:01:19.821134 kernel: raid6: neonx1 gen() 4461 MB/s Aug 13 00:01:19.839133 kernel: raid6: neonx1 xor() 3678 MB/s Aug 13 00:01:19.857134 kernel: raid6: int64x8 gen() 3411 MB/s Aug 13 00:01:19.875134 kernel: raid6: int64x8 xor() 2085 MB/s Aug 13 00:01:19.893134 kernel: raid6: int64x4 gen() 3780 MB/s Aug 13 00:01:19.911134 kernel: raid6: int64x4 xor() 2191 MB/s Aug 13 00:01:19.929135 kernel: raid6: int64x2 gen() 3563 MB/s Aug 13 00:01:19.947134 kernel: raid6: int64x2 xor() 1948 MB/s Aug 13 00:01:19.965134 kernel: raid6: int64x1 gen() 2762 MB/s Aug 13 00:01:19.984691 kernel: raid6: int64x1 xor() 1451 MB/s Aug 13 00:01:19.984720 kernel: raid6: using algorithm neonx4 gen() 6473 MB/s Aug 13 00:01:19.984744 kernel: raid6: .... xor() 4913 MB/s, rmw enabled Aug 13 00:01:19.986536 kernel: raid6: using neon recovery algorithm Aug 13 00:01:20.005141 kernel: xor: measuring software checksum speed Aug 13 00:01:20.007134 kernel: 8regs : 8792 MB/sec Aug 13 00:01:20.007172 kernel: 32regs : 10031 MB/sec Aug 13 00:01:20.010755 kernel: arm64_neon : 9557 MB/sec Aug 13 00:01:20.010784 kernel: xor: using function: 32regs (10031 MB/sec) Aug 13 00:01:20.109164 kernel: Btrfs loaded, crc32c=crc32c-generic, zoned=no, fsverity=no Aug 13 00:01:20.126461 systemd[1]: Finished dracut-pre-udev.service. Aug 13 00:01:20.127000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:20.128000 audit: BPF prog-id=7 op=LOAD Aug 13 00:01:20.128000 audit: BPF prog-id=8 op=LOAD Aug 13 00:01:20.131805 systemd[1]: Starting systemd-udevd.service... Aug 13 00:01:20.161260 systemd-udevd[509]: Using default interface naming scheme 'v252'. Aug 13 00:01:20.170151 systemd[1]: Started systemd-udevd.service. Aug 13 00:01:20.182000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:20.185366 systemd[1]: Starting dracut-pre-trigger.service... Aug 13 00:01:20.215702 dracut-pre-trigger[528]: rd.md=0: removing MD RAID activation Aug 13 00:01:20.273400 systemd[1]: Finished dracut-pre-trigger.service. Aug 13 00:01:20.275000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:20.278341 systemd[1]: Starting systemd-udev-trigger.service... Aug 13 00:01:20.379000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:20.376141 systemd[1]: Finished systemd-udev-trigger.service. Aug 13 00:01:20.497715 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Aug 13 00:01:20.497782 kernel: ena 0000:00:05.0: enabling device (0010 -> 0012) Aug 13 00:01:20.520451 kernel: ena 0000:00:05.0: ENA device version: 0.10 Aug 13 00:01:20.520693 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Aug 13 00:01:20.520946 kernel: ACPI: \_SB_.PCI0.GSI0: Enabled at IRQ 35 Aug 13 00:01:20.520972 kernel: nvme nvme0: pci function 0000:00:04.0 Aug 13 00:01:20.521239 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80114000, mac addr 06:50:e9:6c:d4:3d Aug 13 00:01:20.525147 kernel: nvme nvme0: 2/0/0 default/read/poll queues Aug 13 00:01:20.533420 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Aug 13 00:01:20.533462 kernel: GPT:9289727 != 16777215 Aug 13 00:01:20.533485 kernel: GPT:Alternate GPT header not at the end of the disk. Aug 13 00:01:20.535710 kernel: GPT:9289727 != 16777215 Aug 13 00:01:20.537056 kernel: GPT: Use GNU Parted to correct GPT errors. Aug 13 00:01:20.539087 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 00:01:20.545668 (udev-worker)[566]: Network interface NamePolicy= disabled on kernel command line. Aug 13 00:01:20.616152 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/nvme0n1p6 scanned by (udev-worker) (561) Aug 13 00:01:20.672340 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Aug 13 00:01:20.727222 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Aug 13 00:01:20.749396 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Aug 13 00:01:20.761704 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Aug 13 00:01:20.766878 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Aug 13 00:01:20.772867 systemd[1]: Starting disk-uuid.service... Aug 13 00:01:20.784461 disk-uuid[664]: Primary Header is updated. Aug 13 00:01:20.784461 disk-uuid[664]: Secondary Entries is updated. Aug 13 00:01:20.784461 disk-uuid[664]: Secondary Header is updated. Aug 13 00:01:20.795141 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 00:01:20.804154 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 00:01:20.813139 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 00:01:21.814135 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 00:01:21.814924 disk-uuid[665]: The operation has completed successfully. Aug 13 00:01:21.981042 systemd[1]: disk-uuid.service: Deactivated successfully. Aug 13 00:01:21.983000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:21.983000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:21.981259 systemd[1]: Finished disk-uuid.service. Aug 13 00:01:21.997300 systemd[1]: Starting verity-setup.service... Aug 13 00:01:22.036684 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Aug 13 00:01:22.127448 systemd[1]: Found device dev-mapper-usr.device. Aug 13 00:01:22.131761 systemd[1]: Finished verity-setup.service. Aug 13 00:01:22.133000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:22.136589 systemd[1]: Mounting sysusr-usr.mount... Aug 13 00:01:22.227296 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Aug 13 00:01:22.227943 systemd[1]: Mounted sysusr-usr.mount. Aug 13 00:01:22.231184 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Aug 13 00:01:22.235298 systemd[1]: Starting ignition-setup.service... Aug 13 00:01:22.245605 systemd[1]: Starting parse-ip-for-networkd.service... Aug 13 00:01:22.272524 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Aug 13 00:01:22.272589 kernel: BTRFS info (device nvme0n1p6): using free space tree Aug 13 00:01:22.272622 kernel: BTRFS info (device nvme0n1p6): has skinny extents Aug 13 00:01:22.285136 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Aug 13 00:01:22.304193 systemd[1]: mnt-oem.mount: Deactivated successfully. Aug 13 00:01:22.319626 systemd[1]: Finished ignition-setup.service. Aug 13 00:01:22.320000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:22.323162 systemd[1]: Starting ignition-fetch-offline.service... Aug 13 00:01:22.398201 systemd[1]: Finished parse-ip-for-networkd.service. Aug 13 00:01:22.400000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:22.402000 audit: BPF prog-id=9 op=LOAD Aug 13 00:01:22.404158 systemd[1]: Starting systemd-networkd.service... Aug 13 00:01:22.453257 systemd-networkd[1177]: lo: Link UP Aug 13 00:01:22.453278 systemd-networkd[1177]: lo: Gained carrier Aug 13 00:01:22.457065 systemd-networkd[1177]: Enumeration completed Aug 13 00:01:22.456000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:22.457554 systemd-networkd[1177]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 00:01:22.457781 systemd[1]: Started systemd-networkd.service. Aug 13 00:01:22.459855 systemd[1]: Reached target network.target. Aug 13 00:01:22.480000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:22.465778 systemd[1]: Starting iscsiuio.service... Aug 13 00:01:22.479082 systemd[1]: Started iscsiuio.service. Aug 13 00:01:22.483283 systemd-networkd[1177]: eth0: Link UP Aug 13 00:01:22.483291 systemd-networkd[1177]: eth0: Gained carrier Aug 13 00:01:22.507854 iscsid[1182]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Aug 13 00:01:22.507854 iscsid[1182]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Aug 13 00:01:22.507854 iscsid[1182]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Aug 13 00:01:22.507854 iscsid[1182]: If using hardware iscsi like qla4xxx this message can be ignored. Aug 13 00:01:22.507854 iscsid[1182]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Aug 13 00:01:22.507854 iscsid[1182]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Aug 13 00:01:22.513000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:22.545000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:22.490763 systemd[1]: Starting iscsid.service... Aug 13 00:01:22.513258 systemd[1]: Started iscsid.service. Aug 13 00:01:22.516329 systemd[1]: Starting dracut-initqueue.service... Aug 13 00:01:22.516422 systemd-networkd[1177]: eth0: DHCPv4 address 172.31.27.151/20, gateway 172.31.16.1 acquired from 172.31.16.1 Aug 13 00:01:22.542907 systemd[1]: Finished dracut-initqueue.service. Aug 13 00:01:22.547867 systemd[1]: Reached target remote-fs-pre.target. Aug 13 00:01:22.559429 systemd[1]: Reached target remote-cryptsetup.target. Aug 13 00:01:22.573082 systemd[1]: Reached target remote-fs.target. Aug 13 00:01:22.577884 systemd[1]: Starting dracut-pre-mount.service... Aug 13 00:01:22.601000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:22.600696 systemd[1]: Finished dracut-pre-mount.service. Aug 13 00:01:23.073770 ignition[1110]: Ignition 2.14.0 Aug 13 00:01:23.075609 ignition[1110]: Stage: fetch-offline Aug 13 00:01:23.077611 ignition[1110]: reading system config file "/usr/lib/ignition/base.d/base.ign" Aug 13 00:01:23.080095 ignition[1110]: parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b Aug 13 00:01:23.103267 ignition[1110]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Aug 13 00:01:23.104219 ignition[1110]: Ignition finished successfully Aug 13 00:01:23.108936 systemd[1]: Finished ignition-fetch-offline.service. Aug 13 00:01:23.124258 kernel: kauditd_printk_skb: 17 callbacks suppressed Aug 13 00:01:23.124294 kernel: audit: type=1130 audit(1755043283.110:28): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.110000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.112598 systemd[1]: Starting ignition-fetch.service... Aug 13 00:01:23.134274 ignition[1201]: Ignition 2.14.0 Aug 13 00:01:23.134301 ignition[1201]: Stage: fetch Aug 13 00:01:23.134590 ignition[1201]: reading system config file "/usr/lib/ignition/base.d/base.ign" Aug 13 00:01:23.134647 ignition[1201]: parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b Aug 13 00:01:23.149817 ignition[1201]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Aug 13 00:01:23.152287 ignition[1201]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Aug 13 00:01:23.162987 ignition[1201]: INFO : PUT result: OK Aug 13 00:01:23.166687 ignition[1201]: DEBUG : parsed url from cmdline: "" Aug 13 00:01:23.166687 ignition[1201]: INFO : no config URL provided Aug 13 00:01:23.166687 ignition[1201]: INFO : reading system config file "/usr/lib/ignition/user.ign" Aug 13 00:01:23.166687 ignition[1201]: INFO : no config at "/usr/lib/ignition/user.ign" Aug 13 00:01:23.166687 ignition[1201]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Aug 13 00:01:23.178294 ignition[1201]: INFO : PUT result: OK Aug 13 00:01:23.178294 ignition[1201]: INFO : GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Aug 13 00:01:23.178294 ignition[1201]: INFO : GET result: OK Aug 13 00:01:23.178294 ignition[1201]: DEBUG : parsing config with SHA512: 5d0f8f1759941c3e0bc7c993e399bdc21bc88e8475dedfcab5747cc947acea0be17de0b000ea1577fd1f015fee3351e0d11ac7f2fe1ad992a3995c3f263884f0 Aug 13 00:01:23.193882 unknown[1201]: fetched base config from "system" Aug 13 00:01:23.193933 unknown[1201]: fetched base config from "system" Aug 13 00:01:23.193949 unknown[1201]: fetched user config from "aws" Aug 13 00:01:23.200533 ignition[1201]: fetch: fetch complete Aug 13 00:01:23.200579 ignition[1201]: fetch: fetch passed Aug 13 00:01:23.200698 ignition[1201]: Ignition finished successfully Aug 13 00:01:23.205666 systemd[1]: Finished ignition-fetch.service. Aug 13 00:01:23.207000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.210297 systemd[1]: Starting ignition-kargs.service... Aug 13 00:01:23.219835 kernel: audit: type=1130 audit(1755043283.207:29): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.232852 ignition[1207]: Ignition 2.14.0 Aug 13 00:01:23.232886 ignition[1207]: Stage: kargs Aug 13 00:01:23.233255 ignition[1207]: reading system config file "/usr/lib/ignition/base.d/base.ign" Aug 13 00:01:23.233310 ignition[1207]: parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b Aug 13 00:01:23.248883 ignition[1207]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Aug 13 00:01:23.251293 ignition[1207]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Aug 13 00:01:23.254534 ignition[1207]: INFO : PUT result: OK Aug 13 00:01:23.260286 ignition[1207]: kargs: kargs passed Aug 13 00:01:23.260391 ignition[1207]: Ignition finished successfully Aug 13 00:01:23.264981 systemd[1]: Finished ignition-kargs.service. Aug 13 00:01:23.267000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.269680 systemd[1]: Starting ignition-disks.service... Aug 13 00:01:23.287084 kernel: audit: type=1130 audit(1755043283.267:30): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.282927 ignition[1213]: Ignition 2.14.0 Aug 13 00:01:23.282942 ignition[1213]: Stage: disks Aug 13 00:01:23.283305 ignition[1213]: reading system config file "/usr/lib/ignition/base.d/base.ign" Aug 13 00:01:23.283358 ignition[1213]: parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b Aug 13 00:01:23.298470 ignition[1213]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Aug 13 00:01:23.300967 ignition[1213]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Aug 13 00:01:23.304214 ignition[1213]: INFO : PUT result: OK Aug 13 00:01:23.309709 ignition[1213]: disks: disks passed Aug 13 00:01:23.309993 ignition[1213]: Ignition finished successfully Aug 13 00:01:23.311714 systemd[1]: Finished ignition-disks.service. Aug 13 00:01:23.316000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.318176 systemd[1]: Reached target initrd-root-device.target. Aug 13 00:01:23.327585 kernel: audit: type=1130 audit(1755043283.316:31): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.327498 systemd[1]: Reached target local-fs-pre.target. Aug 13 00:01:23.331000 systemd[1]: Reached target local-fs.target. Aug 13 00:01:23.334239 systemd[1]: Reached target sysinit.target. Aug 13 00:01:23.337395 systemd[1]: Reached target basic.target. Aug 13 00:01:23.343312 systemd[1]: Starting systemd-fsck-root.service... Aug 13 00:01:23.382557 systemd-fsck[1221]: ROOT: clean, 629/553520 files, 56026/553472 blocks Aug 13 00:01:23.393075 systemd[1]: Finished systemd-fsck-root.service. Aug 13 00:01:23.404463 kernel: audit: type=1130 audit(1755043283.393:32): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.393000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.396412 systemd[1]: Mounting sysroot.mount... Aug 13 00:01:23.427124 kernel: EXT4-fs (nvme0n1p9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Aug 13 00:01:23.427759 systemd[1]: Mounted sysroot.mount. Aug 13 00:01:23.432425 systemd[1]: Reached target initrd-root-fs.target. Aug 13 00:01:23.437932 systemd[1]: Mounting sysroot-usr.mount... Aug 13 00:01:23.438924 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. Aug 13 00:01:23.440152 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Aug 13 00:01:23.440418 systemd[1]: Reached target ignition-diskful.target. Aug 13 00:01:23.458393 systemd[1]: Mounted sysroot-usr.mount. Aug 13 00:01:23.483598 systemd[1]: Mounting sysroot-usr-share-oem.mount... Aug 13 00:01:23.492353 systemd[1]: Starting initrd-setup-root.service... Aug 13 00:01:23.505716 initrd-setup-root[1243]: cut: /sysroot/etc/passwd: No such file or directory Aug 13 00:01:23.521812 initrd-setup-root[1251]: cut: /sysroot/etc/group: No such file or directory Aug 13 00:01:23.529419 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme0n1p6 scanned by mount (1238) Aug 13 00:01:23.536053 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Aug 13 00:01:23.536142 kernel: BTRFS info (device nvme0n1p6): using free space tree Aug 13 00:01:23.538324 kernel: BTRFS info (device nvme0n1p6): has skinny extents Aug 13 00:01:23.539562 initrd-setup-root[1259]: cut: /sysroot/etc/shadow: No such file or directory Aug 13 00:01:23.547861 initrd-setup-root[1283]: cut: /sysroot/etc/gshadow: No such file or directory Aug 13 00:01:23.553164 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Aug 13 00:01:23.558249 systemd[1]: Mounted sysroot-usr-share-oem.mount. Aug 13 00:01:23.751813 systemd[1]: Finished initrd-setup-root.service. Aug 13 00:01:23.754000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.760470 systemd[1]: Starting ignition-mount.service... Aug 13 00:01:23.766139 kernel: audit: type=1130 audit(1755043283.754:33): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.768804 systemd[1]: Starting sysroot-boot.service... Aug 13 00:01:23.779353 systemd[1]: sysusr-usr-share-oem.mount: Deactivated successfully. Aug 13 00:01:23.779735 systemd[1]: sysroot-usr-share-oem.mount: Deactivated successfully. Aug 13 00:01:23.801004 ignition[1303]: INFO : Ignition 2.14.0 Aug 13 00:01:23.804839 ignition[1303]: INFO : Stage: mount Aug 13 00:01:23.804839 ignition[1303]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Aug 13 00:01:23.804839 ignition[1303]: DEBUG : parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b Aug 13 00:01:23.820370 ignition[1303]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Aug 13 00:01:23.823192 ignition[1303]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Aug 13 00:01:23.827129 ignition[1303]: INFO : PUT result: OK Aug 13 00:01:23.832659 ignition[1303]: INFO : mount: mount passed Aug 13 00:01:23.836208 ignition[1303]: INFO : Ignition finished successfully Aug 13 00:01:23.834483 systemd[1]: Finished ignition-mount.service. Aug 13 00:01:23.840000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.843137 systemd[1]: Starting ignition-files.service... Aug 13 00:01:23.868086 kernel: audit: type=1130 audit(1755043283.840:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.868150 kernel: audit: type=1130 audit(1755043283.858:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.858000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:23.853705 systemd[1]: Finished sysroot-boot.service. Aug 13 00:01:23.874545 systemd[1]: Mounting sysroot-usr-share-oem.mount... Aug 13 00:01:23.898146 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/nvme0n1p6 scanned by mount (1313) Aug 13 00:01:23.904417 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Aug 13 00:01:23.904458 kernel: BTRFS info (device nvme0n1p6): using free space tree Aug 13 00:01:23.904491 kernel: BTRFS info (device nvme0n1p6): has skinny extents Aug 13 00:01:23.914138 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Aug 13 00:01:23.919423 systemd[1]: Mounted sysroot-usr-share-oem.mount. Aug 13 00:01:23.938552 ignition[1332]: INFO : Ignition 2.14.0 Aug 13 00:01:23.938552 ignition[1332]: INFO : Stage: files Aug 13 00:01:23.947179 ignition[1332]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Aug 13 00:01:23.947179 ignition[1332]: DEBUG : parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b Aug 13 00:01:23.958837 ignition[1332]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Aug 13 00:01:23.961528 ignition[1332]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Aug 13 00:01:23.964272 ignition[1332]: INFO : PUT result: OK Aug 13 00:01:23.969420 ignition[1332]: DEBUG : files: compiled without relabeling support, skipping Aug 13 00:01:23.975098 ignition[1332]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Aug 13 00:01:23.978181 ignition[1332]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Aug 13 00:01:24.025679 ignition[1332]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Aug 13 00:01:24.028873 ignition[1332]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Aug 13 00:01:24.033445 unknown[1332]: wrote ssh authorized keys file for user: core Aug 13 00:01:24.036823 ignition[1332]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Aug 13 00:01:24.039695 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Aug 13 00:01:24.039695 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Aug 13 00:01:24.039695 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Aug 13 00:01:24.039695 ignition[1332]: INFO : GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Aug 13 00:01:24.099393 systemd-networkd[1177]: eth0: Gained IPv6LL Aug 13 00:01:24.112922 ignition[1332]: INFO : GET result: OK Aug 13 00:01:24.421014 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Aug 13 00:01:24.425331 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 00:01:24.425331 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 00:01:24.425331 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Aug 13 00:01:24.425331 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Aug 13 00:01:24.425331 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/etc/eks/bootstrap.sh" Aug 13 00:01:24.425331 ignition[1332]: INFO : oem config not found in "/usr/share/oem", looking on oem partition Aug 13 00:01:24.457601 ignition[1332]: INFO : op(1): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem4287933938" Aug 13 00:01:24.457601 ignition[1332]: CRITICAL : op(1): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem4287933938": device or resource busy Aug 13 00:01:24.457601 ignition[1332]: ERROR : failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem4287933938", trying btrfs: device or resource busy Aug 13 00:01:24.457601 ignition[1332]: INFO : op(2): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem4287933938" Aug 13 00:01:24.457601 ignition[1332]: INFO : op(2): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem4287933938" Aug 13 00:01:24.474368 ignition[1332]: INFO : op(3): [started] unmounting "/mnt/oem4287933938" Aug 13 00:01:24.474368 ignition[1332]: INFO : op(3): [finished] unmounting "/mnt/oem4287933938" Aug 13 00:01:24.474368 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/etc/eks/bootstrap.sh" Aug 13 00:01:24.474368 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 00:01:24.474368 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 00:01:24.474368 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 00:01:24.474368 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 00:01:24.474368 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/home/core/install.sh" Aug 13 00:01:24.474368 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/home/core/install.sh" Aug 13 00:01:24.474368 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/home/core/nginx.yaml" Aug 13 00:01:24.474368 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/home/core/nginx.yaml" Aug 13 00:01:24.474368 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/etc/systemd/system/nvidia.service" Aug 13 00:01:24.474368 ignition[1332]: INFO : oem config not found in "/usr/share/oem", looking on oem partition Aug 13 00:01:24.522434 systemd[1]: mnt-oem4287933938.mount: Deactivated successfully. Aug 13 00:01:24.538528 ignition[1332]: INFO : op(4): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2047029073" Aug 13 00:01:24.541835 ignition[1332]: CRITICAL : op(4): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2047029073": device or resource busy Aug 13 00:01:24.541835 ignition[1332]: ERROR : failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem2047029073", trying btrfs: device or resource busy Aug 13 00:01:24.541835 ignition[1332]: INFO : op(5): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2047029073" Aug 13 00:01:24.541835 ignition[1332]: INFO : op(5): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2047029073" Aug 13 00:01:24.555485 ignition[1332]: INFO : op(6): [started] unmounting "/mnt/oem2047029073" Aug 13 00:01:24.557831 ignition[1332]: INFO : op(6): [finished] unmounting "/mnt/oem2047029073" Aug 13 00:01:24.560228 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/etc/systemd/system/nvidia.service" Aug 13 00:01:24.560228 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(d): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Aug 13 00:01:24.568688 ignition[1332]: INFO : GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-arm64.raw: attempt #1 Aug 13 00:01:25.024685 ignition[1332]: INFO : GET result: OK Aug 13 00:01:25.480287 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(d): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Aug 13 00:01:25.485090 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(e): [started] writing file "/sysroot/etc/amazon/ssm/amazon-ssm-agent.json" Aug 13 00:01:25.485090 ignition[1332]: INFO : oem config not found in "/usr/share/oem", looking on oem partition Aug 13 00:01:25.502748 ignition[1332]: INFO : op(7): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1480004065" Aug 13 00:01:25.505840 ignition[1332]: CRITICAL : op(7): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1480004065": device or resource busy Aug 13 00:01:25.505840 ignition[1332]: ERROR : failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem1480004065", trying btrfs: device or resource busy Aug 13 00:01:25.505840 ignition[1332]: INFO : op(8): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1480004065" Aug 13 00:01:25.521908 ignition[1332]: INFO : op(8): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1480004065" Aug 13 00:01:25.521908 ignition[1332]: INFO : op(9): [started] unmounting "/mnt/oem1480004065" Aug 13 00:01:25.521908 ignition[1332]: INFO : op(9): [finished] unmounting "/mnt/oem1480004065" Aug 13 00:01:25.521908 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(e): [finished] writing file "/sysroot/etc/amazon/ssm/amazon-ssm-agent.json" Aug 13 00:01:25.521908 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(f): [started] writing file "/sysroot/etc/amazon/ssm/seelog.xml" Aug 13 00:01:25.521908 ignition[1332]: INFO : oem config not found in "/usr/share/oem", looking on oem partition Aug 13 00:01:25.530372 systemd[1]: mnt-oem1480004065.mount: Deactivated successfully. Aug 13 00:01:25.562968 ignition[1332]: INFO : op(a): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1936296168" Aug 13 00:01:25.570183 ignition[1332]: CRITICAL : op(a): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1936296168": device or resource busy Aug 13 00:01:25.570183 ignition[1332]: ERROR : failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem1936296168", trying btrfs: device or resource busy Aug 13 00:01:25.570183 ignition[1332]: INFO : op(b): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1936296168" Aug 13 00:01:25.570183 ignition[1332]: INFO : op(b): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1936296168" Aug 13 00:01:25.570183 ignition[1332]: INFO : op(c): [started] unmounting "/mnt/oem1936296168" Aug 13 00:01:25.570183 ignition[1332]: INFO : op(c): [finished] unmounting "/mnt/oem1936296168" Aug 13 00:01:25.570183 ignition[1332]: INFO : files: createFilesystemsFiles: createFiles: op(f): [finished] writing file "/sysroot/etc/amazon/ssm/seelog.xml" Aug 13 00:01:25.570183 ignition[1332]: INFO : files: op(10): [started] processing unit "amazon-ssm-agent.service" Aug 13 00:01:25.570183 ignition[1332]: INFO : files: op(10): op(11): [started] writing unit "amazon-ssm-agent.service" at "/sysroot/etc/systemd/system/amazon-ssm-agent.service" Aug 13 00:01:25.570183 ignition[1332]: INFO : files: op(10): op(11): [finished] writing unit "amazon-ssm-agent.service" at "/sysroot/etc/systemd/system/amazon-ssm-agent.service" Aug 13 00:01:25.570183 ignition[1332]: INFO : files: op(10): [finished] processing unit "amazon-ssm-agent.service" Aug 13 00:01:25.570183 ignition[1332]: INFO : files: op(12): [started] processing unit "nvidia.service" Aug 13 00:01:25.570183 ignition[1332]: INFO : files: op(12): [finished] processing unit "nvidia.service" Aug 13 00:01:25.570183 ignition[1332]: INFO : files: op(13): [started] processing unit "coreos-metadata-sshkeys@.service" Aug 13 00:01:25.570183 ignition[1332]: INFO : files: op(13): [finished] processing unit "coreos-metadata-sshkeys@.service" Aug 13 00:01:25.570183 ignition[1332]: INFO : files: op(14): [started] processing unit "containerd.service" Aug 13 00:01:25.570183 ignition[1332]: INFO : files: op(14): op(15): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Aug 13 00:01:25.631303 ignition[1332]: INFO : files: op(14): op(15): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Aug 13 00:01:25.631303 ignition[1332]: INFO : files: op(14): [finished] processing unit "containerd.service" Aug 13 00:01:25.631303 ignition[1332]: INFO : files: op(16): [started] processing unit "prepare-helm.service" Aug 13 00:01:25.631303 ignition[1332]: INFO : files: op(16): op(17): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 00:01:25.589377 systemd[1]: mnt-oem1936296168.mount: Deactivated successfully. Aug 13 00:01:25.647461 ignition[1332]: INFO : files: op(16): op(17): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 00:01:25.647461 ignition[1332]: INFO : files: op(16): [finished] processing unit "prepare-helm.service" Aug 13 00:01:25.647461 ignition[1332]: INFO : files: op(18): [started] setting preset to enabled for "amazon-ssm-agent.service" Aug 13 00:01:25.647461 ignition[1332]: INFO : files: op(18): [finished] setting preset to enabled for "amazon-ssm-agent.service" Aug 13 00:01:25.647461 ignition[1332]: INFO : files: op(19): [started] setting preset to enabled for "nvidia.service" Aug 13 00:01:25.647461 ignition[1332]: INFO : files: op(19): [finished] setting preset to enabled for "nvidia.service" Aug 13 00:01:25.647461 ignition[1332]: INFO : files: op(1a): [started] setting preset to enabled for "coreos-metadata-sshkeys@.service " Aug 13 00:01:25.647461 ignition[1332]: INFO : files: op(1a): [finished] setting preset to enabled for "coreos-metadata-sshkeys@.service " Aug 13 00:01:25.647461 ignition[1332]: INFO : files: op(1b): [started] setting preset to enabled for "prepare-helm.service" Aug 13 00:01:25.647461 ignition[1332]: INFO : files: op(1b): [finished] setting preset to enabled for "prepare-helm.service" Aug 13 00:01:25.680780 ignition[1332]: INFO : files: createResultFile: createFiles: op(1c): [started] writing file "/sysroot/etc/.ignition-result.json" Aug 13 00:01:25.684617 ignition[1332]: INFO : files: createResultFile: createFiles: op(1c): [finished] writing file "/sysroot/etc/.ignition-result.json" Aug 13 00:01:25.688260 ignition[1332]: INFO : files: files passed Aug 13 00:01:25.689997 ignition[1332]: INFO : Ignition finished successfully Aug 13 00:01:25.694580 systemd[1]: Finished ignition-files.service. Aug 13 00:01:25.694000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.704408 systemd[1]: Starting initrd-setup-root-after-ignition.service... Aug 13 00:01:25.716920 kernel: audit: type=1130 audit(1755043285.694:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.712677 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Aug 13 00:01:25.714038 systemd[1]: Starting ignition-quench.service... Aug 13 00:01:25.725000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.725000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.736628 initrd-setup-root-after-ignition[1357]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 00:01:25.740040 kernel: audit: type=1130 audit(1755043285.725:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.721689 systemd[1]: ignition-quench.service: Deactivated successfully. Aug 13 00:01:25.741000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.721879 systemd[1]: Finished ignition-quench.service. Aug 13 00:01:25.735553 systemd[1]: Finished initrd-setup-root-after-ignition.service. Aug 13 00:01:25.742343 systemd[1]: Reached target ignition-complete.target. Aug 13 00:01:25.751216 systemd[1]: Starting initrd-parse-etc.service... Aug 13 00:01:25.785369 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Aug 13 00:01:25.787666 systemd[1]: Finished initrd-parse-etc.service. Aug 13 00:01:25.789000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.790000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.791374 systemd[1]: Reached target initrd-fs.target. Aug 13 00:01:25.794572 systemd[1]: Reached target initrd.target. Aug 13 00:01:25.797746 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Aug 13 00:01:25.801847 systemd[1]: Starting dracut-pre-pivot.service... Aug 13 00:01:25.825000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.824978 systemd[1]: Finished dracut-pre-pivot.service. Aug 13 00:01:25.828586 systemd[1]: Starting initrd-cleanup.service... Aug 13 00:01:25.855765 systemd[1]: Stopped target nss-lookup.target. Aug 13 00:01:25.856428 systemd[1]: Stopped target remote-cryptsetup.target. Aug 13 00:01:25.863000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.857726 systemd[1]: Stopped target timers.target. Aug 13 00:01:25.861021 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Aug 13 00:01:25.862579 systemd[1]: Stopped dracut-pre-pivot.service. Aug 13 00:01:25.866075 systemd[1]: Stopped target initrd.target. Aug 13 00:01:25.867806 systemd[1]: Stopped target basic.target. Aug 13 00:01:25.875462 systemd[1]: Stopped target ignition-complete.target. Aug 13 00:01:25.876572 systemd[1]: Stopped target ignition-diskful.target. Aug 13 00:01:25.880287 systemd[1]: Stopped target initrd-root-device.target. Aug 13 00:01:25.883582 systemd[1]: Stopped target remote-fs.target. Aug 13 00:01:25.886910 systemd[1]: Stopped target remote-fs-pre.target. Aug 13 00:01:25.889990 systemd[1]: Stopped target sysinit.target. Aug 13 00:01:25.907000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.893442 systemd[1]: Stopped target local-fs.target. Aug 13 00:01:25.912000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.896697 systemd[1]: Stopped target local-fs-pre.target. Aug 13 00:01:25.915000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.900039 systemd[1]: Stopped target swap.target. Aug 13 00:01:25.919000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.903269 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Aug 13 00:01:25.903367 systemd[1]: Stopped dracut-pre-mount.service. Aug 13 00:01:25.908274 systemd[1]: Stopped target cryptsetup.target. Aug 13 00:01:25.910069 systemd[1]: dracut-initqueue.service: Deactivated successfully. Aug 13 00:01:25.910514 systemd[1]: Stopped dracut-initqueue.service. Aug 13 00:01:25.942000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.947000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.949000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:25.913740 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Aug 13 00:01:25.913823 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Aug 13 00:01:25.917457 systemd[1]: ignition-files.service: Deactivated successfully. Aug 13 00:01:25.917539 systemd[1]: Stopped ignition-files.service. Aug 13 00:01:25.921807 systemd[1]: Stopping ignition-mount.service... Aug 13 00:01:25.936328 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Aug 13 00:01:25.975170 ignition[1370]: INFO : Ignition 2.14.0 Aug 13 00:01:25.975170 ignition[1370]: INFO : Stage: umount Aug 13 00:01:25.975170 ignition[1370]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Aug 13 00:01:25.975170 ignition[1370]: DEBUG : parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b Aug 13 00:01:25.936442 systemd[1]: Stopped kmod-static-nodes.service. Aug 13 00:01:25.944635 systemd[1]: Stopping sysroot-boot.service... Aug 13 00:01:25.946389 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Aug 13 00:01:25.946538 systemd[1]: Stopped systemd-udev-trigger.service. Aug 13 00:01:25.948692 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Aug 13 00:01:25.948788 systemd[1]: Stopped dracut-pre-trigger.service. Aug 13 00:01:25.951555 systemd[1]: initrd-cleanup.service: Deactivated successfully. Aug 13 00:01:26.007000 systemd[1]: Finished initrd-cleanup.service. Aug 13 00:01:26.009000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.009000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.016157 ignition[1370]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Aug 13 00:01:26.020812 systemd[1]: sysroot-boot.mount: Deactivated successfully. Aug 13 00:01:26.027501 ignition[1370]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Aug 13 00:01:26.033190 ignition[1370]: INFO : PUT result: OK Aug 13 00:01:26.036516 ignition[1370]: INFO : umount: umount passed Aug 13 00:01:26.038400 ignition[1370]: INFO : Ignition finished successfully Aug 13 00:01:26.041334 systemd[1]: sysroot-boot.service: Deactivated successfully. Aug 13 00:01:26.043717 systemd[1]: Stopped sysroot-boot.service. Aug 13 00:01:26.045000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.047300 systemd[1]: ignition-mount.service: Deactivated successfully. Aug 13 00:01:26.049366 systemd[1]: Stopped ignition-mount.service. Aug 13 00:01:26.050000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.053000 systemd[1]: ignition-disks.service: Deactivated successfully. Aug 13 00:01:26.053096 systemd[1]: Stopped ignition-disks.service. Aug 13 00:01:26.057000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.058667 systemd[1]: ignition-kargs.service: Deactivated successfully. Aug 13 00:01:26.058746 systemd[1]: Stopped ignition-kargs.service. Aug 13 00:01:26.061000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.062299 systemd[1]: ignition-fetch.service: Deactivated successfully. Aug 13 00:01:26.064000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.063821 systemd[1]: Stopped ignition-fetch.service. Aug 13 00:01:26.065679 systemd[1]: Stopped target network.target. Aug 13 00:01:26.072541 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Aug 13 00:01:26.072632 systemd[1]: Stopped ignition-fetch-offline.service. Aug 13 00:01:26.075000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.076541 systemd[1]: Stopped target paths.target. Aug 13 00:01:26.081378 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Aug 13 00:01:26.084176 systemd[1]: Stopped systemd-ask-password-console.path. Aug 13 00:01:26.087715 systemd[1]: Stopped target slices.target. Aug 13 00:01:26.089350 systemd[1]: Stopped target sockets.target. Aug 13 00:01:26.092420 systemd[1]: iscsid.socket: Deactivated successfully. Aug 13 00:01:26.093879 systemd[1]: Closed iscsid.socket. Aug 13 00:01:26.095416 systemd[1]: iscsiuio.socket: Deactivated successfully. Aug 13 00:01:26.100444 systemd[1]: Closed iscsiuio.socket. Aug 13 00:01:26.103332 systemd[1]: ignition-setup.service: Deactivated successfully. Aug 13 00:01:26.103426 systemd[1]: Stopped ignition-setup.service. Aug 13 00:01:26.105000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.106911 systemd[1]: initrd-setup-root.service: Deactivated successfully. Aug 13 00:01:26.109000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.108435 systemd[1]: Stopped initrd-setup-root.service. Aug 13 00:01:26.112399 systemd[1]: Stopping systemd-networkd.service... Aug 13 00:01:26.115659 systemd[1]: Stopping systemd-resolved.service... Aug 13 00:01:26.118168 systemd-networkd[1177]: eth0: DHCPv6 lease lost Aug 13 00:01:26.123000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.120504 systemd[1]: systemd-networkd.service: Deactivated successfully. Aug 13 00:01:26.125000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.125000 audit: BPF prog-id=9 op=UNLOAD Aug 13 00:01:26.120700 systemd[1]: Stopped systemd-networkd.service. Aug 13 00:01:26.131000 audit: BPF prog-id=6 op=UNLOAD Aug 13 00:01:26.125735 systemd[1]: systemd-resolved.service: Deactivated successfully. Aug 13 00:01:26.125935 systemd[1]: Stopped systemd-resolved.service. Aug 13 00:01:26.141000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.143000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.127233 systemd[1]: systemd-networkd.socket: Deactivated successfully. Aug 13 00:01:26.127300 systemd[1]: Closed systemd-networkd.socket. Aug 13 00:01:26.149000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.133984 systemd[1]: Stopping network-cleanup.service... Aug 13 00:01:26.138772 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Aug 13 00:01:26.138889 systemd[1]: Stopped parse-ip-for-networkd.service. Aug 13 00:01:26.143073 systemd[1]: systemd-sysctl.service: Deactivated successfully. Aug 13 00:01:26.143187 systemd[1]: Stopped systemd-sysctl.service. Aug 13 00:01:26.146362 systemd[1]: systemd-modules-load.service: Deactivated successfully. Aug 13 00:01:26.146460 systemd[1]: Stopped systemd-modules-load.service. Aug 13 00:01:26.155987 systemd[1]: Stopping systemd-udevd.service... Aug 13 00:01:26.171830 systemd[1]: network-cleanup.service: Deactivated successfully. Aug 13 00:01:26.172045 systemd[1]: Stopped network-cleanup.service. Aug 13 00:01:26.182000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.184441 systemd[1]: systemd-udevd.service: Deactivated successfully. Aug 13 00:01:26.184967 systemd[1]: Stopped systemd-udevd.service. Aug 13 00:01:26.188000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.190424 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Aug 13 00:01:26.191029 systemd[1]: Closed systemd-udevd-control.socket. Aug 13 00:01:26.196253 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Aug 13 00:01:26.197158 systemd[1]: Closed systemd-udevd-kernel.socket. Aug 13 00:01:26.201942 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Aug 13 00:01:26.201000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.202028 systemd[1]: Stopped dracut-pre-udev.service. Aug 13 00:01:26.206000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.202895 systemd[1]: dracut-cmdline.service: Deactivated successfully. Aug 13 00:01:26.209000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.202989 systemd[1]: Stopped dracut-cmdline.service. Aug 13 00:01:26.207365 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 00:01:26.208897 systemd[1]: Stopped dracut-cmdline-ask.service. Aug 13 00:01:26.218966 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Aug 13 00:01:26.224000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.223228 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 00:01:26.223370 systemd[1]: Stopped systemd-vconsole-setup.service. Aug 13 00:01:26.238144 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Aug 13 00:01:26.240732 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Aug 13 00:01:26.243000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.243000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:26.244758 systemd[1]: Reached target initrd-switch-root.target. Aug 13 00:01:26.249709 systemd[1]: Starting initrd-switch-root.service... Aug 13 00:01:26.277247 systemd[1]: Switching root. Aug 13 00:01:26.278000 audit: BPF prog-id=8 op=UNLOAD Aug 13 00:01:26.279000 audit: BPF prog-id=7 op=UNLOAD Aug 13 00:01:26.282000 audit: BPF prog-id=5 op=UNLOAD Aug 13 00:01:26.282000 audit: BPF prog-id=4 op=UNLOAD Aug 13 00:01:26.282000 audit: BPF prog-id=3 op=UNLOAD Aug 13 00:01:26.298359 iscsid[1182]: iscsid shutting down. Aug 13 00:01:26.302067 systemd-journald[310]: Received SIGTERM from PID 1 (systemd). Aug 13 00:01:26.302215 systemd-journald[310]: Journal stopped Aug 13 00:01:32.359035 kernel: SELinux: Class mctp_socket not defined in policy. Aug 13 00:01:32.359183 kernel: SELinux: Class anon_inode not defined in policy. Aug 13 00:01:32.359289 kernel: SELinux: the above unknown classes and permissions will be allowed Aug 13 00:01:32.359339 kernel: SELinux: policy capability network_peer_controls=1 Aug 13 00:01:32.359373 kernel: SELinux: policy capability open_perms=1 Aug 13 00:01:32.359409 kernel: SELinux: policy capability extended_socket_class=1 Aug 13 00:01:32.359441 kernel: SELinux: policy capability always_check_network=0 Aug 13 00:01:32.359471 kernel: SELinux: policy capability cgroup_seclabel=1 Aug 13 00:01:32.359500 kernel: SELinux: policy capability nnp_nosuid_transition=1 Aug 13 00:01:32.359531 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Aug 13 00:01:32.359561 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Aug 13 00:01:32.359597 systemd[1]: Successfully loaded SELinux policy in 123.445ms. Aug 13 00:01:32.360239 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 19.378ms. Aug 13 00:01:32.360284 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Aug 13 00:01:32.360317 systemd[1]: Detected virtualization amazon. Aug 13 00:01:32.360348 systemd[1]: Detected architecture arm64. Aug 13 00:01:32.360378 systemd[1]: Detected first boot. Aug 13 00:01:32.360411 systemd[1]: Initializing machine ID from VM UUID. Aug 13 00:01:32.360480 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Aug 13 00:01:32.360513 systemd[1]: Populated /etc with preset unit settings. Aug 13 00:01:32.360548 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Aug 13 00:01:32.360585 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Aug 13 00:01:32.360618 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 00:01:32.360655 systemd[1]: Queued start job for default target multi-user.target. Aug 13 00:01:32.360697 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device. Aug 13 00:01:32.360732 systemd[1]: Created slice system-addon\x2dconfig.slice. Aug 13 00:01:32.360771 systemd[1]: Created slice system-addon\x2drun.slice. Aug 13 00:01:32.360808 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice. Aug 13 00:01:32.360846 systemd[1]: Created slice system-getty.slice. Aug 13 00:01:32.360883 systemd[1]: Created slice system-modprobe.slice. Aug 13 00:01:32.360913 systemd[1]: Created slice system-serial\x2dgetty.slice. Aug 13 00:01:32.360947 systemd[1]: Created slice system-system\x2dcloudinit.slice. Aug 13 00:01:32.360980 systemd[1]: Created slice system-systemd\x2dfsck.slice. Aug 13 00:01:32.361014 systemd[1]: Created slice user.slice. Aug 13 00:01:32.361048 systemd[1]: Started systemd-ask-password-console.path. Aug 13 00:01:32.361086 systemd[1]: Started systemd-ask-password-wall.path. Aug 13 00:01:32.361179 systemd[1]: Set up automount boot.automount. Aug 13 00:01:32.361213 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Aug 13 00:01:32.361246 systemd[1]: Reached target integritysetup.target. Aug 13 00:01:32.361278 systemd[1]: Reached target remote-cryptsetup.target. Aug 13 00:01:32.361324 systemd[1]: Reached target remote-fs.target. Aug 13 00:01:32.361357 systemd[1]: Reached target slices.target. Aug 13 00:01:32.361390 systemd[1]: Reached target swap.target. Aug 13 00:01:32.361424 systemd[1]: Reached target torcx.target. Aug 13 00:01:32.361459 systemd[1]: Reached target veritysetup.target. Aug 13 00:01:32.361492 systemd[1]: Listening on systemd-coredump.socket. Aug 13 00:01:32.361522 systemd[1]: Listening on systemd-initctl.socket. Aug 13 00:01:32.361553 kernel: kauditd_printk_skb: 54 callbacks suppressed Aug 13 00:01:32.361584 kernel: audit: type=1400 audit(1755043291.986:85): avc: denied { audit_read } for pid=1 comm="systemd" capability=37 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Aug 13 00:01:32.361614 systemd[1]: Listening on systemd-journald-audit.socket. Aug 13 00:01:32.361643 kernel: audit: type=1335 audit(1755043291.988:86): pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Aug 13 00:01:32.361676 systemd[1]: Listening on systemd-journald-dev-log.socket. Aug 13 00:01:32.361708 systemd[1]: Listening on systemd-journald.socket. Aug 13 00:01:32.361741 systemd[1]: Listening on systemd-networkd.socket. Aug 13 00:01:32.361773 systemd[1]: Listening on systemd-udevd-control.socket. Aug 13 00:01:32.361803 systemd[1]: Listening on systemd-udevd-kernel.socket. Aug 13 00:01:32.361832 systemd[1]: Listening on systemd-userdbd.socket. Aug 13 00:01:32.361864 systemd[1]: Mounting dev-hugepages.mount... Aug 13 00:01:32.361896 systemd[1]: Mounting dev-mqueue.mount... Aug 13 00:01:32.361925 systemd[1]: Mounting media.mount... Aug 13 00:01:32.361960 systemd[1]: Mounting sys-kernel-debug.mount... Aug 13 00:01:32.361995 systemd[1]: Mounting sys-kernel-tracing.mount... Aug 13 00:01:32.362027 systemd[1]: Mounting tmp.mount... Aug 13 00:01:32.362057 systemd[1]: Starting flatcar-tmpfiles.service... Aug 13 00:01:32.362088 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Aug 13 00:01:32.379184 systemd[1]: Starting kmod-static-nodes.service... Aug 13 00:01:32.379232 systemd[1]: Starting modprobe@configfs.service... Aug 13 00:01:32.379264 systemd[1]: Starting modprobe@dm_mod.service... Aug 13 00:01:32.379297 systemd[1]: Starting modprobe@drm.service... Aug 13 00:01:32.379335 systemd[1]: Starting modprobe@efi_pstore.service... Aug 13 00:01:32.379366 systemd[1]: Starting modprobe@fuse.service... Aug 13 00:01:32.379398 systemd[1]: Starting modprobe@loop.service... Aug 13 00:01:32.379430 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Aug 13 00:01:32.379460 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Aug 13 00:01:32.379491 systemd[1]: (This warning is only shown for the first unit using IP firewalling.) Aug 13 00:01:32.379520 systemd[1]: Starting systemd-journald.service... Aug 13 00:01:32.379549 systemd[1]: Starting systemd-modules-load.service... Aug 13 00:01:32.379582 systemd[1]: Starting systemd-network-generator.service... Aug 13 00:01:32.379612 systemd[1]: Starting systemd-remount-fs.service... Aug 13 00:01:32.379643 systemd[1]: Starting systemd-udev-trigger.service... Aug 13 00:01:32.379674 systemd[1]: Mounted dev-hugepages.mount. Aug 13 00:01:32.379707 kernel: fuse: init (API version 7.34) Aug 13 00:01:32.379737 systemd[1]: Mounted dev-mqueue.mount. Aug 13 00:01:32.379765 kernel: loop: module loaded Aug 13 00:01:32.379793 systemd[1]: Mounted media.mount. Aug 13 00:01:32.379823 systemd[1]: Mounted sys-kernel-debug.mount. Aug 13 00:01:32.379853 systemd[1]: Mounted sys-kernel-tracing.mount. Aug 13 00:01:32.379886 systemd[1]: Mounted tmp.mount. Aug 13 00:01:32.379916 systemd[1]: Finished kmod-static-nodes.service. Aug 13 00:01:32.379947 kernel: audit: type=1130 audit(1755043292.293:87): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.379978 systemd[1]: modprobe@configfs.service: Deactivated successfully. Aug 13 00:01:32.380008 systemd[1]: Finished modprobe@configfs.service. Aug 13 00:01:32.380039 kernel: audit: type=1130 audit(1755043292.320:88): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.380068 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 00:01:32.380098 kernel: audit: type=1131 audit(1755043292.320:89): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.380152 systemd[1]: Finished modprobe@dm_mod.service. Aug 13 00:01:32.380184 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 00:01:32.380214 kernel: audit: type=1305 audit(1755043292.346:90): op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Aug 13 00:01:32.380245 systemd[1]: Finished modprobe@drm.service. Aug 13 00:01:32.380281 kernel: audit: type=1300 audit(1755043292.346:90): arch=c00000b7 syscall=211 success=yes exit=60 a0=5 a1=fffffc7a9400 a2=4000 a3=1 items=0 ppid=1 pid=1523 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:32.380312 systemd-journald[1523]: Journal started Aug 13 00:01:32.380414 systemd-journald[1523]: Runtime Journal (/run/log/journal/ec2a58ee5e450242648ec4eb845689c5) is 8.0M, max 75.4M, 67.4M free. Aug 13 00:01:31.988000 audit[1]: EVENT_LISTENER pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Aug 13 00:01:32.293000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.320000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.320000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.346000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Aug 13 00:01:32.346000 audit[1523]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=60 a0=5 a1=fffffc7a9400 a2=4000 a3=1 items=0 ppid=1 pid=1523 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:32.388638 systemd[1]: Started systemd-journald.service. Aug 13 00:01:32.388716 kernel: audit: type=1327 audit(1755043292.346:90): proctitle="/usr/lib/systemd/systemd-journald" Aug 13 00:01:32.346000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Aug 13 00:01:32.399851 kernel: audit: type=1130 audit(1755043292.346:91): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.346000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.346000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.403074 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 00:01:32.403475 systemd[1]: Finished modprobe@efi_pstore.service. Aug 13 00:01:32.381000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.381000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.399000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.419163 kernel: audit: type=1131 audit(1755043292.346:92): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.417000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.417000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.419441 systemd[1]: modprobe@fuse.service: Deactivated successfully. Aug 13 00:01:32.419813 systemd[1]: Finished modprobe@fuse.service. Aug 13 00:01:32.423000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.423000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.425084 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 00:01:32.425523 systemd[1]: Finished modprobe@loop.service. Aug 13 00:01:32.430000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.430000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.432564 systemd[1]: Finished systemd-modules-load.service. Aug 13 00:01:32.437000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.439656 systemd[1]: Finished systemd-network-generator.service. Aug 13 00:01:32.446000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.448730 systemd[1]: Finished flatcar-tmpfiles.service. Aug 13 00:01:32.451000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.454433 systemd[1]: Finished systemd-remount-fs.service. Aug 13 00:01:32.457000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.459722 systemd[1]: Reached target network-pre.target. Aug 13 00:01:32.470324 systemd[1]: Mounting sys-fs-fuse-connections.mount... Aug 13 00:01:32.481383 systemd[1]: Mounting sys-kernel-config.mount... Aug 13 00:01:32.485542 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Aug 13 00:01:32.504321 systemd[1]: Starting systemd-hwdb-update.service... Aug 13 00:01:32.513476 systemd[1]: Starting systemd-journal-flush.service... Aug 13 00:01:32.517753 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 00:01:32.520242 systemd[1]: Starting systemd-random-seed.service... Aug 13 00:01:32.528306 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Aug 13 00:01:32.530807 systemd[1]: Starting systemd-sysctl.service... Aug 13 00:01:32.537349 systemd[1]: Starting systemd-sysusers.service... Aug 13 00:01:32.548301 systemd[1]: Mounted sys-fs-fuse-connections.mount. Aug 13 00:01:32.555954 systemd[1]: Mounted sys-kernel-config.mount. Aug 13 00:01:32.566917 systemd[1]: Finished systemd-random-seed.service. Aug 13 00:01:32.570000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.571706 systemd[1]: Reached target first-boot-complete.target. Aug 13 00:01:32.577339 systemd-journald[1523]: Time spent on flushing to /var/log/journal/ec2a58ee5e450242648ec4eb845689c5 is 83.759ms for 1071 entries. Aug 13 00:01:32.577339 systemd-journald[1523]: System Journal (/var/log/journal/ec2a58ee5e450242648ec4eb845689c5) is 8.0M, max 195.6M, 187.6M free. Aug 13 00:01:32.683172 systemd-journald[1523]: Received client request to flush runtime journal. Aug 13 00:01:32.628000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.650000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.625385 systemd[1]: Finished systemd-sysctl.service. Aug 13 00:01:32.683828 udevadm[1572]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Aug 13 00:01:32.646965 systemd[1]: Finished systemd-udev-trigger.service. Aug 13 00:01:32.654223 systemd[1]: Starting systemd-udev-settle.service... Aug 13 00:01:32.690000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.686310 systemd[1]: Finished systemd-journal-flush.service. Aug 13 00:01:32.816509 systemd[1]: Finished systemd-sysusers.service. Aug 13 00:01:32.819000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:32.823489 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Aug 13 00:01:32.940033 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Aug 13 00:01:32.946000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:33.399087 systemd[1]: Finished systemd-hwdb-update.service. Aug 13 00:01:33.406000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:33.409826 systemd[1]: Starting systemd-udevd.service... Aug 13 00:01:33.449330 systemd-udevd[1580]: Using default interface naming scheme 'v252'. Aug 13 00:01:33.500654 systemd[1]: Started systemd-udevd.service. Aug 13 00:01:33.501000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:33.505770 systemd[1]: Starting systemd-networkd.service... Aug 13 00:01:33.516590 systemd[1]: Starting systemd-userdbd.service... Aug 13 00:01:33.587395 systemd[1]: Found device dev-ttyS0.device. Aug 13 00:01:33.628173 (udev-worker)[1593]: Network interface NamePolicy= disabled on kernel command line. Aug 13 00:01:33.636560 systemd[1]: Started systemd-userdbd.service. Aug 13 00:01:33.637000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:33.828660 systemd-networkd[1587]: lo: Link UP Aug 13 00:01:33.828683 systemd-networkd[1587]: lo: Gained carrier Aug 13 00:01:33.829667 systemd-networkd[1587]: Enumeration completed Aug 13 00:01:33.832000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:33.829885 systemd[1]: Started systemd-networkd.service. Aug 13 00:01:33.832188 systemd-networkd[1587]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 00:01:33.836543 systemd[1]: Starting systemd-networkd-wait-online.service... Aug 13 00:01:33.848190 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Aug 13 00:01:33.842820 systemd-networkd[1587]: eth0: Link UP Aug 13 00:01:33.843166 systemd-networkd[1587]: eth0: Gained carrier Aug 13 00:01:33.871351 systemd-networkd[1587]: eth0: DHCPv4 address 172.31.27.151/20, gateway 172.31.16.1 acquired from 172.31.16.1 Aug 13 00:01:34.008520 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Aug 13 00:01:34.013000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:34.012959 systemd[1]: Finished systemd-udev-settle.service. Aug 13 00:01:34.044627 systemd[1]: Starting lvm2-activation-early.service... Aug 13 00:01:34.104029 lvm[1700]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Aug 13 00:01:34.141831 systemd[1]: Finished lvm2-activation-early.service. Aug 13 00:01:34.143000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:34.144496 systemd[1]: Reached target cryptsetup.target. Aug 13 00:01:34.149298 systemd[1]: Starting lvm2-activation.service... Aug 13 00:01:34.158633 lvm[1702]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Aug 13 00:01:34.191758 systemd[1]: Finished lvm2-activation.service. Aug 13 00:01:34.192000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:34.193894 systemd[1]: Reached target local-fs-pre.target. Aug 13 00:01:34.195891 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Aug 13 00:01:34.195947 systemd[1]: Reached target local-fs.target. Aug 13 00:01:34.197801 systemd[1]: Reached target machines.target. Aug 13 00:01:34.207215 systemd[1]: Starting ldconfig.service... Aug 13 00:01:34.216740 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Aug 13 00:01:34.216854 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Aug 13 00:01:34.219252 systemd[1]: Starting systemd-boot-update.service... Aug 13 00:01:34.223600 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Aug 13 00:01:34.228866 systemd[1]: Starting systemd-machine-id-commit.service... Aug 13 00:01:34.234087 systemd[1]: Starting systemd-sysext.service... Aug 13 00:01:34.254651 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1705 (bootctl) Aug 13 00:01:34.257189 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Aug 13 00:01:34.272238 systemd[1]: Unmounting usr-share-oem.mount... Aug 13 00:01:34.282763 systemd[1]: usr-share-oem.mount: Deactivated successfully. Aug 13 00:01:34.283419 systemd[1]: Unmounted usr-share-oem.mount. Aug 13 00:01:34.311671 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Aug 13 00:01:34.317000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:34.330193 kernel: loop0: detected capacity change from 0 to 203944 Aug 13 00:01:34.443141 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Aug 13 00:01:34.456779 systemd-fsck[1719]: fsck.fat 4.2 (2021-01-31) Aug 13 00:01:34.456779 systemd-fsck[1719]: /dev/nvme0n1p1: 236 files, 117307/258078 clusters Aug 13 00:01:34.459610 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Aug 13 00:01:34.461000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:34.464847 systemd[1]: Mounting boot.mount... Aug 13 00:01:34.489223 kernel: loop1: detected capacity change from 0 to 203944 Aug 13 00:01:34.506875 systemd[1]: Mounted boot.mount. Aug 13 00:01:34.527920 (sd-sysext)[1725]: Using extensions 'kubernetes'. Aug 13 00:01:34.534224 (sd-sysext)[1725]: Merged extensions into '/usr'. Aug 13 00:01:34.540000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:34.539968 systemd[1]: Finished systemd-boot-update.service. Aug 13 00:01:34.594196 systemd[1]: Mounting usr-share-oem.mount... Aug 13 00:01:34.596866 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Aug 13 00:01:34.599619 systemd[1]: Starting modprobe@dm_mod.service... Aug 13 00:01:34.606453 systemd[1]: Starting modprobe@efi_pstore.service... Aug 13 00:01:34.617341 systemd[1]: Starting modprobe@loop.service... Aug 13 00:01:34.619220 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Aug 13 00:01:34.619555 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Aug 13 00:01:34.630909 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Aug 13 00:01:34.638810 systemd[1]: Finished systemd-machine-id-commit.service. Aug 13 00:01:34.642000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:34.643900 systemd[1]: Mounted usr-share-oem.mount. Aug 13 00:01:34.646665 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 00:01:34.647057 systemd[1]: Finished modprobe@dm_mod.service. Aug 13 00:01:34.648000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:34.648000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:34.652000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:34.652000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:34.650676 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 00:01:34.651082 systemd[1]: Finished modprobe@efi_pstore.service. Aug 13 00:01:34.654575 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 00:01:34.655867 systemd[1]: Finished modprobe@loop.service. Aug 13 00:01:34.656000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:34.657000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:34.658896 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 00:01:34.659147 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Aug 13 00:01:34.662052 systemd[1]: Finished systemd-sysext.service. Aug 13 00:01:34.666000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:34.670492 systemd[1]: Starting ensure-sysext.service... Aug 13 00:01:34.682667 systemd[1]: Starting systemd-tmpfiles-setup.service... Aug 13 00:01:34.699438 systemd[1]: Reloading. Aug 13 00:01:34.712761 systemd-tmpfiles[1754]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Aug 13 00:01:34.723543 systemd-tmpfiles[1754]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Aug 13 00:01:34.735045 systemd-tmpfiles[1754]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Aug 13 00:01:34.812347 /usr/lib/systemd/system-generators/torcx-generator[1773]: time="2025-08-13T00:01:34Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Aug 13 00:01:34.812411 /usr/lib/systemd/system-generators/torcx-generator[1773]: time="2025-08-13T00:01:34Z" level=info msg="torcx already run" Aug 13 00:01:35.068389 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Aug 13 00:01:35.068916 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Aug 13 00:01:35.126804 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 00:01:35.170328 systemd-networkd[1587]: eth0: Gained IPv6LL Aug 13 00:01:35.324275 systemd[1]: Finished systemd-networkd-wait-online.service. Aug 13 00:01:35.323000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd-wait-online comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.328983 systemd[1]: Finished systemd-tmpfiles-setup.service. Aug 13 00:01:35.330000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.345055 systemd[1]: Starting audit-rules.service... Aug 13 00:01:35.350552 systemd[1]: Starting clean-ca-certificates.service... Aug 13 00:01:35.361854 systemd[1]: Starting systemd-journal-catalog-update.service... Aug 13 00:01:35.367803 systemd[1]: Starting systemd-resolved.service... Aug 13 00:01:35.375075 systemd[1]: Starting systemd-timesyncd.service... Aug 13 00:01:35.385761 systemd[1]: Starting systemd-update-utmp.service... Aug 13 00:01:35.389775 systemd[1]: Finished clean-ca-certificates.service. Aug 13 00:01:35.397000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.418266 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Aug 13 00:01:35.424270 systemd[1]: Starting modprobe@dm_mod.service... Aug 13 00:01:35.432040 systemd[1]: Starting modprobe@efi_pstore.service... Aug 13 00:01:35.441944 systemd[1]: Starting modprobe@loop.service... Aug 13 00:01:35.444045 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Aug 13 00:01:35.444995 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Aug 13 00:01:35.445913 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 00:01:35.457043 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Aug 13 00:01:35.458066 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Aug 13 00:01:35.458452 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Aug 13 00:01:35.459154 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 00:01:35.469792 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Aug 13 00:01:35.476341 systemd[1]: Starting modprobe@drm.service... Aug 13 00:01:35.478275 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Aug 13 00:01:35.478582 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Aug 13 00:01:35.478927 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 00:01:35.489217 systemd[1]: Finished ensure-sysext.service. Aug 13 00:01:35.491821 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 00:01:35.492233 systemd[1]: Finished modprobe@loop.service. Aug 13 00:01:35.490000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ensure-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.491000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.491000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.511370 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 00:01:35.511761 systemd[1]: Finished modprobe@dm_mod.service. Aug 13 00:01:35.514046 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Aug 13 00:01:35.512000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.512000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.519000 audit[1847]: SYSTEM_BOOT pid=1847 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.521000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.521000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.519834 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 00:01:35.520448 systemd[1]: Finished modprobe@efi_pstore.service. Aug 13 00:01:35.522990 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 00:01:35.527000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.526448 systemd[1]: Finished systemd-update-utmp.service. Aug 13 00:01:35.548000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.548000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.555000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:35.549376 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 00:01:35.549761 systemd[1]: Finished modprobe@drm.service. Aug 13 00:01:35.554582 systemd[1]: Finished systemd-journal-catalog-update.service. Aug 13 00:01:35.612000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Aug 13 00:01:35.612000 audit[1871]: SYSCALL arch=c00000b7 syscall=206 success=yes exit=1056 a0=3 a1=ffffd73a2950 a2=420 a3=0 items=0 ppid=1838 pid=1871 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:35.612000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Aug 13 00:01:35.615955 augenrules[1871]: No rules Aug 13 00:01:35.617134 systemd[1]: Finished audit-rules.service. Aug 13 00:01:35.697184 ldconfig[1704]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Aug 13 00:01:35.698133 systemd[1]: Started systemd-timesyncd.service. Aug 13 00:01:35.700271 systemd[1]: Reached target time-set.target. Aug 13 00:01:35.709243 systemd[1]: Finished ldconfig.service. Aug 13 00:01:35.713802 systemd[1]: Starting systemd-update-done.service... Aug 13 00:01:35.740550 systemd[1]: Finished systemd-update-done.service. Aug 13 00:01:35.741018 systemd-resolved[1842]: Positive Trust Anchors: Aug 13 00:01:35.741036 systemd-resolved[1842]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 00:01:35.741088 systemd-resolved[1842]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Aug 13 00:01:35.795609 systemd-timesyncd[1843]: Contacted time server 141.11.228.173:123 (0.flatcar.pool.ntp.org). Aug 13 00:01:35.795949 systemd-timesyncd[1843]: Initial clock synchronization to Wed 2025-08-13 00:01:35.778181 UTC. Aug 13 00:01:35.800040 systemd-resolved[1842]: Defaulting to hostname 'linux'. Aug 13 00:01:35.803727 systemd[1]: Started systemd-resolved.service. Aug 13 00:01:35.806242 systemd[1]: Reached target network.target. Aug 13 00:01:35.808274 systemd[1]: Reached target network-online.target. Aug 13 00:01:35.810523 systemd[1]: Reached target nss-lookup.target. Aug 13 00:01:35.812554 systemd[1]: Reached target sysinit.target. Aug 13 00:01:35.814726 systemd[1]: Started motdgen.path. Aug 13 00:01:35.816621 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Aug 13 00:01:35.819644 systemd[1]: Started logrotate.timer. Aug 13 00:01:35.821716 systemd[1]: Started mdadm.timer. Aug 13 00:01:35.823441 systemd[1]: Started systemd-tmpfiles-clean.timer. Aug 13 00:01:35.825518 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Aug 13 00:01:35.825582 systemd[1]: Reached target paths.target. Aug 13 00:01:35.827404 systemd[1]: Reached target timers.target. Aug 13 00:01:35.835702 systemd[1]: Listening on dbus.socket. Aug 13 00:01:35.840214 systemd[1]: Starting docker.socket... Aug 13 00:01:35.845329 systemd[1]: Listening on sshd.socket. Aug 13 00:01:35.847685 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Aug 13 00:01:35.848442 systemd[1]: Listening on docker.socket. Aug 13 00:01:35.850923 systemd[1]: Reached target sockets.target. Aug 13 00:01:35.853232 systemd[1]: Reached target basic.target. Aug 13 00:01:35.855803 systemd[1]: System is tainted: cgroupsv1 Aug 13 00:01:35.855912 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Aug 13 00:01:35.855970 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Aug 13 00:01:35.858892 systemd[1]: Started amazon-ssm-agent.service. Aug 13 00:01:35.864958 systemd[1]: Starting containerd.service... Aug 13 00:01:35.872326 systemd[1]: Starting coreos-metadata-sshkeys@core.service... Aug 13 00:01:35.883057 systemd[1]: Starting dbus.service... Aug 13 00:01:35.898943 systemd[1]: Starting enable-oem-cloudinit.service... Aug 13 00:01:35.910696 systemd[1]: Starting extend-filesystems.service... Aug 13 00:01:35.934170 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Aug 13 00:01:35.940812 systemd[1]: Starting kubelet.service... Aug 13 00:01:35.945169 systemd[1]: Starting motdgen.service... Aug 13 00:01:35.950845 systemd[1]: Started nvidia.service. Aug 13 00:01:35.960261 systemd[1]: Starting prepare-helm.service... Aug 13 00:01:35.964785 systemd[1]: Starting ssh-key-proc-cmdline.service... Aug 13 00:01:35.976551 systemd[1]: Starting sshd-keygen.service... Aug 13 00:01:35.986485 systemd[1]: Starting systemd-logind.service... Aug 13 00:01:35.991090 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Aug 13 00:01:35.991332 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Aug 13 00:01:35.996375 systemd[1]: Starting update-engine.service... Aug 13 00:01:36.000877 systemd[1]: Starting update-ssh-keys-after-ignition.service... Aug 13 00:01:36.018748 jq[1904]: true Aug 13 00:01:36.084032 jq[1888]: false Aug 13 00:01:36.098846 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Aug 13 00:01:36.141181 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Aug 13 00:01:36.142207 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Aug 13 00:01:36.143008 systemd[1]: Finished ssh-key-proc-cmdline.service. Aug 13 00:01:36.154141 jq[1911]: true Aug 13 00:01:36.190200 tar[1908]: linux-arm64/helm Aug 13 00:01:36.202201 dbus-daemon[1887]: [system] SELinux support is enabled Aug 13 00:01:36.211698 systemd[1]: Started dbus.service. Aug 13 00:01:36.217078 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Aug 13 00:01:36.217165 systemd[1]: Reached target system-config.target. Aug 13 00:01:36.219284 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Aug 13 00:01:36.219336 systemd[1]: Reached target user-config.target. Aug 13 00:01:36.224679 extend-filesystems[1890]: Found loop1 Aug 13 00:01:36.224679 extend-filesystems[1890]: Found nvme0n1 Aug 13 00:01:36.224679 extend-filesystems[1890]: Found nvme0n1p1 Aug 13 00:01:36.224679 extend-filesystems[1890]: Found nvme0n1p2 Aug 13 00:01:36.224679 extend-filesystems[1890]: Found nvme0n1p3 Aug 13 00:01:36.224679 extend-filesystems[1890]: Found usr Aug 13 00:01:36.224679 extend-filesystems[1890]: Found nvme0n1p4 Aug 13 00:01:36.224679 extend-filesystems[1890]: Found nvme0n1p6 Aug 13 00:01:36.224679 extend-filesystems[1890]: Found nvme0n1p7 Aug 13 00:01:36.224679 extend-filesystems[1890]: Found nvme0n1p9 Aug 13 00:01:36.224679 extend-filesystems[1890]: Checking size of /dev/nvme0n1p9 Aug 13 00:01:36.260058 systemd[1]: motdgen.service: Deactivated successfully. Aug 13 00:01:36.259903 dbus-daemon[1887]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1587 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Aug 13 00:01:36.260711 systemd[1]: Finished motdgen.service. Aug 13 00:01:36.263372 dbus-daemon[1887]: [system] Successfully activated service 'org.freedesktop.systemd1' Aug 13 00:01:36.269918 systemd[1]: Starting systemd-hostnamed.service... Aug 13 00:01:36.320156 extend-filesystems[1890]: Resized partition /dev/nvme0n1p9 Aug 13 00:01:36.348664 extend-filesystems[1953]: resize2fs 1.46.5 (30-Dec-2021) Aug 13 00:01:36.397204 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks Aug 13 00:01:36.542308 update_engine[1902]: I0813 00:01:36.539832 1902 main.cc:92] Flatcar Update Engine starting Aug 13 00:01:36.569241 systemd[1]: Started update-engine.service. Aug 13 00:01:36.577898 systemd[1]: Started locksmithd.service. Aug 13 00:01:36.581087 update_engine[1902]: I0813 00:01:36.581004 1902 update_check_scheduler.cc:74] Next update check in 9m57s Aug 13 00:01:36.584141 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 Aug 13 00:01:36.600490 extend-filesystems[1953]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Aug 13 00:01:36.600490 extend-filesystems[1953]: old_desc_blocks = 1, new_desc_blocks = 1 Aug 13 00:01:36.600490 extend-filesystems[1953]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. Aug 13 00:01:36.617632 extend-filesystems[1890]: Resized filesystem in /dev/nvme0n1p9 Aug 13 00:01:36.627668 amazon-ssm-agent[1883]: 2025/08/13 00:01:36 Failed to load instance info from vault. RegistrationKey does not exist. Aug 13 00:01:36.627147 systemd[1]: extend-filesystems.service: Deactivated successfully. Aug 13 00:01:36.627723 systemd[1]: Finished extend-filesystems.service. Aug 13 00:01:36.636120 bash[1957]: Updated "/home/core/.ssh/authorized_keys" Aug 13 00:01:36.640649 env[1913]: time="2025-08-13T00:01:36.638593205Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Aug 13 00:01:36.640937 systemd[1]: Finished update-ssh-keys-after-ignition.service. Aug 13 00:01:36.696698 amazon-ssm-agent[1883]: Initializing new seelog logger Aug 13 00:01:36.697336 amazon-ssm-agent[1883]: New Seelog Logger Creation Complete Aug 13 00:01:36.698923 amazon-ssm-agent[1883]: 2025/08/13 00:01:36 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Aug 13 00:01:36.699088 amazon-ssm-agent[1883]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Aug 13 00:01:36.699832 amazon-ssm-agent[1883]: 2025/08/13 00:01:36 processing appconfig overrides Aug 13 00:01:36.706467 systemd[1]: nvidia.service: Deactivated successfully. Aug 13 00:01:36.770284 systemd-logind[1901]: Watching system buttons on /dev/input/event0 (Power Button) Aug 13 00:01:36.770889 systemd-logind[1901]: Watching system buttons on /dev/input/event1 (Sleep Button) Aug 13 00:01:36.779088 systemd-logind[1901]: New seat seat0. Aug 13 00:01:36.785013 systemd[1]: Started systemd-logind.service. Aug 13 00:01:36.883540 env[1913]: time="2025-08-13T00:01:36.883393297Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Aug 13 00:01:36.883670 env[1913]: time="2025-08-13T00:01:36.883648083Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Aug 13 00:01:36.895563 env[1913]: time="2025-08-13T00:01:36.895480322Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.189-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Aug 13 00:01:36.895563 env[1913]: time="2025-08-13T00:01:36.895553575Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Aug 13 00:01:36.896093 env[1913]: time="2025-08-13T00:01:36.896028623Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 00:01:36.896093 env[1913]: time="2025-08-13T00:01:36.896085742Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Aug 13 00:01:36.900993 env[1913]: time="2025-08-13T00:01:36.900921093Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Aug 13 00:01:36.901166 env[1913]: time="2025-08-13T00:01:36.900994142Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Aug 13 00:01:36.901312 env[1913]: time="2025-08-13T00:01:36.901264331Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Aug 13 00:01:36.901986 env[1913]: time="2025-08-13T00:01:36.901930526Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Aug 13 00:01:36.902404 env[1913]: time="2025-08-13T00:01:36.902330379Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 00:01:36.902404 env[1913]: time="2025-08-13T00:01:36.902386670Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Aug 13 00:01:36.902575 env[1913]: time="2025-08-13T00:01:36.902528094Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Aug 13 00:01:36.902697 env[1913]: time="2025-08-13T00:01:36.902565865Z" level=info msg="metadata content store policy set" policy=shared Aug 13 00:01:36.920933 dbus-daemon[1887]: [system] Successfully activated service 'org.freedesktop.hostname1' Aug 13 00:01:36.921199 systemd[1]: Started systemd-hostnamed.service. Aug 13 00:01:36.923501 dbus-daemon[1887]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.6' (uid=0 pid=1939 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Aug 13 00:01:36.931847 systemd[1]: Starting polkit.service... Aug 13 00:01:36.944241 env[1913]: time="2025-08-13T00:01:36.940783662Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Aug 13 00:01:36.944241 env[1913]: time="2025-08-13T00:01:36.940865498Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Aug 13 00:01:36.944241 env[1913]: time="2025-08-13T00:01:36.940901483Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Aug 13 00:01:36.944241 env[1913]: time="2025-08-13T00:01:36.940971056Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Aug 13 00:01:36.944241 env[1913]: time="2025-08-13T00:01:36.941009163Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Aug 13 00:01:36.944241 env[1913]: time="2025-08-13T00:01:36.941041924Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Aug 13 00:01:36.944241 env[1913]: time="2025-08-13T00:01:36.941097496Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Aug 13 00:01:36.944241 env[1913]: time="2025-08-13T00:01:36.941679613Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Aug 13 00:01:36.944241 env[1913]: time="2025-08-13T00:01:36.941726878Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Aug 13 00:01:36.944241 env[1913]: time="2025-08-13T00:01:36.941761017Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Aug 13 00:01:36.944241 env[1913]: time="2025-08-13T00:01:36.941815079Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Aug 13 00:01:36.944241 env[1913]: time="2025-08-13T00:01:36.941849302Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Aug 13 00:01:36.944241 env[1913]: time="2025-08-13T00:01:36.942075547Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Aug 13 00:01:36.944241 env[1913]: time="2025-08-13T00:01:36.942273262Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.942829834Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.942880564Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.942916525Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.943156986Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.943192731Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.943224054Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.943253985Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.943285607Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.943318464Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.943348707Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.943377476Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.943411675Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.943738887Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.943783611Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Aug 13 00:01:36.945028 env[1913]: time="2025-08-13T00:01:36.943816360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Aug 13 00:01:36.945793 env[1913]: time="2025-08-13T00:01:36.943850631Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Aug 13 00:01:36.945793 env[1913]: time="2025-08-13T00:01:36.943888618Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Aug 13 00:01:36.945793 env[1913]: time="2025-08-13T00:01:36.943916919Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Aug 13 00:01:36.945793 env[1913]: time="2025-08-13T00:01:36.943951574Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Aug 13 00:01:36.945793 env[1913]: time="2025-08-13T00:01:36.944030832Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Aug 13 00:01:36.946077 env[1913]: time="2025-08-13T00:01:36.944399076Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Aug 13 00:01:36.946077 env[1913]: time="2025-08-13T00:01:36.944506540Z" level=info msg="Connect containerd service" Aug 13 00:01:36.946077 env[1913]: time="2025-08-13T00:01:36.944576545Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Aug 13 00:01:36.961602 env[1913]: time="2025-08-13T00:01:36.961531369Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 00:01:36.964132 env[1913]: time="2025-08-13T00:01:36.961966213Z" level=info msg="Start subscribing containerd event" Aug 13 00:01:36.964132 env[1913]: time="2025-08-13T00:01:36.962062337Z" level=info msg="Start recovering state" Aug 13 00:01:36.964132 env[1913]: time="2025-08-13T00:01:36.962148692Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Aug 13 00:01:36.964132 env[1913]: time="2025-08-13T00:01:36.962266825Z" level=info msg="Start event monitor" Aug 13 00:01:36.964132 env[1913]: time="2025-08-13T00:01:36.962271572Z" level=info msg=serving... address=/run/containerd/containerd.sock Aug 13 00:01:36.964132 env[1913]: time="2025-08-13T00:01:36.962308456Z" level=info msg="Start snapshots syncer" Aug 13 00:01:36.964132 env[1913]: time="2025-08-13T00:01:36.962363237Z" level=info msg="Start cni network conf syncer for default" Aug 13 00:01:36.964132 env[1913]: time="2025-08-13T00:01:36.962387224Z" level=info msg="Start streaming server" Aug 13 00:01:36.962522 systemd[1]: Started containerd.service. Aug 13 00:01:36.983868 env[1913]: time="2025-08-13T00:01:36.983790481Z" level=info msg="containerd successfully booted in 0.384433s" Aug 13 00:01:36.987129 polkitd[2004]: Started polkitd version 121 Aug 13 00:01:37.033056 polkitd[2004]: Loading rules from directory /etc/polkit-1/rules.d Aug 13 00:01:37.033794 polkitd[2004]: Loading rules from directory /usr/share/polkit-1/rules.d Aug 13 00:01:37.038295 polkitd[2004]: Finished loading, compiling and executing 2 rules Aug 13 00:01:37.040446 dbus-daemon[1887]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Aug 13 00:01:37.040985 systemd[1]: Started polkit.service. Aug 13 00:01:37.044043 polkitd[2004]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Aug 13 00:01:37.084580 systemd-hostnamed[1939]: Hostname set to (transient) Aug 13 00:01:37.084743 systemd-resolved[1842]: System hostname changed to 'ip-172-31-27-151'. Aug 13 00:01:37.135892 coreos-metadata[1885]: Aug 13 00:01:37.135 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Aug 13 00:01:37.137853 coreos-metadata[1885]: Aug 13 00:01:37.137 INFO Fetching http://169.254.169.254/2019-10-01/meta-data/public-keys: Attempt #1 Aug 13 00:01:37.138685 coreos-metadata[1885]: Aug 13 00:01:37.138 INFO Fetch successful Aug 13 00:01:37.139970 coreos-metadata[1885]: Aug 13 00:01:37.139 INFO Fetching http://169.254.169.254/2019-10-01/meta-data/public-keys/0/openssh-key: Attempt #1 Aug 13 00:01:37.140293 coreos-metadata[1885]: Aug 13 00:01:37.140 INFO Fetch successful Aug 13 00:01:37.145074 unknown[1885]: wrote ssh authorized keys file for user: core Aug 13 00:01:37.169209 update-ssh-keys[2023]: Updated "/home/core/.ssh/authorized_keys" Aug 13 00:01:37.170049 systemd[1]: Finished coreos-metadata-sshkeys@core.service. Aug 13 00:01:37.584913 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO Create new startup processor Aug 13 00:01:37.585281 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [LongRunningPluginsManager] registered plugins: {} Aug 13 00:01:37.585420 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO Initializing bookkeeping folders Aug 13 00:01:37.585420 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO removing the completed state files Aug 13 00:01:37.585420 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO Initializing bookkeeping folders for long running plugins Aug 13 00:01:37.585420 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO Initializing replies folder for MDS reply requests that couldn't reach the service Aug 13 00:01:37.585420 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO Initializing healthcheck folders for long running plugins Aug 13 00:01:37.585420 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO Initializing locations for inventory plugin Aug 13 00:01:37.585729 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO Initializing default location for custom inventory Aug 13 00:01:37.585729 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO Initializing default location for file inventory Aug 13 00:01:37.585729 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO Initializing default location for role inventory Aug 13 00:01:37.585729 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO Init the cloudwatchlogs publisher Aug 13 00:01:37.585729 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [instanceID=i-0c2478391bcf4e049] Successfully loaded platform independent plugin aws:runPowerShellScript Aug 13 00:01:37.585729 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [instanceID=i-0c2478391bcf4e049] Successfully loaded platform independent plugin aws:runDockerAction Aug 13 00:01:37.585729 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [instanceID=i-0c2478391bcf4e049] Successfully loaded platform independent plugin aws:refreshAssociation Aug 13 00:01:37.585729 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [instanceID=i-0c2478391bcf4e049] Successfully loaded platform independent plugin aws:configurePackage Aug 13 00:01:37.585729 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [instanceID=i-0c2478391bcf4e049] Successfully loaded platform independent plugin aws:runDocument Aug 13 00:01:37.585729 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [instanceID=i-0c2478391bcf4e049] Successfully loaded platform independent plugin aws:softwareInventory Aug 13 00:01:37.585729 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [instanceID=i-0c2478391bcf4e049] Successfully loaded platform independent plugin aws:updateSsmAgent Aug 13 00:01:37.585729 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [instanceID=i-0c2478391bcf4e049] Successfully loaded platform independent plugin aws:configureDocker Aug 13 00:01:37.586421 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [instanceID=i-0c2478391bcf4e049] Successfully loaded platform independent plugin aws:downloadContent Aug 13 00:01:37.586421 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [instanceID=i-0c2478391bcf4e049] Successfully loaded platform dependent plugin aws:runShellScript Aug 13 00:01:37.586421 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO Starting Agent: amazon-ssm-agent - v2.3.1319.0 Aug 13 00:01:37.586421 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO OS: linux, Arch: arm64 Aug 13 00:01:37.587393 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessagingDeliveryService] Starting document processing engine... Aug 13 00:01:37.599033 amazon-ssm-agent[1883]: datastore file /var/lib/amazon/ssm/i-0c2478391bcf4e049/longrunningplugins/datastore/store doesn't exist - no long running plugins to execute Aug 13 00:01:37.687279 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessagingDeliveryService] [EngineProcessor] Starting Aug 13 00:01:37.785050 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessagingDeliveryService] [EngineProcessor] Initial processing Aug 13 00:01:37.880281 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessagingDeliveryService] Starting message polling Aug 13 00:01:37.974989 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessagingDeliveryService] Starting send replies to MDS Aug 13 00:01:38.069956 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [instanceID=i-0c2478391bcf4e049] Starting association polling Aug 13 00:01:38.165177 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessagingDeliveryService] [Association] [EngineProcessor] Starting Aug 13 00:01:38.260426 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessagingDeliveryService] [Association] Launching response handler Aug 13 00:01:38.293739 tar[1908]: linux-arm64/LICENSE Aug 13 00:01:38.294681 tar[1908]: linux-arm64/README.md Aug 13 00:01:38.309182 systemd[1]: Finished prepare-helm.service. Aug 13 00:01:38.356083 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessagingDeliveryService] [Association] [EngineProcessor] Initial processing Aug 13 00:01:38.451944 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessagingDeliveryService] [Association] Initializing association scheduling service Aug 13 00:01:38.495250 locksmithd[1973]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Aug 13 00:01:38.547751 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessagingDeliveryService] [Association] Association scheduling service initialized Aug 13 00:01:38.643821 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [HealthCheck] HealthCheck reporting agent health. Aug 13 00:01:38.740248 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessageGatewayService] Starting session document processing engine... Aug 13 00:01:38.836653 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessageGatewayService] [EngineProcessor] Starting Aug 13 00:01:38.874212 systemd[1]: Started kubelet.service. Aug 13 00:01:38.933399 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessageGatewayService] SSM Agent is trying to setup control channel for Session Manager module. Aug 13 00:01:39.030425 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessageGatewayService] Setting up websocket for controlchannel for instance: i-0c2478391bcf4e049, requestId: a91917c3-08fa-4e93-8568-a64401577375 Aug 13 00:01:39.127546 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [OfflineService] Starting document processing engine... Aug 13 00:01:39.224842 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [OfflineService] [EngineProcessor] Starting Aug 13 00:01:39.322454 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [OfflineService] [EngineProcessor] Initial processing Aug 13 00:01:39.420043 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [OfflineService] Starting message polling Aug 13 00:01:39.518056 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [OfflineService] Starting send replies to MDS Aug 13 00:01:39.616181 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [LongRunningPluginsManager] starting long running plugin manager Aug 13 00:01:39.714512 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [LongRunningPluginsManager] there aren't any long running plugin to execute Aug 13 00:01:39.807721 sshd_keygen[1931]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Aug 13 00:01:39.812952 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessageGatewayService] listening reply. Aug 13 00:01:39.862489 systemd[1]: Finished sshd-keygen.service. Aug 13 00:01:39.867787 systemd[1]: Starting issuegen.service... Aug 13 00:01:39.884442 systemd[1]: issuegen.service: Deactivated successfully. Aug 13 00:01:39.885060 systemd[1]: Finished issuegen.service. Aug 13 00:01:39.890636 systemd[1]: Starting systemd-user-sessions.service... Aug 13 00:01:39.911648 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [LongRunningPluginsManager] There are no long running plugins currently getting executed - skipping their healthcheck Aug 13 00:01:39.912959 kubelet[2117]: E0813 00:01:39.912900 2117 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 00:01:39.917545 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 00:01:39.917949 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 00:01:39.920146 systemd[1]: Finished systemd-user-sessions.service. Aug 13 00:01:39.925802 systemd[1]: Started getty@tty1.service. Aug 13 00:01:39.934368 systemd[1]: Started serial-getty@ttyS0.service. Aug 13 00:01:39.936901 systemd[1]: Reached target getty.target. Aug 13 00:01:39.938916 systemd[1]: Reached target multi-user.target. Aug 13 00:01:39.944643 systemd[1]: Starting systemd-update-utmp-runlevel.service... Aug 13 00:01:39.961401 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Aug 13 00:01:39.961968 systemd[1]: Finished systemd-update-utmp-runlevel.service. Aug 13 00:01:39.970340 systemd[1]: Startup finished in 9.959s (kernel) + 12.600s (userspace) = 22.559s. Aug 13 00:01:40.010631 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [StartupProcessor] Executing startup processor tasks Aug 13 00:01:40.109730 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [StartupProcessor] Write to serial port: Amazon SSM Agent v2.3.1319.0 is running Aug 13 00:01:40.209094 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [StartupProcessor] Write to serial port: OsProductName: Flatcar Container Linux by Kinvolk Aug 13 00:01:40.308518 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [StartupProcessor] Write to serial port: OsVersion: 3510.3.8 Aug 13 00:01:40.408149 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessageGatewayService] Opening websocket connection to: wss://ssmmessages.us-west-2.amazonaws.com/v1/control-channel/i-0c2478391bcf4e049?role=subscribe&stream=input Aug 13 00:01:40.508210 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessageGatewayService] Successfully opened websocket connection to: wss://ssmmessages.us-west-2.amazonaws.com/v1/control-channel/i-0c2478391bcf4e049?role=subscribe&stream=input Aug 13 00:01:40.608127 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessageGatewayService] Starting receiving message from control channel Aug 13 00:01:40.708494 amazon-ssm-agent[1883]: 2025-08-13 00:01:37 INFO [MessageGatewayService] [EngineProcessor] Initial processing Aug 13 00:01:44.521044 systemd[1]: Created slice system-sshd.slice. Aug 13 00:01:44.523364 systemd[1]: Started sshd@0-172.31.27.151:22-139.178.89.65:42802.service. Aug 13 00:01:44.878161 sshd[2143]: Accepted publickey for core from 139.178.89.65 port 42802 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:01:44.884175 sshd[2143]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:01:44.904936 systemd[1]: Created slice user-500.slice. Aug 13 00:01:44.907450 systemd[1]: Starting user-runtime-dir@500.service... Aug 13 00:01:44.918199 systemd-logind[1901]: New session 1 of user core. Aug 13 00:01:44.929662 systemd[1]: Finished user-runtime-dir@500.service. Aug 13 00:01:44.934773 systemd[1]: Starting user@500.service... Aug 13 00:01:44.946830 (systemd)[2148]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:01:45.130186 systemd[2148]: Queued start job for default target default.target. Aug 13 00:01:45.131390 systemd[2148]: Reached target paths.target. Aug 13 00:01:45.131451 systemd[2148]: Reached target sockets.target. Aug 13 00:01:45.131483 systemd[2148]: Reached target timers.target. Aug 13 00:01:45.131513 systemd[2148]: Reached target basic.target. Aug 13 00:01:45.131721 systemd[1]: Started user@500.service. Aug 13 00:01:45.132668 systemd[2148]: Reached target default.target. Aug 13 00:01:45.132936 systemd[2148]: Startup finished in 173ms. Aug 13 00:01:45.133612 systemd[1]: Started session-1.scope. Aug 13 00:01:45.281818 systemd[1]: Started sshd@1-172.31.27.151:22-139.178.89.65:42812.service. Aug 13 00:01:45.464441 sshd[2157]: Accepted publickey for core from 139.178.89.65 port 42812 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:01:45.467644 sshd[2157]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:01:45.475183 systemd-logind[1901]: New session 2 of user core. Aug 13 00:01:45.477255 systemd[1]: Started session-2.scope. Aug 13 00:01:45.611965 sshd[2157]: pam_unix(sshd:session): session closed for user core Aug 13 00:01:45.616824 systemd[1]: sshd@1-172.31.27.151:22-139.178.89.65:42812.service: Deactivated successfully. Aug 13 00:01:45.618704 systemd-logind[1901]: Session 2 logged out. Waiting for processes to exit. Aug 13 00:01:45.618884 systemd[1]: session-2.scope: Deactivated successfully. Aug 13 00:01:45.621707 systemd-logind[1901]: Removed session 2. Aug 13 00:01:45.637072 systemd[1]: Started sshd@2-172.31.27.151:22-139.178.89.65:42824.service. Aug 13 00:01:45.814654 sshd[2164]: Accepted publickey for core from 139.178.89.65 port 42824 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:01:45.816562 sshd[2164]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:01:45.825227 systemd-logind[1901]: New session 3 of user core. Aug 13 00:01:45.825975 systemd[1]: Started session-3.scope. Aug 13 00:01:45.949617 sshd[2164]: pam_unix(sshd:session): session closed for user core Aug 13 00:01:45.954584 systemd[1]: sshd@2-172.31.27.151:22-139.178.89.65:42824.service: Deactivated successfully. Aug 13 00:01:45.955958 systemd[1]: session-3.scope: Deactivated successfully. Aug 13 00:01:45.958739 systemd-logind[1901]: Session 3 logged out. Waiting for processes to exit. Aug 13 00:01:45.961355 systemd-logind[1901]: Removed session 3. Aug 13 00:01:45.975013 systemd[1]: Started sshd@3-172.31.27.151:22-139.178.89.65:42828.service. Aug 13 00:01:46.151898 sshd[2171]: Accepted publickey for core from 139.178.89.65 port 42828 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:01:46.155070 sshd[2171]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:01:46.162811 systemd-logind[1901]: New session 4 of user core. Aug 13 00:01:46.163762 systemd[1]: Started session-4.scope. Aug 13 00:01:46.300450 sshd[2171]: pam_unix(sshd:session): session closed for user core Aug 13 00:01:46.305786 systemd-logind[1901]: Session 4 logged out. Waiting for processes to exit. Aug 13 00:01:46.307725 systemd[1]: sshd@3-172.31.27.151:22-139.178.89.65:42828.service: Deactivated successfully. Aug 13 00:01:46.309332 systemd[1]: session-4.scope: Deactivated successfully. Aug 13 00:01:46.311836 systemd-logind[1901]: Removed session 4. Aug 13 00:01:46.326837 systemd[1]: Started sshd@4-172.31.27.151:22-139.178.89.65:42832.service. Aug 13 00:01:46.504734 sshd[2178]: Accepted publickey for core from 139.178.89.65 port 42832 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:01:46.507992 sshd[2178]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:01:46.517192 systemd[1]: Started session-5.scope. Aug 13 00:01:46.518169 systemd-logind[1901]: New session 5 of user core. Aug 13 00:01:46.673419 sudo[2182]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Aug 13 00:01:46.674003 sudo[2182]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Aug 13 00:01:46.690423 dbus-daemon[1887]: avc: received setenforce notice (enforcing=1) Aug 13 00:01:46.693281 sudo[2182]: pam_unix(sudo:session): session closed for user root Aug 13 00:01:46.718680 sshd[2178]: pam_unix(sshd:session): session closed for user core Aug 13 00:01:46.724633 systemd-logind[1901]: Session 5 logged out. Waiting for processes to exit. Aug 13 00:01:46.725437 systemd[1]: sshd@4-172.31.27.151:22-139.178.89.65:42832.service: Deactivated successfully. Aug 13 00:01:46.727005 systemd[1]: session-5.scope: Deactivated successfully. Aug 13 00:01:46.728316 systemd-logind[1901]: Removed session 5. Aug 13 00:01:46.745692 systemd[1]: Started sshd@5-172.31.27.151:22-139.178.89.65:42838.service. Aug 13 00:01:46.932707 sshd[2186]: Accepted publickey for core from 139.178.89.65 port 42838 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:01:46.934813 sshd[2186]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:01:46.944413 systemd[1]: Started session-6.scope. Aug 13 00:01:46.944880 systemd-logind[1901]: New session 6 of user core. Aug 13 00:01:47.058575 sudo[2191]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Aug 13 00:01:47.059202 sudo[2191]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Aug 13 00:01:47.064719 sudo[2191]: pam_unix(sudo:session): session closed for user root Aug 13 00:01:47.074586 sudo[2190]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Aug 13 00:01:47.075722 sudo[2190]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Aug 13 00:01:47.096209 systemd[1]: Stopping audit-rules.service... Aug 13 00:01:47.097000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Aug 13 00:01:47.098422 auditctl[2194]: No rules Aug 13 00:01:47.099955 kernel: kauditd_printk_skb: 55 callbacks suppressed Aug 13 00:01:47.100043 kernel: audit: type=1305 audit(1755043307.097:146): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 Aug 13 00:01:47.100691 systemd[1]: audit-rules.service: Deactivated successfully. Aug 13 00:01:47.101247 systemd[1]: Stopped audit-rules.service. Aug 13 00:01:47.105698 systemd[1]: Starting audit-rules.service... Aug 13 00:01:47.097000 audit[2194]: SYSCALL arch=c00000b7 syscall=206 success=yes exit=1056 a0=3 a1=fffff15de370 a2=420 a3=0 items=0 ppid=1 pid=2194 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:47.117926 kernel: audit: type=1300 audit(1755043307.097:146): arch=c00000b7 syscall=206 success=yes exit=1056 a0=3 a1=fffff15de370 a2=420 a3=0 items=0 ppid=1 pid=2194 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:47.097000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D44 Aug 13 00:01:47.123146 kernel: audit: type=1327 audit(1755043307.097:146): proctitle=2F7362696E2F617564697463746C002D44 Aug 13 00:01:47.101000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:47.134940 kernel: audit: type=1131 audit(1755043307.101:147): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:47.155530 augenrules[2212]: No rules Aug 13 00:01:47.157000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:47.157322 systemd[1]: Finished audit-rules.service. Aug 13 00:01:47.166903 sudo[2190]: pam_unix(sudo:session): session closed for user root Aug 13 00:01:47.166000 audit[2190]: USER_END pid=2190 uid=500 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Aug 13 00:01:47.178302 kernel: audit: type=1130 audit(1755043307.157:148): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:47.178412 kernel: audit: type=1106 audit(1755043307.166:149): pid=2190 uid=500 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Aug 13 00:01:47.167000 audit[2190]: CRED_DISP pid=2190 uid=500 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Aug 13 00:01:47.187497 kernel: audit: type=1104 audit(1755043307.167:150): pid=2190 uid=500 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Aug 13 00:01:47.193205 sshd[2186]: pam_unix(sshd:session): session closed for user core Aug 13 00:01:47.194000 audit[2186]: USER_END pid=2186 uid=0 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:01:47.195000 audit[2186]: CRED_DISP pid=2186 uid=0 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:01:47.217321 kernel: audit: type=1106 audit(1755043307.194:151): pid=2186 uid=0 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:01:47.217459 kernel: audit: type=1104 audit(1755043307.195:152): pid=2186 uid=0 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:01:47.208220 systemd[1]: sshd@5-172.31.27.151:22-139.178.89.65:42838.service: Deactivated successfully. Aug 13 00:01:47.209461 systemd[1]: session-6.scope: Deactivated successfully. Aug 13 00:01:47.217368 systemd-logind[1901]: Session 6 logged out. Waiting for processes to exit. Aug 13 00:01:47.207000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@5-172.31.27.151:22-139.178.89.65:42838 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:47.225409 systemd[1]: Started sshd@6-172.31.27.151:22-139.178.89.65:42846.service. Aug 13 00:01:47.228620 kernel: audit: type=1131 audit(1755043307.207:153): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@5-172.31.27.151:22-139.178.89.65:42838 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:47.224000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-172.31.27.151:22-139.178.89.65:42846 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:47.230010 systemd-logind[1901]: Removed session 6. Aug 13 00:01:47.312234 amazon-ssm-agent[1883]: 2025-08-13 00:01:47 INFO [MessagingDeliveryService] [Association] No associations on boot. Requerying for associations after 30 seconds. Aug 13 00:01:47.408000 audit[2219]: USER_ACCT pid=2219 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:01:47.410276 sshd[2219]: Accepted publickey for core from 139.178.89.65 port 42846 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:01:47.410000 audit[2219]: CRED_ACQ pid=2219 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:01:47.411000 audit[2219]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffc7a303d0 a2=3 a3=1 items=0 ppid=1 pid=2219 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=7 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:47.411000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:01:47.413510 sshd[2219]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:01:47.422942 systemd[1]: Started session-7.scope. Aug 13 00:01:47.425533 systemd-logind[1901]: New session 7 of user core. Aug 13 00:01:47.436000 audit[2219]: USER_START pid=2219 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:01:47.438000 audit[2222]: CRED_ACQ pid=2222 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:01:47.538000 audit[2223]: USER_ACCT pid=2223 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Aug 13 00:01:47.539811 sudo[2223]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Aug 13 00:01:47.539000 audit[2223]: CRED_REFR pid=2223 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Aug 13 00:01:47.541074 sudo[2223]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Aug 13 00:01:47.544000 audit[2223]: USER_START pid=2223 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Aug 13 00:01:47.624809 systemd[1]: Starting docker.service... Aug 13 00:01:47.745916 env[2233]: time="2025-08-13T00:01:47.745834662Z" level=info msg="Starting up" Aug 13 00:01:47.753754 env[2233]: time="2025-08-13T00:01:47.753657737Z" level=info msg="parsed scheme: \"unix\"" module=grpc Aug 13 00:01:47.753754 env[2233]: time="2025-08-13T00:01:47.753715870Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Aug 13 00:01:47.753968 env[2233]: time="2025-08-13T00:01:47.753771749Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Aug 13 00:01:47.753968 env[2233]: time="2025-08-13T00:01:47.753800725Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Aug 13 00:01:47.759027 env[2233]: time="2025-08-13T00:01:47.758969438Z" level=info msg="parsed scheme: \"unix\"" module=grpc Aug 13 00:01:47.759551 env[2233]: time="2025-08-13T00:01:47.759501140Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Aug 13 00:01:47.759825 env[2233]: time="2025-08-13T00:01:47.759759735Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Aug 13 00:01:47.759973 env[2233]: time="2025-08-13T00:01:47.759940767Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Aug 13 00:01:47.986978 env[2233]: time="2025-08-13T00:01:47.986818278Z" level=warning msg="Your kernel does not support cgroup blkio weight" Aug 13 00:01:47.986978 env[2233]: time="2025-08-13T00:01:47.986869550Z" level=warning msg="Your kernel does not support cgroup blkio weight_device" Aug 13 00:01:47.987368 env[2233]: time="2025-08-13T00:01:47.987329627Z" level=info msg="Loading containers: start." Aug 13 00:01:48.151000 audit[2264]: NETFILTER_CFG table=nat:2 family=2 entries=2 op=nft_register_chain pid=2264 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.151000 audit[2264]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=116 a0=3 a1=fffff3f5b5a0 a2=0 a3=1 items=0 ppid=2233 pid=2264 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.151000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 Aug 13 00:01:48.156000 audit[2266]: NETFILTER_CFG table=filter:3 family=2 entries=2 op=nft_register_chain pid=2266 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.156000 audit[2266]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=124 a0=3 a1=ffffd4890ee0 a2=0 a3=1 items=0 ppid=2233 pid=2266 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.156000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 Aug 13 00:01:48.161000 audit[2268]: NETFILTER_CFG table=filter:4 family=2 entries=1 op=nft_register_chain pid=2268 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.161000 audit[2268]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=112 a0=3 a1=ffffdef99010 a2=0 a3=1 items=0 ppid=2233 pid=2268 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.161000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 Aug 13 00:01:48.166000 audit[2270]: NETFILTER_CFG table=filter:5 family=2 entries=1 op=nft_register_chain pid=2270 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.166000 audit[2270]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=112 a0=3 a1=ffffd83856d0 a2=0 a3=1 items=0 ppid=2233 pid=2270 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.166000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 Aug 13 00:01:48.188000 audit[2272]: NETFILTER_CFG table=filter:6 family=2 entries=1 op=nft_register_rule pid=2272 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.188000 audit[2272]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=228 a0=3 a1=ffffd05200c0 a2=0 a3=1 items=0 ppid=2233 pid=2272 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.188000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4100444F434B45522D49534F4C4154494F4E2D53544147452D31002D6A0052455455524E Aug 13 00:01:48.217000 audit[2277]: NETFILTER_CFG table=filter:7 family=2 entries=1 op=nft_register_rule pid=2277 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.217000 audit[2277]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=228 a0=3 a1=ffffef3f36e0 a2=0 a3=1 items=0 ppid=2233 pid=2277 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.217000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4100444F434B45522D49534F4C4154494F4E2D53544147452D32002D6A0052455455524E Aug 13 00:01:48.235000 audit[2279]: NETFILTER_CFG table=filter:8 family=2 entries=1 op=nft_register_chain pid=2279 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.235000 audit[2279]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=96 a0=3 a1=fffffbccd100 a2=0 a3=1 items=0 ppid=2233 pid=2279 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.235000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 Aug 13 00:01:48.241000 audit[2281]: NETFILTER_CFG table=filter:9 family=2 entries=1 op=nft_register_rule pid=2281 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.241000 audit[2281]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=212 a0=3 a1=ffffd77467e0 a2=0 a3=1 items=0 ppid=2233 pid=2281 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.241000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E Aug 13 00:01:48.245000 audit[2283]: NETFILTER_CFG table=filter:10 family=2 entries=2 op=nft_register_chain pid=2283 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.245000 audit[2283]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=308 a0=3 a1=ffffdddf4fa0 a2=0 a3=1 items=0 ppid=2233 pid=2283 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.245000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Aug 13 00:01:48.264000 audit[2287]: NETFILTER_CFG table=filter:11 family=2 entries=1 op=nft_unregister_rule pid=2287 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.264000 audit[2287]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=216 a0=3 a1=ffffc5cf5410 a2=0 a3=1 items=0 ppid=2233 pid=2287 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.264000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4400464F5257415244002D6A00444F434B45522D55534552 Aug 13 00:01:48.271000 audit[2288]: NETFILTER_CFG table=filter:12 family=2 entries=1 op=nft_register_rule pid=2288 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.271000 audit[2288]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=224 a0=3 a1=ffffcb66fa80 a2=0 a3=1 items=0 ppid=2233 pid=2288 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.271000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Aug 13 00:01:48.302162 kernel: Initializing XFRM netlink socket Aug 13 00:01:48.373190 env[2233]: time="2025-08-13T00:01:48.373068578Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" Aug 13 00:01:48.377678 (udev-worker)[2244]: Network interface NamePolicy= disabled on kernel command line. Aug 13 00:01:48.416000 audit[2296]: NETFILTER_CFG table=nat:13 family=2 entries=2 op=nft_register_chain pid=2296 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.416000 audit[2296]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=492 a0=3 a1=fffff4e6f750 a2=0 a3=1 items=0 ppid=2233 pid=2296 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.416000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D4900504F5354524F5554494E47002D73003137322E31372E302E302F31360000002D6F00646F636B657230002D6A004D415351554552414445 Aug 13 00:01:48.436000 audit[2299]: NETFILTER_CFG table=nat:14 family=2 entries=1 op=nft_register_rule pid=2299 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.436000 audit[2299]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=288 a0=3 a1=ffffcdec5d30 a2=0 a3=1 items=0 ppid=2233 pid=2299 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.436000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D4900444F434B4552002D6900646F636B657230002D6A0052455455524E Aug 13 00:01:48.443000 audit[2302]: NETFILTER_CFG table=filter:15 family=2 entries=1 op=nft_register_rule pid=2302 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.443000 audit[2302]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=376 a0=3 a1=ffffd623a750 a2=0 a3=1 items=0 ppid=2233 pid=2302 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.443000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6900646F636B657230002D6F00646F636B657230002D6A00414343455054 Aug 13 00:01:48.448000 audit[2304]: NETFILTER_CFG table=filter:16 family=2 entries=1 op=nft_register_rule pid=2304 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.448000 audit[2304]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=376 a0=3 a1=ffffd7ed6e00 a2=0 a3=1 items=0 ppid=2233 pid=2304 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.448000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6900646F636B6572300000002D6F00646F636B657230002D6A00414343455054 Aug 13 00:01:48.453000 audit[2306]: NETFILTER_CFG table=nat:17 family=2 entries=2 op=nft_register_chain pid=2306 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.453000 audit[2306]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=356 a0=3 a1=fffff33a05d0 a2=0 a3=1 items=0 ppid=2233 pid=2306 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.453000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 Aug 13 00:01:48.458000 audit[2308]: NETFILTER_CFG table=nat:18 family=2 entries=2 op=nft_register_chain pid=2308 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.458000 audit[2308]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=444 a0=3 a1=ffffdee58ad0 a2=0 a3=1 items=0 ppid=2233 pid=2308 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.458000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003132372E302E302E302F38 Aug 13 00:01:48.462000 audit[2310]: NETFILTER_CFG table=filter:19 family=2 entries=1 op=nft_register_rule pid=2310 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.462000 audit[2310]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=304 a0=3 a1=ffffeb215ae0 a2=0 a3=1 items=0 ppid=2233 pid=2310 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.462000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6F00646F636B657230002D6A00444F434B4552 Aug 13 00:01:48.491000 audit[2313]: NETFILTER_CFG table=filter:20 family=2 entries=1 op=nft_register_rule pid=2313 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.491000 audit[2313]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=508 a0=3 a1=ffffc0ff4170 a2=0 a3=1 items=0 ppid=2233 pid=2313 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.491000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6F00646F636B657230002D6D00636F6E6E747261636B002D2D637473746174650052454C415445442C45535441424C4953484544002D6A00414343455054 Aug 13 00:01:48.498000 audit[2315]: NETFILTER_CFG table=filter:21 family=2 entries=1 op=nft_register_rule pid=2315 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.498000 audit[2315]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=240 a0=3 a1=ffffea6a6e50 a2=0 a3=1 items=0 ppid=2233 pid=2315 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.498000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 Aug 13 00:01:48.503000 audit[2317]: NETFILTER_CFG table=filter:22 family=2 entries=1 op=nft_register_rule pid=2317 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.503000 audit[2317]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=428 a0=3 a1=ffffc9898b80 a2=0 a3=1 items=0 ppid=2233 pid=2317 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.503000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4900444F434B45522D49534F4C4154494F4E2D53544147452D31002D6900646F636B6572300000002D6F00646F636B657230002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D32 Aug 13 00:01:48.508000 audit[2319]: NETFILTER_CFG table=filter:23 family=2 entries=1 op=nft_register_rule pid=2319 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.508000 audit[2319]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=312 a0=3 a1=ffffe5d87a00 a2=0 a3=1 items=0 ppid=2233 pid=2319 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.508000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4900444F434B45522D49534F4C4154494F4E2D53544147452D32002D6F00646F636B657230002D6A0044524F50 Aug 13 00:01:48.510383 systemd-networkd[1587]: docker0: Link UP Aug 13 00:01:48.533000 audit[2323]: NETFILTER_CFG table=filter:24 family=2 entries=1 op=nft_unregister_rule pid=2323 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.533000 audit[2323]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=228 a0=3 a1=ffffed515940 a2=0 a3=1 items=0 ppid=2233 pid=2323 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.533000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4400464F5257415244002D6A00444F434B45522D55534552 Aug 13 00:01:48.541000 audit[2324]: NETFILTER_CFG table=filter:25 family=2 entries=1 op=nft_register_rule pid=2324 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:01:48.541000 audit[2324]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=224 a0=3 a1=ffffddaa0660 a2=0 a3=1 items=0 ppid=2233 pid=2324 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:01:48.541000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 Aug 13 00:01:48.543060 env[2233]: time="2025-08-13T00:01:48.542974045Z" level=info msg="Loading containers: done." Aug 13 00:01:48.578096 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck4182380240-merged.mount: Deactivated successfully. Aug 13 00:01:48.600161 env[2233]: time="2025-08-13T00:01:48.600042998Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Aug 13 00:01:48.600950 env[2233]: time="2025-08-13T00:01:48.600882821Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 Aug 13 00:01:48.601502 env[2233]: time="2025-08-13T00:01:48.601463357Z" level=info msg="Daemon has completed initialization" Aug 13 00:01:48.637000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=docker comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:48.638292 systemd[1]: Started docker.service. Aug 13 00:01:48.655940 env[2233]: time="2025-08-13T00:01:48.655834092Z" level=info msg="API listen on /run/docker.sock" Aug 13 00:01:49.845067 env[1913]: time="2025-08-13T00:01:49.844856444Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.11\"" Aug 13 00:01:50.168000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:50.169000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:50.169705 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Aug 13 00:01:50.170058 systemd[1]: Stopped kubelet.service. Aug 13 00:01:50.174004 systemd[1]: Starting kubelet.service... Aug 13 00:01:50.540876 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount696372883.mount: Deactivated successfully. Aug 13 00:01:50.572446 systemd[1]: Started kubelet.service. Aug 13 00:01:50.571000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:01:50.695555 kubelet[2365]: E0813 00:01:50.695461 2365 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 00:01:50.703782 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 00:01:50.704303 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 00:01:50.703000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Aug 13 00:01:52.780306 env[1913]: time="2025-08-13T00:01:52.780236739Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:52.784851 env[1913]: time="2025-08-13T00:01:52.784788659Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:00a68b619a4bfa14c989a2181a7aa0726a5cb1272a7f65394e6a594ad6eade27,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:52.791603 env[1913]: time="2025-08-13T00:01:52.791506576Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:52.794133 env[1913]: time="2025-08-13T00:01:52.794055395Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:a3d1c4440817725a1b503a7ccce94f3dce2b208ebf257b405dc2d97817df3dde,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:52.796239 env[1913]: time="2025-08-13T00:01:52.796158453Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.11\" returns image reference \"sha256:00a68b619a4bfa14c989a2181a7aa0726a5cb1272a7f65394e6a594ad6eade27\"" Aug 13 00:01:52.798612 env[1913]: time="2025-08-13T00:01:52.798565474Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.11\"" Aug 13 00:01:55.072640 env[1913]: time="2025-08-13T00:01:55.072575113Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:55.079061 env[1913]: time="2025-08-13T00:01:55.078959641Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:5c5dc52b837451e0fe6108fdfb9cfa431191ce227ce71d103dec8a8c655c4e71,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:55.082907 env[1913]: time="2025-08-13T00:01:55.082823597Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:55.087362 env[1913]: time="2025-08-13T00:01:55.087287315Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:0f19de157f3d251f5ddeb6e9d026895bc55cb02592874b326fa345c57e5e2848,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:55.089335 env[1913]: time="2025-08-13T00:01:55.089272772Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.11\" returns image reference \"sha256:5c5dc52b837451e0fe6108fdfb9cfa431191ce227ce71d103dec8a8c655c4e71\"" Aug 13 00:01:55.090337 env[1913]: time="2025-08-13T00:01:55.090281929Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.11\"" Aug 13 00:01:56.887012 env[1913]: time="2025-08-13T00:01:56.886948024Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:56.889753 env[1913]: time="2025-08-13T00:01:56.889674973Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:89be0efdc4ab1793b9b1b05e836e33dc50f5b2911b57609b315b58608b2d3746,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:56.893818 env[1913]: time="2025-08-13T00:01:56.893746412Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:56.897583 env[1913]: time="2025-08-13T00:01:56.897506746Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:1a9b59b3bfa6c1f1911f6f865a795620c461d079e413061bb71981cadd67f39d,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:56.899791 env[1913]: time="2025-08-13T00:01:56.899694239Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.11\" returns image reference \"sha256:89be0efdc4ab1793b9b1b05e836e33dc50f5b2911b57609b315b58608b2d3746\"" Aug 13 00:01:56.900755 env[1913]: time="2025-08-13T00:01:56.900697981Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.11\"" Aug 13 00:01:58.292872 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3804376659.mount: Deactivated successfully. Aug 13 00:01:59.188791 env[1913]: time="2025-08-13T00:01:59.188705571Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:59.197488 env[1913]: time="2025-08-13T00:01:59.197419898Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:7d1e7db6660181423f98acbe3a495b3fe5cec9b85cdef245540cc2cb3b180ab0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:59.203185 env[1913]: time="2025-08-13T00:01:59.203090998Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.31.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:59.210536 env[1913]: time="2025-08-13T00:01:59.210476709Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:a31da847792c5e7e92e91b78da1ad21d693e4b2b48d0e9f4610c8764dc2a5d79,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:01:59.211195 env[1913]: time="2025-08-13T00:01:59.211151760Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.11\" returns image reference \"sha256:7d1e7db6660181423f98acbe3a495b3fe5cec9b85cdef245540cc2cb3b180ab0\"" Aug 13 00:01:59.212363 env[1913]: time="2025-08-13T00:01:59.212317576Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Aug 13 00:01:59.811691 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2885655542.mount: Deactivated successfully. Aug 13 00:02:00.901923 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Aug 13 00:02:00.912532 kernel: kauditd_printk_skb: 88 callbacks suppressed Aug 13 00:02:00.912620 kernel: audit: type=1130 audit(1755043320.902:192): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:00.902000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:00.902285 systemd[1]: Stopped kubelet.service. Aug 13 00:02:00.905039 systemd[1]: Starting kubelet.service... Aug 13 00:02:00.921578 kernel: audit: type=1131 audit(1755043320.902:193): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:00.902000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:01.268000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:01.269345 systemd[1]: Started kubelet.service. Aug 13 00:02:01.283490 kernel: audit: type=1130 audit(1755043321.268:194): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:01.380818 kubelet[2381]: E0813 00:02:01.380762 2381 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 00:02:01.384000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Aug 13 00:02:01.384689 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 00:02:01.385063 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 00:02:01.394215 kernel: audit: type=1131 audit(1755043321.384:195): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Aug 13 00:02:01.827840 env[1913]: time="2025-08-13T00:02:01.827755221Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.11.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:01.833559 env[1913]: time="2025-08-13T00:02:01.833467820Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:01.837661 env[1913]: time="2025-08-13T00:02:01.837613207Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.11.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:01.841604 env[1913]: time="2025-08-13T00:02:01.841556366Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:01.843233 env[1913]: time="2025-08-13T00:02:01.843187752Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" Aug 13 00:02:01.844008 env[1913]: time="2025-08-13T00:02:01.843962114Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Aug 13 00:02:02.303129 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2695532548.mount: Deactivated successfully. Aug 13 00:02:02.315779 env[1913]: time="2025-08-13T00:02:02.315718730Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:02.320306 env[1913]: time="2025-08-13T00:02:02.320260825Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:02.323997 env[1913]: time="2025-08-13T00:02:02.323933467Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:02.327471 env[1913]: time="2025-08-13T00:02:02.327409480Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:02.328741 env[1913]: time="2025-08-13T00:02:02.328696558Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Aug 13 00:02:02.329503 env[1913]: time="2025-08-13T00:02:02.329459172Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Aug 13 00:02:02.895405 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount424075134.mount: Deactivated successfully. Aug 13 00:02:06.486020 env[1913]: time="2025-08-13T00:02:06.485931894Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.15-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:06.495596 env[1913]: time="2025-08-13T00:02:06.495526061Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:06.500018 env[1913]: time="2025-08-13T00:02:06.499927038Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.15-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:06.510693 env[1913]: time="2025-08-13T00:02:06.510629712Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:06.512529 env[1913]: time="2025-08-13T00:02:06.512449254Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\"" Aug 13 00:02:07.099310 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Aug 13 00:02:07.098000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hostnamed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:07.111143 kernel: audit: type=1131 audit(1755043327.098:196): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hostnamed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:11.401913 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Aug 13 00:02:11.402304 systemd[1]: Stopped kubelet.service. Aug 13 00:02:11.401000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:11.405003 systemd[1]: Starting kubelet.service... Aug 13 00:02:11.401000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:11.423414 kernel: audit: type=1130 audit(1755043331.401:197): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:11.423548 kernel: audit: type=1131 audit(1755043331.401:198): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:11.724824 systemd[1]: Started kubelet.service. Aug 13 00:02:11.724000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:11.742162 kernel: audit: type=1130 audit(1755043331.724:199): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:11.830925 kubelet[2415]: E0813 00:02:11.830851 2415 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 00:02:11.833713 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 00:02:11.834135 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 00:02:11.833000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Aug 13 00:02:11.844148 kernel: audit: type=1131 audit(1755043331.833:200): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Aug 13 00:02:12.212285 systemd[1]: Stopped kubelet.service. Aug 13 00:02:12.211000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:12.218000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:12.223465 systemd[1]: Starting kubelet.service... Aug 13 00:02:12.229382 kernel: audit: type=1130 audit(1755043332.211:201): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:12.229509 kernel: audit: type=1131 audit(1755043332.218:202): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:12.281220 systemd[1]: Reloading. Aug 13 00:02:12.456658 /usr/lib/systemd/system-generators/torcx-generator[2450]: time="2025-08-13T00:02:12Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Aug 13 00:02:12.456729 /usr/lib/systemd/system-generators/torcx-generator[2450]: time="2025-08-13T00:02:12Z" level=info msg="torcx already run" Aug 13 00:02:12.677890 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Aug 13 00:02:12.678430 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Aug 13 00:02:12.717269 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 00:02:12.939169 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Aug 13 00:02:12.939684 systemd[1]: kubelet.service: Failed with result 'signal'. Aug 13 00:02:12.940567 systemd[1]: Stopped kubelet.service. Aug 13 00:02:12.939000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Aug 13 00:02:12.953161 kernel: audit: type=1130 audit(1755043332.939:203): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' Aug 13 00:02:12.954238 systemd[1]: Starting kubelet.service... Aug 13 00:02:13.264000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:13.263975 systemd[1]: Started kubelet.service. Aug 13 00:02:13.276140 kernel: audit: type=1130 audit(1755043333.264:204): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:13.352733 kubelet[2524]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 00:02:13.353339 kubelet[2524]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Aug 13 00:02:13.353445 kubelet[2524]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 00:02:13.353694 kubelet[2524]: I0813 00:02:13.353635 2524 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 00:02:14.159497 kubelet[2524]: I0813 00:02:14.159431 2524 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Aug 13 00:02:14.159497 kubelet[2524]: I0813 00:02:14.159480 2524 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 00:02:14.160789 kubelet[2524]: I0813 00:02:14.160745 2524 server.go:934] "Client rotation is on, will bootstrap in background" Aug 13 00:02:14.201604 kubelet[2524]: E0813 00:02:14.201556 2524 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.27.151:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.27.151:6443: connect: connection refused" logger="UnhandledError" Aug 13 00:02:14.203801 kubelet[2524]: I0813 00:02:14.203745 2524 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 00:02:14.214921 kubelet[2524]: E0813 00:02:14.214854 2524 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Aug 13 00:02:14.214921 kubelet[2524]: I0813 00:02:14.214908 2524 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Aug 13 00:02:14.222697 kubelet[2524]: I0813 00:02:14.222636 2524 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 00:02:14.223823 kubelet[2524]: I0813 00:02:14.223778 2524 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Aug 13 00:02:14.224149 kubelet[2524]: I0813 00:02:14.224068 2524 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 00:02:14.224447 kubelet[2524]: I0813 00:02:14.224149 2524 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-27-151","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Aug 13 00:02:14.224620 kubelet[2524]: I0813 00:02:14.224576 2524 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 00:02:14.224620 kubelet[2524]: I0813 00:02:14.224599 2524 container_manager_linux.go:300] "Creating device plugin manager" Aug 13 00:02:14.224943 kubelet[2524]: I0813 00:02:14.224903 2524 state_mem.go:36] "Initialized new in-memory state store" Aug 13 00:02:14.236095 kubelet[2524]: I0813 00:02:14.236041 2524 kubelet.go:408] "Attempting to sync node with API server" Aug 13 00:02:14.236095 kubelet[2524]: I0813 00:02:14.236091 2524 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 00:02:14.236283 kubelet[2524]: I0813 00:02:14.236156 2524 kubelet.go:314] "Adding apiserver pod source" Aug 13 00:02:14.236344 kubelet[2524]: I0813 00:02:14.236311 2524 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 00:02:14.268021 kubelet[2524]: W0813 00:02:14.267928 2524 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.27.151:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-27-151&limit=500&resourceVersion=0": dial tcp 172.31.27.151:6443: connect: connection refused Aug 13 00:02:14.268329 kubelet[2524]: E0813 00:02:14.268296 2524 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.27.151:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-27-151&limit=500&resourceVersion=0\": dial tcp 172.31.27.151:6443: connect: connection refused" logger="UnhandledError" Aug 13 00:02:14.268837 kubelet[2524]: I0813 00:02:14.268807 2524 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Aug 13 00:02:14.270277 kubelet[2524]: I0813 00:02:14.270245 2524 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Aug 13 00:02:14.270769 kubelet[2524]: W0813 00:02:14.270746 2524 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Aug 13 00:02:14.278080 kubelet[2524]: I0813 00:02:14.278043 2524 server.go:1274] "Started kubelet" Aug 13 00:02:14.304061 kubelet[2524]: W0813 00:02:14.303954 2524 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.27.151:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.27.151:6443: connect: connection refused Aug 13 00:02:14.304260 kubelet[2524]: E0813 00:02:14.304074 2524 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.27.151:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.27.151:6443: connect: connection refused" logger="UnhandledError" Aug 13 00:02:14.308000 audit[2524]: AVC avc: denied { mac_admin } for pid=2524 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:02:14.314339 kubelet[2524]: I0813 00:02:14.312193 2524 kubelet.go:1430] "Unprivileged containerized plugins might not work, could not set selinux context on plugin registration dir" path="/var/lib/kubelet/plugins_registry" err="setxattr /var/lib/kubelet/plugins_registry: invalid argument" Aug 13 00:02:14.314339 kubelet[2524]: I0813 00:02:14.312274 2524 kubelet.go:1434] "Unprivileged containerized plugins might not work, could not set selinux context on plugins dir" path="/var/lib/kubelet/plugins" err="setxattr /var/lib/kubelet/plugins: invalid argument" Aug 13 00:02:14.314339 kubelet[2524]: I0813 00:02:14.312432 2524 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 00:02:14.308000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" Aug 13 00:02:14.321194 kernel: audit: type=1400 audit(1755043334.308:205): avc: denied { mac_admin } for pid=2524 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:02:14.321334 kernel: audit: type=1401 audit(1755043334.308:205): op=setxattr invalid_context="system_u:object_r:container_file_t:s0" Aug 13 00:02:14.321382 kernel: audit: type=1300 audit(1755043334.308:205): arch=c00000b7 syscall=5 success=no exit=-22 a0=4000a78390 a1=4000c08468 a2=4000a78360 a3=25 items=0 ppid=1 pid=2524 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.308000 audit[2524]: SYSCALL arch=c00000b7 syscall=5 success=no exit=-22 a0=4000a78390 a1=4000c08468 a2=4000a78360 a3=25 items=0 ppid=1 pid=2524 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.308000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 Aug 13 00:02:14.343762 kernel: audit: type=1327 audit(1755043334.308:205): proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 Aug 13 00:02:14.343826 kernel: audit: type=1400 audit(1755043334.312:206): avc: denied { mac_admin } for pid=2524 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:02:14.312000 audit[2524]: AVC avc: denied { mac_admin } for pid=2524 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:02:14.346272 kubelet[2524]: E0813 00:02:14.309501 2524 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.27.151:6443/api/v1/namespaces/default/events\": dial tcp 172.31.27.151:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-27-151.185b2a9d290ec71b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-27-151,UID:ip-172-31-27-151,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-27-151,},FirstTimestamp:2025-08-13 00:02:14.278006555 +0000 UTC m=+0.994380776,LastTimestamp:2025-08-13 00:02:14.278006555 +0000 UTC m=+0.994380776,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-27-151,}" Aug 13 00:02:14.346906 kubelet[2524]: I0813 00:02:14.346862 2524 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 00:02:14.312000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" Aug 13 00:02:14.358720 kubelet[2524]: I0813 00:02:14.358682 2524 server.go:449] "Adding debug handlers to kubelet server" Aug 13 00:02:14.359376 kernel: audit: type=1401 audit(1755043334.312:206): op=setxattr invalid_context="system_u:object_r:container_file_t:s0" Aug 13 00:02:14.312000 audit[2524]: SYSCALL arch=c00000b7 syscall=5 success=no exit=-22 a0=4000a26a40 a1=4000c08480 a2=4000a78420 a3=25 items=0 ppid=1 pid=2524 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.312000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 Aug 13 00:02:14.353000 audit[2536]: NETFILTER_CFG table=mangle:26 family=2 entries=2 op=nft_register_chain pid=2536 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:14.353000 audit[2536]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=136 a0=3 a1=ffffe44d2c50 a2=0 a3=1 items=0 ppid=2524 pid=2536 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.353000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Aug 13 00:02:14.359000 audit[2537]: NETFILTER_CFG table=filter:27 family=2 entries=1 op=nft_register_chain pid=2537 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:14.359000 audit[2537]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffc7605c50 a2=0 a3=1 items=0 ppid=2524 pid=2537 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.359000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 Aug 13 00:02:14.362726 kubelet[2524]: I0813 00:02:14.347736 2524 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 00:02:14.364360 kubelet[2524]: I0813 00:02:14.347240 2524 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 00:02:14.364360 kubelet[2524]: I0813 00:02:14.364357 2524 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 00:02:14.364582 kubelet[2524]: E0813 00:02:14.352407 2524 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ip-172-31-27-151\" not found" Aug 13 00:02:14.364582 kubelet[2524]: I0813 00:02:14.354688 2524 factory.go:221] Registration of the systemd container factory successfully Aug 13 00:02:14.364716 kubelet[2524]: I0813 00:02:14.364623 2524 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 00:02:14.365127 kubelet[2524]: I0813 00:02:14.352034 2524 volume_manager.go:289] "Starting Kubelet Volume Manager" Aug 13 00:02:14.365623 kubelet[2524]: I0813 00:02:14.352085 2524 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Aug 13 00:02:14.365623 kubelet[2524]: E0813 00:02:14.354959 2524 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.27.151:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-27-151?timeout=10s\": dial tcp 172.31.27.151:6443: connect: connection refused" interval="200ms" Aug 13 00:02:14.365794 kubelet[2524]: W0813 00:02:14.354827 2524 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.27.151:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.27.151:6443: connect: connection refused Aug 13 00:02:14.365794 kubelet[2524]: E0813 00:02:14.365686 2524 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.27.151:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.27.151:6443: connect: connection refused" logger="UnhandledError" Aug 13 00:02:14.366097 kubelet[2524]: I0813 00:02:14.366059 2524 reconciler.go:26] "Reconciler: start to sync state" Aug 13 00:02:14.366262 kubelet[2524]: E0813 00:02:14.366235 2524 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 00:02:14.368535 kubelet[2524]: I0813 00:02:14.368483 2524 factory.go:221] Registration of the containerd container factory successfully Aug 13 00:02:14.368000 audit[2539]: NETFILTER_CFG table=filter:28 family=2 entries=2 op=nft_register_chain pid=2539 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:14.368000 audit[2539]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=312 a0=3 a1=ffffca530890 a2=0 a3=1 items=0 ppid=2524 pid=2539 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.368000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Aug 13 00:02:14.374000 audit[2541]: NETFILTER_CFG table=filter:29 family=2 entries=2 op=nft_register_chain pid=2541 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:14.374000 audit[2541]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=312 a0=3 a1=fffffe2eaf40 a2=0 a3=1 items=0 ppid=2524 pid=2541 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.374000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Aug 13 00:02:14.401000 audit[2547]: NETFILTER_CFG table=filter:30 family=2 entries=1 op=nft_register_rule pid=2547 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:14.401000 audit[2547]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=924 a0=3 a1=ffffe8468770 a2=0 a3=1 items=0 ppid=2524 pid=2547 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.401000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D41004B5542452D4649524557414C4C002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E7400626C6F636B20696E636F6D696E67206C6F63616C6E657420636F6E6E656374696F6E73002D2D647374003132372E302E302E302F38 Aug 13 00:02:14.403577 kubelet[2524]: I0813 00:02:14.403524 2524 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Aug 13 00:02:14.404000 audit[2548]: NETFILTER_CFG table=mangle:31 family=10 entries=2 op=nft_register_chain pid=2548 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:14.404000 audit[2548]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=136 a0=3 a1=ffffd593bb60 a2=0 a3=1 items=0 ppid=2524 pid=2548 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.404000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 Aug 13 00:02:14.406462 kubelet[2524]: I0813 00:02:14.406426 2524 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Aug 13 00:02:14.406658 kubelet[2524]: I0813 00:02:14.406634 2524 status_manager.go:217] "Starting to sync pod status with apiserver" Aug 13 00:02:14.406785 kubelet[2524]: I0813 00:02:14.406764 2524 kubelet.go:2321] "Starting kubelet main sync loop" Aug 13 00:02:14.406971 kubelet[2524]: E0813 00:02:14.406941 2524 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 00:02:14.408000 audit[2549]: NETFILTER_CFG table=mangle:32 family=2 entries=1 op=nft_register_chain pid=2549 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:14.408000 audit[2549]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=fffffa98aea0 a2=0 a3=1 items=0 ppid=2524 pid=2549 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.408000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Aug 13 00:02:14.414000 audit[2550]: NETFILTER_CFG table=mangle:33 family=10 entries=1 op=nft_register_chain pid=2550 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:14.414000 audit[2550]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffeee54f70 a2=0 a3=1 items=0 ppid=2524 pid=2550 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.414000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 Aug 13 00:02:14.415000 audit[2551]: NETFILTER_CFG table=nat:34 family=2 entries=1 op=nft_register_chain pid=2551 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:14.415000 audit[2551]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffc425db10 a2=0 a3=1 items=0 ppid=2524 pid=2551 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.415000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Aug 13 00:02:14.417000 audit[2552]: NETFILTER_CFG table=nat:35 family=10 entries=2 op=nft_register_chain pid=2552 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:14.417000 audit[2552]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=128 a0=3 a1=ffffd70d0180 a2=0 a3=1 items=0 ppid=2524 pid=2552 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.417000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 Aug 13 00:02:14.418000 audit[2553]: NETFILTER_CFG table=filter:36 family=2 entries=1 op=nft_register_chain pid=2553 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:14.418000 audit[2553]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffd62ef1e0 a2=0 a3=1 items=0 ppid=2524 pid=2553 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.418000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Aug 13 00:02:14.419000 audit[2554]: NETFILTER_CFG table=filter:37 family=10 entries=2 op=nft_register_chain pid=2554 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:14.419000 audit[2554]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=136 a0=3 a1=ffffe1f60340 a2=0 a3=1 items=0 ppid=2524 pid=2554 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.419000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 Aug 13 00:02:14.423879 kubelet[2524]: W0813 00:02:14.423818 2524 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.27.151:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.27.151:6443: connect: connection refused Aug 13 00:02:14.424072 kubelet[2524]: E0813 00:02:14.423893 2524 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.27.151:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.27.151:6443: connect: connection refused" logger="UnhandledError" Aug 13 00:02:14.424724 kubelet[2524]: I0813 00:02:14.424676 2524 cpu_manager.go:214] "Starting CPU manager" policy="none" Aug 13 00:02:14.424724 kubelet[2524]: I0813 00:02:14.424709 2524 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Aug 13 00:02:14.424897 kubelet[2524]: I0813 00:02:14.424742 2524 state_mem.go:36] "Initialized new in-memory state store" Aug 13 00:02:14.430149 kubelet[2524]: I0813 00:02:14.430080 2524 policy_none.go:49] "None policy: Start" Aug 13 00:02:14.431283 kubelet[2524]: I0813 00:02:14.431235 2524 memory_manager.go:170] "Starting memorymanager" policy="None" Aug 13 00:02:14.431283 kubelet[2524]: I0813 00:02:14.431285 2524 state_mem.go:35] "Initializing new in-memory state store" Aug 13 00:02:14.444468 kubelet[2524]: I0813 00:02:14.444428 2524 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Aug 13 00:02:14.443000 audit[2524]: AVC avc: denied { mac_admin } for pid=2524 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:02:14.443000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" Aug 13 00:02:14.443000 audit[2524]: SYSCALL arch=c00000b7 syscall=5 success=no exit=-22 a0=4000ee6f00 a1=4000e498d8 a2=4000ee6ed0 a3=25 items=0 ppid=1 pid=2524 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:14.443000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 Aug 13 00:02:14.445179 kubelet[2524]: I0813 00:02:14.445087 2524 server.go:88] "Unprivileged containerized plugins might not work. Could not set selinux context on socket dir" path="/var/lib/kubelet/device-plugins/" err="setxattr /var/lib/kubelet/device-plugins/: invalid argument" Aug 13 00:02:14.445482 kubelet[2524]: I0813 00:02:14.445459 2524 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 00:02:14.445634 kubelet[2524]: I0813 00:02:14.445582 2524 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 00:02:14.449167 kubelet[2524]: I0813 00:02:14.449079 2524 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 00:02:14.452260 kubelet[2524]: E0813 00:02:14.452226 2524 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-27-151\" not found" Aug 13 00:02:14.548174 kubelet[2524]: I0813 00:02:14.548082 2524 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-27-151" Aug 13 00:02:14.548959 kubelet[2524]: E0813 00:02:14.548913 2524 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.27.151:6443/api/v1/nodes\": dial tcp 172.31.27.151:6443: connect: connection refused" node="ip-172-31-27-151" Aug 13 00:02:14.566648 kubelet[2524]: E0813 00:02:14.566601 2524 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.27.151:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-27-151?timeout=10s\": dial tcp 172.31.27.151:6443: connect: connection refused" interval="400ms" Aug 13 00:02:14.567857 kubelet[2524]: I0813 00:02:14.567820 2524 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5e3ffd7fdb1d7770aed8e1a0ec687436-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-27-151\" (UID: \"5e3ffd7fdb1d7770aed8e1a0ec687436\") " pod="kube-system/kube-controller-manager-ip-172-31-27-151" Aug 13 00:02:14.567958 kubelet[2524]: I0813 00:02:14.567878 2524 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5e3ffd7fdb1d7770aed8e1a0ec687436-k8s-certs\") pod \"kube-controller-manager-ip-172-31-27-151\" (UID: \"5e3ffd7fdb1d7770aed8e1a0ec687436\") " pod="kube-system/kube-controller-manager-ip-172-31-27-151" Aug 13 00:02:14.567958 kubelet[2524]: I0813 00:02:14.567920 2524 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d246503ef314f9f2381c6bdef310f4af-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-27-151\" (UID: \"d246503ef314f9f2381c6bdef310f4af\") " pod="kube-system/kube-apiserver-ip-172-31-27-151" Aug 13 00:02:14.568121 kubelet[2524]: I0813 00:02:14.567958 2524 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d246503ef314f9f2381c6bdef310f4af-ca-certs\") pod \"kube-apiserver-ip-172-31-27-151\" (UID: \"d246503ef314f9f2381c6bdef310f4af\") " pod="kube-system/kube-apiserver-ip-172-31-27-151" Aug 13 00:02:14.568121 kubelet[2524]: I0813 00:02:14.567995 2524 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d246503ef314f9f2381c6bdef310f4af-k8s-certs\") pod \"kube-apiserver-ip-172-31-27-151\" (UID: \"d246503ef314f9f2381c6bdef310f4af\") " pod="kube-system/kube-apiserver-ip-172-31-27-151" Aug 13 00:02:14.568121 kubelet[2524]: I0813 00:02:14.568029 2524 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5e3ffd7fdb1d7770aed8e1a0ec687436-ca-certs\") pod \"kube-controller-manager-ip-172-31-27-151\" (UID: \"5e3ffd7fdb1d7770aed8e1a0ec687436\") " pod="kube-system/kube-controller-manager-ip-172-31-27-151" Aug 13 00:02:14.568121 kubelet[2524]: I0813 00:02:14.568064 2524 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5e3ffd7fdb1d7770aed8e1a0ec687436-kubeconfig\") pod \"kube-controller-manager-ip-172-31-27-151\" (UID: \"5e3ffd7fdb1d7770aed8e1a0ec687436\") " pod="kube-system/kube-controller-manager-ip-172-31-27-151" Aug 13 00:02:14.568354 kubelet[2524]: I0813 00:02:14.568120 2524 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5e3ffd7fdb1d7770aed8e1a0ec687436-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-27-151\" (UID: \"5e3ffd7fdb1d7770aed8e1a0ec687436\") " pod="kube-system/kube-controller-manager-ip-172-31-27-151" Aug 13 00:02:14.568354 kubelet[2524]: I0813 00:02:14.568163 2524 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e8468136087d101cffab1e36b9bf9565-kubeconfig\") pod \"kube-scheduler-ip-172-31-27-151\" (UID: \"e8468136087d101cffab1e36b9bf9565\") " pod="kube-system/kube-scheduler-ip-172-31-27-151" Aug 13 00:02:14.753175 kubelet[2524]: I0813 00:02:14.750987 2524 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-27-151" Aug 13 00:02:14.753545 kubelet[2524]: E0813 00:02:14.753503 2524 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.27.151:6443/api/v1/nodes\": dial tcp 172.31.27.151:6443: connect: connection refused" node="ip-172-31-27-151" Aug 13 00:02:14.821918 env[1913]: time="2025-08-13T00:02:14.821843772Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-27-151,Uid:d246503ef314f9f2381c6bdef310f4af,Namespace:kube-system,Attempt:0,}" Aug 13 00:02:14.832533 env[1913]: time="2025-08-13T00:02:14.832479419Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-27-151,Uid:5e3ffd7fdb1d7770aed8e1a0ec687436,Namespace:kube-system,Attempt:0,}" Aug 13 00:02:14.836791 env[1913]: time="2025-08-13T00:02:14.836704419Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-27-151,Uid:e8468136087d101cffab1e36b9bf9565,Namespace:kube-system,Attempt:0,}" Aug 13 00:02:14.967891 kubelet[2524]: E0813 00:02:14.967808 2524 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.27.151:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-27-151?timeout=10s\": dial tcp 172.31.27.151:6443: connect: connection refused" interval="800ms" Aug 13 00:02:15.156456 kubelet[2524]: I0813 00:02:15.155948 2524 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-27-151" Aug 13 00:02:15.156456 kubelet[2524]: E0813 00:02:15.156409 2524 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.27.151:6443/api/v1/nodes\": dial tcp 172.31.27.151:6443: connect: connection refused" node="ip-172-31-27-151" Aug 13 00:02:15.473449 kubelet[2524]: W0813 00:02:15.473256 2524 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.27.151:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.27.151:6443: connect: connection refused Aug 13 00:02:15.473449 kubelet[2524]: E0813 00:02:15.473359 2524 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.27.151:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.27.151:6443: connect: connection refused" logger="UnhandledError" Aug 13 00:02:15.608171 kubelet[2524]: W0813 00:02:15.608053 2524 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.27.151:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.27.151:6443: connect: connection refused Aug 13 00:02:15.608323 kubelet[2524]: E0813 00:02:15.608183 2524 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.27.151:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.27.151:6443: connect: connection refused" logger="UnhandledError" Aug 13 00:02:15.661518 kubelet[2524]: W0813 00:02:15.661375 2524 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.27.151:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.27.151:6443: connect: connection refused Aug 13 00:02:15.661518 kubelet[2524]: E0813 00:02:15.661465 2524 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.27.151:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.27.151:6443: connect: connection refused" logger="UnhandledError" Aug 13 00:02:15.769335 kubelet[2524]: E0813 00:02:15.769184 2524 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.27.151:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-27-151?timeout=10s\": dial tcp 172.31.27.151:6443: connect: connection refused" interval="1.6s" Aug 13 00:02:15.779997 kubelet[2524]: W0813 00:02:15.779890 2524 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.27.151:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-27-151&limit=500&resourceVersion=0": dial tcp 172.31.27.151:6443: connect: connection refused Aug 13 00:02:15.780219 kubelet[2524]: E0813 00:02:15.780004 2524 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.27.151:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-27-151&limit=500&resourceVersion=0\": dial tcp 172.31.27.151:6443: connect: connection refused" logger="UnhandledError" Aug 13 00:02:15.959022 kubelet[2524]: I0813 00:02:15.958980 2524 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-27-151" Aug 13 00:02:15.959502 kubelet[2524]: E0813 00:02:15.959460 2524 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.27.151:6443/api/v1/nodes\": dial tcp 172.31.27.151:6443: connect: connection refused" node="ip-172-31-27-151" Aug 13 00:02:16.210933 kubelet[2524]: E0813 00:02:16.210871 2524 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.27.151:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.27.151:6443: connect: connection refused" logger="UnhandledError" Aug 13 00:02:17.350211 amazon-ssm-agent[1883]: 2025-08-13 00:02:17 INFO [MessagingDeliveryService] [Association] Schedule manager refreshed with 0 associations, 0 new associations associated Aug 13 00:02:17.370313 kubelet[2524]: E0813 00:02:17.370227 2524 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.27.151:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-27-151?timeout=10s\": dial tcp 172.31.27.151:6443: connect: connection refused" interval="3.2s" Aug 13 00:02:17.538741 kubelet[2524]: W0813 00:02:17.538627 2524 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.27.151:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.27.151:6443: connect: connection refused Aug 13 00:02:17.538741 kubelet[2524]: E0813 00:02:17.538693 2524 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.27.151:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.27.151:6443: connect: connection refused" logger="UnhandledError" Aug 13 00:02:17.551500 kubelet[2524]: W0813 00:02:17.551397 2524 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.27.151:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.27.151:6443: connect: connection refused Aug 13 00:02:17.551500 kubelet[2524]: E0813 00:02:17.551456 2524 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.27.151:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.27.151:6443: connect: connection refused" logger="UnhandledError" Aug 13 00:02:17.561683 kubelet[2524]: I0813 00:02:17.561651 2524 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-27-151" Aug 13 00:02:17.562402 kubelet[2524]: E0813 00:02:17.562319 2524 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.27.151:6443/api/v1/nodes\": dial tcp 172.31.27.151:6443: connect: connection refused" node="ip-172-31-27-151" Aug 13 00:02:17.621538 kubelet[2524]: W0813 00:02:17.621390 2524 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.27.151:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.27.151:6443: connect: connection refused Aug 13 00:02:17.621538 kubelet[2524]: E0813 00:02:17.621470 2524 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.27.151:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.27.151:6443: connect: connection refused" logger="UnhandledError" Aug 13 00:02:18.589426 amazon-ssm-agent[1883]: 2025-08-13 00:02:18 INFO [HealthCheck] HealthCheck reporting agent health. Aug 13 00:02:18.822176 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount382913974.mount: Deactivated successfully. Aug 13 00:02:18.842391 env[1913]: time="2025-08-13T00:02:18.841911214Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:18.846189 env[1913]: time="2025-08-13T00:02:18.846092807Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:18.848422 env[1913]: time="2025-08-13T00:02:18.848368396Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:18.851419 env[1913]: time="2025-08-13T00:02:18.851332508Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:18.856148 env[1913]: time="2025-08-13T00:02:18.856047799Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:18.860444 env[1913]: time="2025-08-13T00:02:18.860396552Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:18.862480 env[1913]: time="2025-08-13T00:02:18.862436322Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:18.864856 env[1913]: time="2025-08-13T00:02:18.864812392Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:18.869263 env[1913]: time="2025-08-13T00:02:18.869210653Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:18.873538 env[1913]: time="2025-08-13T00:02:18.873488047Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:18.875837 env[1913]: time="2025-08-13T00:02:18.875738318Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:18.881423 env[1913]: time="2025-08-13T00:02:18.881372227Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:18.946728 kubelet[2524]: W0813 00:02:18.946670 2524 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.27.151:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-27-151&limit=500&resourceVersion=0": dial tcp 172.31.27.151:6443: connect: connection refused Aug 13 00:02:18.947367 kubelet[2524]: E0813 00:02:18.946743 2524 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.27.151:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-27-151&limit=500&resourceVersion=0\": dial tcp 172.31.27.151:6443: connect: connection refused" logger="UnhandledError" Aug 13 00:02:18.955378 env[1913]: time="2025-08-13T00:02:18.955252357Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:02:18.955540 env[1913]: time="2025-08-13T00:02:18.955417921Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:02:18.955540 env[1913]: time="2025-08-13T00:02:18.955504807Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:02:18.956299 env[1913]: time="2025-08-13T00:02:18.956196701Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/0ad07cd3a1ae7e9859a09c7d9b4e6c9ee783000534ce682bef3fcae8a67509a8 pid=2573 runtime=io.containerd.runc.v2 Aug 13 00:02:18.966277 env[1913]: time="2025-08-13T00:02:18.966016295Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:02:18.966277 env[1913]: time="2025-08-13T00:02:18.966198046Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:02:18.966556 env[1913]: time="2025-08-13T00:02:18.966230023Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:02:18.966802 env[1913]: time="2025-08-13T00:02:18.966665760Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:02:18.966909 env[1913]: time="2025-08-13T00:02:18.966768401Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:02:18.966909 env[1913]: time="2025-08-13T00:02:18.966821197Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:02:18.968091 env[1913]: time="2025-08-13T00:02:18.967994197Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/7b13e48260a3daa908e01728139506669e308358000cb0d65df84fa293b42635 pid=2577 runtime=io.containerd.runc.v2 Aug 13 00:02:18.969637 env[1913]: time="2025-08-13T00:02:18.969518652Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/95d41164c573ec79b9b8ea7a51e300381cb70349b0c4ea7fb1f6739cf9bab936 pid=2584 runtime=io.containerd.runc.v2 Aug 13 00:02:19.145607 env[1913]: time="2025-08-13T00:02:19.143935691Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-27-151,Uid:d246503ef314f9f2381c6bdef310f4af,Namespace:kube-system,Attempt:0,} returns sandbox id \"95d41164c573ec79b9b8ea7a51e300381cb70349b0c4ea7fb1f6739cf9bab936\"" Aug 13 00:02:19.155002 env[1913]: time="2025-08-13T00:02:19.154624430Z" level=info msg="CreateContainer within sandbox \"95d41164c573ec79b9b8ea7a51e300381cb70349b0c4ea7fb1f6739cf9bab936\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Aug 13 00:02:19.160818 env[1913]: time="2025-08-13T00:02:19.160740180Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-27-151,Uid:e8468136087d101cffab1e36b9bf9565,Namespace:kube-system,Attempt:0,} returns sandbox id \"7b13e48260a3daa908e01728139506669e308358000cb0d65df84fa293b42635\"" Aug 13 00:02:19.182301 env[1913]: time="2025-08-13T00:02:19.181287542Z" level=info msg="CreateContainer within sandbox \"7b13e48260a3daa908e01728139506669e308358000cb0d65df84fa293b42635\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Aug 13 00:02:19.186383 env[1913]: time="2025-08-13T00:02:19.186289862Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-27-151,Uid:5e3ffd7fdb1d7770aed8e1a0ec687436,Namespace:kube-system,Attempt:0,} returns sandbox id \"0ad07cd3a1ae7e9859a09c7d9b4e6c9ee783000534ce682bef3fcae8a67509a8\"" Aug 13 00:02:19.195781 env[1913]: time="2025-08-13T00:02:19.195725957Z" level=info msg="CreateContainer within sandbox \"0ad07cd3a1ae7e9859a09c7d9b4e6c9ee783000534ce682bef3fcae8a67509a8\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Aug 13 00:02:19.218413 env[1913]: time="2025-08-13T00:02:19.218346636Z" level=info msg="CreateContainer within sandbox \"95d41164c573ec79b9b8ea7a51e300381cb70349b0c4ea7fb1f6739cf9bab936\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"1969ca48869008ead227382e6eb0f00cacc74e8d15cd58e754b9cd79090540f3\"" Aug 13 00:02:19.219518 env[1913]: time="2025-08-13T00:02:19.219463893Z" level=info msg="StartContainer for \"1969ca48869008ead227382e6eb0f00cacc74e8d15cd58e754b9cd79090540f3\"" Aug 13 00:02:19.226085 env[1913]: time="2025-08-13T00:02:19.226013066Z" level=info msg="CreateContainer within sandbox \"7b13e48260a3daa908e01728139506669e308358000cb0d65df84fa293b42635\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"cf15ca7f436b364bb2bd07bc5ede63041902ab077512643f2681b8d2059cf9db\"" Aug 13 00:02:19.227121 env[1913]: time="2025-08-13T00:02:19.226968813Z" level=info msg="StartContainer for \"cf15ca7f436b364bb2bd07bc5ede63041902ab077512643f2681b8d2059cf9db\"" Aug 13 00:02:19.241291 env[1913]: time="2025-08-13T00:02:19.241211594Z" level=info msg="CreateContainer within sandbox \"0ad07cd3a1ae7e9859a09c7d9b4e6c9ee783000534ce682bef3fcae8a67509a8\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"cbfb4f47c3489a82e524f9b8fc898e2372b9e472524b9f4c97f7759399b8ea45\"" Aug 13 00:02:19.242064 env[1913]: time="2025-08-13T00:02:19.242001669Z" level=info msg="StartContainer for \"cbfb4f47c3489a82e524f9b8fc898e2372b9e472524b9f4c97f7759399b8ea45\"" Aug 13 00:02:19.426836 env[1913]: time="2025-08-13T00:02:19.426778907Z" level=info msg="StartContainer for \"1969ca48869008ead227382e6eb0f00cacc74e8d15cd58e754b9cd79090540f3\" returns successfully" Aug 13 00:02:19.527169 env[1913]: time="2025-08-13T00:02:19.526311018Z" level=info msg="StartContainer for \"cf15ca7f436b364bb2bd07bc5ede63041902ab077512643f2681b8d2059cf9db\" returns successfully" Aug 13 00:02:19.572346 env[1913]: time="2025-08-13T00:02:19.572264648Z" level=info msg="StartContainer for \"cbfb4f47c3489a82e524f9b8fc898e2372b9e472524b9f4c97f7759399b8ea45\" returns successfully" Aug 13 00:02:20.764400 kubelet[2524]: I0813 00:02:20.764364 2524 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-27-151" Aug 13 00:02:21.426956 update_engine[1902]: I0813 00:02:21.426160 1902 update_attempter.cc:509] Updating boot flags... Aug 13 00:02:24.102035 kubelet[2524]: E0813 00:02:24.101987 2524 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-172-31-27-151\" not found" node="ip-172-31-27-151" Aug 13 00:02:24.127111 kubelet[2524]: E0813 00:02:24.126904 2524 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ip-172-31-27-151.185b2a9d290ec71b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-27-151,UID:ip-172-31-27-151,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-27-151,},FirstTimestamp:2025-08-13 00:02:14.278006555 +0000 UTC m=+0.994380776,LastTimestamp:2025-08-13 00:02:14.278006555 +0000 UTC m=+0.994380776,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-27-151,}" Aug 13 00:02:24.181491 kubelet[2524]: I0813 00:02:24.181446 2524 kubelet_node_status.go:75] "Successfully registered node" node="ip-172-31-27-151" Aug 13 00:02:24.181734 kubelet[2524]: E0813 00:02:24.181706 2524 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"ip-172-31-27-151\": node \"ip-172-31-27-151\" not found" Aug 13 00:02:24.235935 kubelet[2524]: E0813 00:02:24.235795 2524 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ip-172-31-27-151.185b2a9d2b1c559c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-27-151,UID:ip-172-31-27-151,APIVersion:,ResourceVersion:,FieldPath:,},Reason:CgroupV1,Message:Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.,Source:EventSource{Component:kubelet,Host:ip-172-31-27-151,},FirstTimestamp:2025-08-13 00:02:14.312449436 +0000 UTC m=+1.028823681,LastTimestamp:2025-08-13 00:02:14.312449436 +0000 UTC m=+1.028823681,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-27-151,}" Aug 13 00:02:24.297363 kubelet[2524]: I0813 00:02:24.297274 2524 apiserver.go:52] "Watching apiserver" Aug 13 00:02:24.366678 kubelet[2524]: I0813 00:02:24.366506 2524 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Aug 13 00:02:26.273706 systemd[1]: Reloading. Aug 13 00:02:26.423210 /usr/lib/systemd/system-generators/torcx-generator[2917]: time="2025-08-13T00:02:26Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Aug 13 00:02:26.424148 /usr/lib/systemd/system-generators/torcx-generator[2917]: time="2025-08-13T00:02:26Z" level=info msg="torcx already run" Aug 13 00:02:26.613011 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Aug 13 00:02:26.613054 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Aug 13 00:02:26.655489 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 00:02:26.885999 systemd[1]: Stopping kubelet.service... Aug 13 00:02:26.910000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:26.911174 systemd[1]: kubelet.service: Deactivated successfully. Aug 13 00:02:26.911805 systemd[1]: Stopped kubelet.service. Aug 13 00:02:26.913983 kernel: kauditd_printk_skb: 42 callbacks suppressed Aug 13 00:02:26.914073 kernel: audit: type=1131 audit(1755043346.910:220): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:26.916895 systemd[1]: Starting kubelet.service... Aug 13 00:02:27.240740 systemd[1]: Started kubelet.service. Aug 13 00:02:27.240000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:27.272198 kernel: audit: type=1130 audit(1755043347.240:221): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:27.364922 kubelet[2984]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 00:02:27.364922 kubelet[2984]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Aug 13 00:02:27.364922 kubelet[2984]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 00:02:27.364922 kubelet[2984]: I0813 00:02:27.363604 2984 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 00:02:27.381084 kubelet[2984]: I0813 00:02:27.381040 2984 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Aug 13 00:02:27.381299 kubelet[2984]: I0813 00:02:27.381276 2984 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 00:02:27.381858 kubelet[2984]: I0813 00:02:27.381830 2984 server.go:934] "Client rotation is on, will bootstrap in background" Aug 13 00:02:27.384901 kubelet[2984]: I0813 00:02:27.384852 2984 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Aug 13 00:02:27.397059 kubelet[2984]: I0813 00:02:27.396986 2984 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 00:02:27.407602 kubelet[2984]: E0813 00:02:27.407545 2984 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Aug 13 00:02:27.407602 kubelet[2984]: I0813 00:02:27.407603 2984 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Aug 13 00:02:27.413477 kubelet[2984]: I0813 00:02:27.412870 2984 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 00:02:27.413673 kubelet[2984]: I0813 00:02:27.413638 2984 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Aug 13 00:02:27.413906 kubelet[2984]: I0813 00:02:27.413849 2984 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 00:02:27.415839 kubelet[2984]: I0813 00:02:27.413908 2984 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-27-151","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Aug 13 00:02:27.415839 kubelet[2984]: I0813 00:02:27.414266 2984 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 00:02:27.415839 kubelet[2984]: I0813 00:02:27.414289 2984 container_manager_linux.go:300] "Creating device plugin manager" Aug 13 00:02:27.415839 kubelet[2984]: I0813 00:02:27.414352 2984 state_mem.go:36] "Initialized new in-memory state store" Aug 13 00:02:27.415839 kubelet[2984]: I0813 00:02:27.414529 2984 kubelet.go:408] "Attempting to sync node with API server" Aug 13 00:02:27.416343 kubelet[2984]: I0813 00:02:27.414552 2984 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 00:02:27.421211 kubelet[2984]: I0813 00:02:27.421151 2984 kubelet.go:314] "Adding apiserver pod source" Aug 13 00:02:27.421211 kubelet[2984]: I0813 00:02:27.421219 2984 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 00:02:27.431570 kubelet[2984]: I0813 00:02:27.431511 2984 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Aug 13 00:02:27.432388 kubelet[2984]: I0813 00:02:27.432346 2984 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Aug 13 00:02:27.433750 kubelet[2984]: I0813 00:02:27.433044 2984 server.go:1274] "Started kubelet" Aug 13 00:02:27.436000 audit[2984]: AVC avc: denied { mac_admin } for pid=2984 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:02:27.447012 kubelet[2984]: I0813 00:02:27.437227 2984 kubelet.go:1430] "Unprivileged containerized plugins might not work, could not set selinux context on plugin registration dir" path="/var/lib/kubelet/plugins_registry" err="setxattr /var/lib/kubelet/plugins_registry: invalid argument" Aug 13 00:02:27.447012 kubelet[2984]: I0813 00:02:27.437292 2984 kubelet.go:1434] "Unprivileged containerized plugins might not work, could not set selinux context on plugins dir" path="/var/lib/kubelet/plugins" err="setxattr /var/lib/kubelet/plugins: invalid argument" Aug 13 00:02:27.447012 kubelet[2984]: I0813 00:02:27.437347 2984 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 00:02:27.447012 kubelet[2984]: I0813 00:02:27.446139 2984 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 00:02:27.447659 kubelet[2984]: I0813 00:02:27.447612 2984 server.go:449] "Adding debug handlers to kubelet server" Aug 13 00:02:27.456860 kernel: audit: type=1400 audit(1755043347.436:222): avc: denied { mac_admin } for pid=2984 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:02:27.456997 kernel: audit: type=1401 audit(1755043347.436:222): op=setxattr invalid_context="system_u:object_r:container_file_t:s0" Aug 13 00:02:27.436000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" Aug 13 00:02:27.457628 kubelet[2984]: I0813 00:02:27.457571 2984 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 00:02:27.473958 kernel: audit: type=1300 audit(1755043347.436:222): arch=c00000b7 syscall=5 success=no exit=-22 a0=4000ba0360 a1=4000ae25a0 a2=4000ba0330 a3=25 items=0 ppid=1 pid=2984 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:27.436000 audit[2984]: SYSCALL arch=c00000b7 syscall=5 success=no exit=-22 a0=4000ba0360 a1=4000ae25a0 a2=4000ba0330 a3=25 items=0 ppid=1 pid=2984 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:27.474265 kubelet[2984]: I0813 00:02:27.461465 2984 volume_manager.go:289] "Starting Kubelet Volume Manager" Aug 13 00:02:27.474265 kubelet[2984]: E0813 00:02:27.461886 2984 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ip-172-31-27-151\" not found" Aug 13 00:02:27.474265 kubelet[2984]: I0813 00:02:27.465483 2984 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Aug 13 00:02:27.474265 kubelet[2984]: I0813 00:02:27.466014 2984 reconciler.go:26] "Reconciler: start to sync state" Aug 13 00:02:27.474265 kubelet[2984]: I0813 00:02:27.469896 2984 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Aug 13 00:02:27.474265 kubelet[2984]: I0813 00:02:27.474010 2984 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Aug 13 00:02:27.474265 kubelet[2984]: I0813 00:02:27.474054 2984 status_manager.go:217] "Starting to sync pod status with apiserver" Aug 13 00:02:27.474265 kubelet[2984]: I0813 00:02:27.474087 2984 kubelet.go:2321] "Starting kubelet main sync loop" Aug 13 00:02:27.474265 kubelet[2984]: E0813 00:02:27.474217 2984 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 00:02:27.477728 kubelet[2984]: I0813 00:02:27.477592 2984 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 00:02:27.480450 kubelet[2984]: I0813 00:02:27.480422 2984 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 00:02:27.436000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 Aug 13 00:02:27.501194 kernel: audit: type=1327 audit(1755043347.436:222): proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 Aug 13 00:02:27.436000 audit[2984]: AVC avc: denied { mac_admin } for pid=2984 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:02:27.519672 kernel: audit: type=1400 audit(1755043347.436:223): avc: denied { mac_admin } for pid=2984 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:02:27.519785 kernel: audit: type=1401 audit(1755043347.436:223): op=setxattr invalid_context="system_u:object_r:container_file_t:s0" Aug 13 00:02:27.436000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" Aug 13 00:02:27.436000 audit[2984]: SYSCALL arch=c00000b7 syscall=5 success=no exit=-22 a0=4000a7cc60 a1=4000ae25b8 a2=4000ba03f0 a3=25 items=0 ppid=1 pid=2984 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:27.540205 kernel: audit: type=1300 audit(1755043347.436:223): arch=c00000b7 syscall=5 success=no exit=-22 a0=4000a7cc60 a1=4000ae25b8 a2=4000ba03f0 a3=25 items=0 ppid=1 pid=2984 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:27.436000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 Aug 13 00:02:27.552194 kernel: audit: type=1327 audit(1755043347.436:223): proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 Aug 13 00:02:27.553406 kubelet[2984]: I0813 00:02:27.553368 2984 factory.go:221] Registration of the systemd container factory successfully Aug 13 00:02:27.562076 kubelet[2984]: E0813 00:02:27.562039 2984 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ip-172-31-27-151\" not found" Aug 13 00:02:27.563051 kubelet[2984]: I0813 00:02:27.563008 2984 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 00:02:27.581299 kubelet[2984]: E0813 00:02:27.581254 2984 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 00:02:27.589253 kubelet[2984]: I0813 00:02:27.588483 2984 factory.go:221] Registration of the containerd container factory successfully Aug 13 00:02:27.597352 kubelet[2984]: E0813 00:02:27.595930 2984 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 00:02:27.740821 kubelet[2984]: I0813 00:02:27.740697 2984 cpu_manager.go:214] "Starting CPU manager" policy="none" Aug 13 00:02:27.740996 kubelet[2984]: I0813 00:02:27.740834 2984 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Aug 13 00:02:27.740996 kubelet[2984]: I0813 00:02:27.740872 2984 state_mem.go:36] "Initialized new in-memory state store" Aug 13 00:02:27.741258 kubelet[2984]: I0813 00:02:27.741220 2984 state_mem.go:88] "Updated default CPUSet" cpuSet="" Aug 13 00:02:27.741352 kubelet[2984]: I0813 00:02:27.741255 2984 state_mem.go:96] "Updated CPUSet assignments" assignments={} Aug 13 00:02:27.741352 kubelet[2984]: I0813 00:02:27.741291 2984 policy_none.go:49] "None policy: Start" Aug 13 00:02:27.742851 kubelet[2984]: I0813 00:02:27.742761 2984 memory_manager.go:170] "Starting memorymanager" policy="None" Aug 13 00:02:27.743000 kubelet[2984]: I0813 00:02:27.742962 2984 state_mem.go:35] "Initializing new in-memory state store" Aug 13 00:02:27.743363 kubelet[2984]: I0813 00:02:27.743327 2984 state_mem.go:75] "Updated machine memory state" Aug 13 00:02:27.745000 audit[2984]: AVC avc: denied { mac_admin } for pid=2984 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:02:27.745000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" Aug 13 00:02:27.745000 audit[2984]: SYSCALL arch=c00000b7 syscall=5 success=no exit=-22 a0=4000ba0d20 a1=4000d19248 a2=4000ba0b70 a3=25 items=0 ppid=1 pid=2984 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:27.745000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 Aug 13 00:02:27.748082 kubelet[2984]: I0813 00:02:27.746004 2984 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Aug 13 00:02:27.748082 kubelet[2984]: I0813 00:02:27.746197 2984 server.go:88] "Unprivileged containerized plugins might not work. Could not set selinux context on socket dir" path="/var/lib/kubelet/device-plugins/" err="setxattr /var/lib/kubelet/device-plugins/: invalid argument" Aug 13 00:02:27.748082 kubelet[2984]: I0813 00:02:27.746429 2984 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 00:02:27.748082 kubelet[2984]: I0813 00:02:27.746447 2984 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 00:02:27.756550 kubelet[2984]: I0813 00:02:27.756326 2984 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 00:02:27.800212 kubelet[2984]: E0813 00:02:27.799841 2984 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ip-172-31-27-151\" already exists" pod="kube-system/kube-scheduler-ip-172-31-27-151" Aug 13 00:02:27.866269 kubelet[2984]: I0813 00:02:27.866227 2984 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-27-151" Aug 13 00:02:27.871132 kubelet[2984]: I0813 00:02:27.871068 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5e3ffd7fdb1d7770aed8e1a0ec687436-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-27-151\" (UID: \"5e3ffd7fdb1d7770aed8e1a0ec687436\") " pod="kube-system/kube-controller-manager-ip-172-31-27-151" Aug 13 00:02:27.871396 kubelet[2984]: I0813 00:02:27.871361 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5e3ffd7fdb1d7770aed8e1a0ec687436-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-27-151\" (UID: \"5e3ffd7fdb1d7770aed8e1a0ec687436\") " pod="kube-system/kube-controller-manager-ip-172-31-27-151" Aug 13 00:02:27.871583 kubelet[2984]: I0813 00:02:27.871552 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e8468136087d101cffab1e36b9bf9565-kubeconfig\") pod \"kube-scheduler-ip-172-31-27-151\" (UID: \"e8468136087d101cffab1e36b9bf9565\") " pod="kube-system/kube-scheduler-ip-172-31-27-151" Aug 13 00:02:27.871736 kubelet[2984]: I0813 00:02:27.871710 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d246503ef314f9f2381c6bdef310f4af-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-27-151\" (UID: \"d246503ef314f9f2381c6bdef310f4af\") " pod="kube-system/kube-apiserver-ip-172-31-27-151" Aug 13 00:02:27.871942 kubelet[2984]: I0813 00:02:27.871916 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d246503ef314f9f2381c6bdef310f4af-k8s-certs\") pod \"kube-apiserver-ip-172-31-27-151\" (UID: \"d246503ef314f9f2381c6bdef310f4af\") " pod="kube-system/kube-apiserver-ip-172-31-27-151" Aug 13 00:02:27.872198 kubelet[2984]: I0813 00:02:27.872171 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5e3ffd7fdb1d7770aed8e1a0ec687436-ca-certs\") pod \"kube-controller-manager-ip-172-31-27-151\" (UID: \"5e3ffd7fdb1d7770aed8e1a0ec687436\") " pod="kube-system/kube-controller-manager-ip-172-31-27-151" Aug 13 00:02:27.872357 kubelet[2984]: I0813 00:02:27.872332 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5e3ffd7fdb1d7770aed8e1a0ec687436-k8s-certs\") pod \"kube-controller-manager-ip-172-31-27-151\" (UID: \"5e3ffd7fdb1d7770aed8e1a0ec687436\") " pod="kube-system/kube-controller-manager-ip-172-31-27-151" Aug 13 00:02:27.872504 kubelet[2984]: I0813 00:02:27.872479 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5e3ffd7fdb1d7770aed8e1a0ec687436-kubeconfig\") pod \"kube-controller-manager-ip-172-31-27-151\" (UID: \"5e3ffd7fdb1d7770aed8e1a0ec687436\") " pod="kube-system/kube-controller-manager-ip-172-31-27-151" Aug 13 00:02:27.872751 kubelet[2984]: I0813 00:02:27.872720 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d246503ef314f9f2381c6bdef310f4af-ca-certs\") pod \"kube-apiserver-ip-172-31-27-151\" (UID: \"d246503ef314f9f2381c6bdef310f4af\") " pod="kube-system/kube-apiserver-ip-172-31-27-151" Aug 13 00:02:27.879366 kubelet[2984]: I0813 00:02:27.879294 2984 kubelet_node_status.go:111] "Node was previously registered" node="ip-172-31-27-151" Aug 13 00:02:27.879517 kubelet[2984]: I0813 00:02:27.879463 2984 kubelet_node_status.go:75] "Successfully registered node" node="ip-172-31-27-151" Aug 13 00:02:28.421844 kubelet[2984]: I0813 00:02:28.421797 2984 apiserver.go:52] "Watching apiserver" Aug 13 00:02:28.466280 kubelet[2984]: I0813 00:02:28.466232 2984 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Aug 13 00:02:28.766522 kubelet[2984]: I0813 00:02:28.766354 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-27-151" podStartSLOduration=1.766334278 podStartE2EDuration="1.766334278s" podCreationTimestamp="2025-08-13 00:02:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:02:28.76492691 +0000 UTC m=+1.512105842" watchObservedRunningTime="2025-08-13 00:02:28.766334278 +0000 UTC m=+1.513513198" Aug 13 00:02:28.811255 kubelet[2984]: I0813 00:02:28.811180 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-27-151" podStartSLOduration=3.811154708 podStartE2EDuration="3.811154708s" podCreationTimestamp="2025-08-13 00:02:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:02:28.796031771 +0000 UTC m=+1.543210715" watchObservedRunningTime="2025-08-13 00:02:28.811154708 +0000 UTC m=+1.558333640" Aug 13 00:02:28.832381 kubelet[2984]: I0813 00:02:28.832265 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-27-151" podStartSLOduration=1.832241637 podStartE2EDuration="1.832241637s" podCreationTimestamp="2025-08-13 00:02:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:02:28.811546589 +0000 UTC m=+1.558725509" watchObservedRunningTime="2025-08-13 00:02:28.832241637 +0000 UTC m=+1.579420545" Aug 13 00:02:32.881403 kubelet[2984]: I0813 00:02:32.881356 2984 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Aug 13 00:02:32.882529 env[1913]: time="2025-08-13T00:02:32.882480983Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Aug 13 00:02:32.883654 kubelet[2984]: I0813 00:02:32.883608 2984 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Aug 13 00:02:33.217933 kubelet[2984]: I0813 00:02:33.217858 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvhfk\" (UniqueName: \"kubernetes.io/projected/492e2f65-355d-47a0-bfc0-1f2573490bad-kube-api-access-cvhfk\") pod \"kube-proxy-g2wxc\" (UID: \"492e2f65-355d-47a0-bfc0-1f2573490bad\") " pod="kube-system/kube-proxy-g2wxc" Aug 13 00:02:33.218209 kubelet[2984]: I0813 00:02:33.217962 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/492e2f65-355d-47a0-bfc0-1f2573490bad-xtables-lock\") pod \"kube-proxy-g2wxc\" (UID: \"492e2f65-355d-47a0-bfc0-1f2573490bad\") " pod="kube-system/kube-proxy-g2wxc" Aug 13 00:02:33.218209 kubelet[2984]: I0813 00:02:33.218058 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/492e2f65-355d-47a0-bfc0-1f2573490bad-lib-modules\") pod \"kube-proxy-g2wxc\" (UID: \"492e2f65-355d-47a0-bfc0-1f2573490bad\") " pod="kube-system/kube-proxy-g2wxc" Aug 13 00:02:33.218209 kubelet[2984]: I0813 00:02:33.218161 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/492e2f65-355d-47a0-bfc0-1f2573490bad-kube-proxy\") pod \"kube-proxy-g2wxc\" (UID: \"492e2f65-355d-47a0-bfc0-1f2573490bad\") " pod="kube-system/kube-proxy-g2wxc" Aug 13 00:02:33.333855 kubelet[2984]: E0813 00:02:33.333811 2984 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Aug 13 00:02:33.334069 kubelet[2984]: E0813 00:02:33.334045 2984 projected.go:194] Error preparing data for projected volume kube-api-access-cvhfk for pod kube-system/kube-proxy-g2wxc: configmap "kube-root-ca.crt" not found Aug 13 00:02:33.334333 kubelet[2984]: E0813 00:02:33.334311 2984 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/492e2f65-355d-47a0-bfc0-1f2573490bad-kube-api-access-cvhfk podName:492e2f65-355d-47a0-bfc0-1f2573490bad nodeName:}" failed. No retries permitted until 2025-08-13 00:02:33.834279406 +0000 UTC m=+6.581458314 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cvhfk" (UniqueName: "kubernetes.io/projected/492e2f65-355d-47a0-bfc0-1f2573490bad-kube-api-access-cvhfk") pod "kube-proxy-g2wxc" (UID: "492e2f65-355d-47a0-bfc0-1f2573490bad") : configmap "kube-root-ca.crt" not found Aug 13 00:02:33.932140 kubelet[2984]: I0813 00:02:33.931088 2984 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Aug 13 00:02:34.031388 kubelet[2984]: I0813 00:02:34.031345 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gp7wj\" (UniqueName: \"kubernetes.io/projected/3bbe2390-6aa1-4e54-89d1-ba802bea3845-kube-api-access-gp7wj\") pod \"tigera-operator-5bf8dfcb4-6q4vb\" (UID: \"3bbe2390-6aa1-4e54-89d1-ba802bea3845\") " pod="tigera-operator/tigera-operator-5bf8dfcb4-6q4vb" Aug 13 00:02:34.031687 kubelet[2984]: I0813 00:02:34.031635 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/3bbe2390-6aa1-4e54-89d1-ba802bea3845-var-lib-calico\") pod \"tigera-operator-5bf8dfcb4-6q4vb\" (UID: \"3bbe2390-6aa1-4e54-89d1-ba802bea3845\") " pod="tigera-operator/tigera-operator-5bf8dfcb4-6q4vb" Aug 13 00:02:34.083040 env[1913]: time="2025-08-13T00:02:34.082965506Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-g2wxc,Uid:492e2f65-355d-47a0-bfc0-1f2573490bad,Namespace:kube-system,Attempt:0,}" Aug 13 00:02:34.122927 env[1913]: time="2025-08-13T00:02:34.122422709Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:02:34.122927 env[1913]: time="2025-08-13T00:02:34.122637120Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:02:34.122927 env[1913]: time="2025-08-13T00:02:34.122663507Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:02:34.123214 env[1913]: time="2025-08-13T00:02:34.123036350Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/422cc4dc468812e5e305ddabebf38815c4f755258a566a00d843fdc6d8eaf132 pid=3036 runtime=io.containerd.runc.v2 Aug 13 00:02:34.242183 env[1913]: time="2025-08-13T00:02:34.241914417Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-g2wxc,Uid:492e2f65-355d-47a0-bfc0-1f2573490bad,Namespace:kube-system,Attempt:0,} returns sandbox id \"422cc4dc468812e5e305ddabebf38815c4f755258a566a00d843fdc6d8eaf132\"" Aug 13 00:02:34.251643 env[1913]: time="2025-08-13T00:02:34.251534392Z" level=info msg="CreateContainer within sandbox \"422cc4dc468812e5e305ddabebf38815c4f755258a566a00d843fdc6d8eaf132\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Aug 13 00:02:34.282342 env[1913]: time="2025-08-13T00:02:34.282266870Z" level=info msg="CreateContainer within sandbox \"422cc4dc468812e5e305ddabebf38815c4f755258a566a00d843fdc6d8eaf132\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"2d906f1915e48e88b58120cbebf8857a9513d23c2f964f9b96f9c7ce83ea7c27\"" Aug 13 00:02:34.285624 env[1913]: time="2025-08-13T00:02:34.285176243Z" level=info msg="StartContainer for \"2d906f1915e48e88b58120cbebf8857a9513d23c2f964f9b96f9c7ce83ea7c27\"" Aug 13 00:02:34.286981 env[1913]: time="2025-08-13T00:02:34.286878436Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-5bf8dfcb4-6q4vb,Uid:3bbe2390-6aa1-4e54-89d1-ba802bea3845,Namespace:tigera-operator,Attempt:0,}" Aug 13 00:02:34.332675 env[1913]: time="2025-08-13T00:02:34.332441527Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:02:34.332854 env[1913]: time="2025-08-13T00:02:34.332731356Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:02:34.332854 env[1913]: time="2025-08-13T00:02:34.332822062Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:02:34.335136 env[1913]: time="2025-08-13T00:02:34.333756526Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/046e44291ace527a4c6139cecc0cda76979228ae7a44565e87bc672c8ee2bd9f pid=3096 runtime=io.containerd.runc.v2 Aug 13 00:02:34.439278 env[1913]: time="2025-08-13T00:02:34.439215624Z" level=info msg="StartContainer for \"2d906f1915e48e88b58120cbebf8857a9513d23c2f964f9b96f9c7ce83ea7c27\" returns successfully" Aug 13 00:02:34.484819 env[1913]: time="2025-08-13T00:02:34.484764591Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-5bf8dfcb4-6q4vb,Uid:3bbe2390-6aa1-4e54-89d1-ba802bea3845,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"046e44291ace527a4c6139cecc0cda76979228ae7a44565e87bc672c8ee2bd9f\"" Aug 13 00:02:34.494775 env[1913]: time="2025-08-13T00:02:34.494627068Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\"" Aug 13 00:02:34.742749 kernel: kauditd_printk_skb: 4 callbacks suppressed Aug 13 00:02:34.742870 kernel: audit: type=1325 audit(1755043354.733:225): table=mangle:38 family=2 entries=1 op=nft_register_chain pid=3182 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.733000 audit[3182]: NETFILTER_CFG table=mangle:38 family=2 entries=1 op=nft_register_chain pid=3182 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.733000 audit[3182]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffe0880c70 a2=0 a3=1 items=0 ppid=3113 pid=3182 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.759606 kernel: audit: type=1300 audit(1755043354.733:225): arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffe0880c70 a2=0 a3=1 items=0 ppid=3113 pid=3182 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.733000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Aug 13 00:02:34.768180 kernel: audit: type=1327 audit(1755043354.733:225): proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Aug 13 00:02:34.769935 kernel: audit: type=1325 audit(1755043354.733:226): table=nat:39 family=2 entries=1 op=nft_register_chain pid=3183 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.733000 audit[3183]: NETFILTER_CFG table=nat:39 family=2 entries=1 op=nft_register_chain pid=3183 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.733000 audit[3183]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=fffffc86d1f0 a2=0 a3=1 items=0 ppid=3113 pid=3183 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.789259 kernel: audit: type=1300 audit(1755043354.733:226): arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=fffffc86d1f0 a2=0 a3=1 items=0 ppid=3113 pid=3183 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.733000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Aug 13 00:02:34.795981 kernel: audit: type=1327 audit(1755043354.733:226): proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Aug 13 00:02:34.796236 kernel: audit: type=1325 audit(1755043354.735:227): table=filter:40 family=2 entries=1 op=nft_register_chain pid=3184 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.735000 audit[3184]: NETFILTER_CFG table=filter:40 family=2 entries=1 op=nft_register_chain pid=3184 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.735000 audit[3184]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffd170ec40 a2=0 a3=1 items=0 ppid=3113 pid=3184 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.817420 kernel: audit: type=1300 audit(1755043354.735:227): arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffd170ec40 a2=0 a3=1 items=0 ppid=3113 pid=3184 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.735000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Aug 13 00:02:34.823596 kernel: audit: type=1327 audit(1755043354.735:227): proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Aug 13 00:02:34.735000 audit[3185]: NETFILTER_CFG table=mangle:41 family=10 entries=1 op=nft_register_chain pid=3185 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:34.829738 kernel: audit: type=1325 audit(1755043354.735:228): table=mangle:41 family=10 entries=1 op=nft_register_chain pid=3185 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:34.735000 audit[3185]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffeeac6b70 a2=0 a3=1 items=0 ppid=3113 pid=3185 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.735000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 Aug 13 00:02:34.735000 audit[3186]: NETFILTER_CFG table=nat:42 family=10 entries=1 op=nft_register_chain pid=3186 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:34.735000 audit[3186]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffd935ee90 a2=0 a3=1 items=0 ppid=3113 pid=3186 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.735000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 Aug 13 00:02:34.743000 audit[3187]: NETFILTER_CFG table=filter:43 family=10 entries=1 op=nft_register_chain pid=3187 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:34.743000 audit[3187]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffdcac2150 a2=0 a3=1 items=0 ppid=3113 pid=3187 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.743000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 Aug 13 00:02:34.849000 audit[3188]: NETFILTER_CFG table=filter:44 family=2 entries=1 op=nft_register_chain pid=3188 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.849000 audit[3188]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=108 a0=3 a1=ffffd007b6f0 a2=0 a3=1 items=0 ppid=3113 pid=3188 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.849000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 Aug 13 00:02:34.856000 audit[3190]: NETFILTER_CFG table=filter:45 family=2 entries=1 op=nft_register_rule pid=3190 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.856000 audit[3190]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=752 a0=3 a1=ffffe0692c00 a2=0 a3=1 items=0 ppid=3113 pid=3190 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.856000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C652073657276696365 Aug 13 00:02:34.864000 audit[3193]: NETFILTER_CFG table=filter:46 family=2 entries=1 op=nft_register_rule pid=3193 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.864000 audit[3193]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=752 a0=3 a1=ffffd976a500 a2=0 a3=1 items=0 ppid=3113 pid=3193 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.864000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669 Aug 13 00:02:34.867000 audit[3194]: NETFILTER_CFG table=filter:47 family=2 entries=1 op=nft_register_chain pid=3194 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.867000 audit[3194]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffd51f2770 a2=0 a3=1 items=0 ppid=3113 pid=3194 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.867000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4E4F4445504F525453002D740066696C746572 Aug 13 00:02:34.873000 audit[3196]: NETFILTER_CFG table=filter:48 family=2 entries=1 op=nft_register_rule pid=3196 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.873000 audit[3196]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=528 a0=3 a1=ffffc2cfcb00 a2=0 a3=1 items=0 ppid=3113 pid=3196 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.873000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 Aug 13 00:02:34.876000 audit[3197]: NETFILTER_CFG table=filter:49 family=2 entries=1 op=nft_register_chain pid=3197 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.876000 audit[3197]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffc7088250 a2=0 a3=1 items=0 ppid=3113 pid=3197 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.876000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D740066696C746572 Aug 13 00:02:34.885000 audit[3199]: NETFILTER_CFG table=filter:50 family=2 entries=1 op=nft_register_rule pid=3199 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.885000 audit[3199]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=744 a0=3 a1=ffffe88a33a0 a2=0 a3=1 items=0 ppid=3113 pid=3199 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.885000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D Aug 13 00:02:34.895000 audit[3202]: NETFILTER_CFG table=filter:51 family=2 entries=1 op=nft_register_rule pid=3202 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.895000 audit[3202]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=744 a0=3 a1=ffffd0950930 a2=0 a3=1 items=0 ppid=3113 pid=3202 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.895000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D53 Aug 13 00:02:34.898000 audit[3203]: NETFILTER_CFG table=filter:52 family=2 entries=1 op=nft_register_chain pid=3203 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.898000 audit[3203]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffc102cbf0 a2=0 a3=1 items=0 ppid=3113 pid=3203 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.898000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D464F5257415244002D740066696C746572 Aug 13 00:02:34.904000 audit[3205]: NETFILTER_CFG table=filter:53 family=2 entries=1 op=nft_register_rule pid=3205 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.904000 audit[3205]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=528 a0=3 a1=ffffee9d9520 a2=0 a3=1 items=0 ppid=3113 pid=3205 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.904000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 Aug 13 00:02:34.907000 audit[3206]: NETFILTER_CFG table=filter:54 family=2 entries=1 op=nft_register_chain pid=3206 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.907000 audit[3206]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=ffffc624e5f0 a2=0 a3=1 items=0 ppid=3113 pid=3206 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.907000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 Aug 13 00:02:34.913000 audit[3208]: NETFILTER_CFG table=filter:55 family=2 entries=1 op=nft_register_rule pid=3208 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.913000 audit[3208]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=748 a0=3 a1=fffff0dc8bd0 a2=0 a3=1 items=0 ppid=3113 pid=3208 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.913000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Aug 13 00:02:34.920000 audit[3211]: NETFILTER_CFG table=filter:56 family=2 entries=1 op=nft_register_rule pid=3211 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.920000 audit[3211]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=748 a0=3 a1=ffffd7e88360 a2=0 a3=1 items=0 ppid=3113 pid=3211 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.920000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Aug 13 00:02:34.929000 audit[3214]: NETFILTER_CFG table=filter:57 family=2 entries=1 op=nft_register_rule pid=3214 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.929000 audit[3214]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=748 a0=3 a1=ffffcfcb61a0 a2=0 a3=1 items=0 ppid=3113 pid=3214 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.929000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D Aug 13 00:02:34.935000 audit[3215]: NETFILTER_CFG table=nat:58 family=2 entries=1 op=nft_register_chain pid=3215 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.935000 audit[3215]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=96 a0=3 a1=fffff28e3370 a2=0 a3=1 items=0 ppid=3113 pid=3215 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.935000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D74006E6174 Aug 13 00:02:34.950000 audit[3217]: NETFILTER_CFG table=nat:59 family=2 entries=1 op=nft_register_rule pid=3217 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.950000 audit[3217]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=524 a0=3 a1=ffffe1cdd790 a2=0 a3=1 items=0 ppid=3113 pid=3217 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.950000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Aug 13 00:02:34.963000 audit[3220]: NETFILTER_CFG table=nat:60 family=2 entries=1 op=nft_register_rule pid=3220 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.963000 audit[3220]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=528 a0=3 a1=fffff81fbfa0 a2=0 a3=1 items=0 ppid=3113 pid=3220 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.963000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Aug 13 00:02:34.966000 audit[3221]: NETFILTER_CFG table=nat:61 family=2 entries=1 op=nft_register_chain pid=3221 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.966000 audit[3221]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffcd43ad40 a2=0 a3=1 items=0 ppid=3113 pid=3221 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.966000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Aug 13 00:02:34.971000 audit[3223]: NETFILTER_CFG table=nat:62 family=2 entries=1 op=nft_register_rule pid=3223 subj=system_u:system_r:kernel_t:s0 comm="iptables" Aug 13 00:02:34.971000 audit[3223]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=532 a0=3 a1=fffffd35b040 a2=0 a3=1 items=0 ppid=3113 pid=3223 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:34.971000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 Aug 13 00:02:35.020000 audit[3229]: NETFILTER_CFG table=filter:63 family=2 entries=8 op=nft_register_rule pid=3229 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:02:35.020000 audit[3229]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=5248 a0=3 a1=ffffdd490310 a2=0 a3=1 items=0 ppid=3113 pid=3229 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.020000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:02:35.033000 audit[3229]: NETFILTER_CFG table=nat:64 family=2 entries=14 op=nft_register_chain pid=3229 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:02:35.033000 audit[3229]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=5508 a0=3 a1=ffffdd490310 a2=0 a3=1 items=0 ppid=3113 pid=3229 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.033000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:02:35.035000 audit[3234]: NETFILTER_CFG table=filter:65 family=10 entries=1 op=nft_register_chain pid=3234 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.035000 audit[3234]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=108 a0=3 a1=ffffced07310 a2=0 a3=1 items=0 ppid=3113 pid=3234 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.035000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 Aug 13 00:02:35.041000 audit[3236]: NETFILTER_CFG table=filter:66 family=10 entries=2 op=nft_register_chain pid=3236 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.041000 audit[3236]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=836 a0=3 a1=ffffdedad2d0 a2=0 a3=1 items=0 ppid=3113 pid=3236 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.041000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C6520736572766963 Aug 13 00:02:35.049000 audit[3239]: NETFILTER_CFG table=filter:67 family=10 entries=2 op=nft_register_chain pid=3239 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.049000 audit[3239]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=836 a0=3 a1=ffffe3e3d760 a2=0 a3=1 items=0 ppid=3113 pid=3239 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.049000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C652073657276 Aug 13 00:02:35.052000 audit[3240]: NETFILTER_CFG table=filter:68 family=10 entries=1 op=nft_register_chain pid=3240 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.052000 audit[3240]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffd0f3fd80 a2=0 a3=1 items=0 ppid=3113 pid=3240 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.052000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4E4F4445504F525453002D740066696C746572 Aug 13 00:02:35.057000 audit[3242]: NETFILTER_CFG table=filter:69 family=10 entries=1 op=nft_register_rule pid=3242 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.057000 audit[3242]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=528 a0=3 a1=ffffd69cb7b0 a2=0 a3=1 items=0 ppid=3113 pid=3242 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.057000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 Aug 13 00:02:35.060000 audit[3243]: NETFILTER_CFG table=filter:70 family=10 entries=1 op=nft_register_chain pid=3243 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.060000 audit[3243]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffdbd3fb10 a2=0 a3=1 items=0 ppid=3113 pid=3243 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.060000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D740066696C746572 Aug 13 00:02:35.066000 audit[3245]: NETFILTER_CFG table=filter:71 family=10 entries=1 op=nft_register_rule pid=3245 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.066000 audit[3245]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=744 a0=3 a1=fffff63e0a10 a2=0 a3=1 items=0 ppid=3113 pid=3245 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.066000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B554245 Aug 13 00:02:35.079000 audit[3248]: NETFILTER_CFG table=filter:72 family=10 entries=2 op=nft_register_chain pid=3248 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.079000 audit[3248]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=828 a0=3 a1=ffffe6b4a280 a2=0 a3=1 items=0 ppid=3113 pid=3248 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.079000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D Aug 13 00:02:35.081000 audit[3249]: NETFILTER_CFG table=filter:73 family=10 entries=1 op=nft_register_chain pid=3249 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.081000 audit[3249]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffdaae78b0 a2=0 a3=1 items=0 ppid=3113 pid=3249 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.081000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D464F5257415244002D740066696C746572 Aug 13 00:02:35.093000 audit[3251]: NETFILTER_CFG table=filter:74 family=10 entries=1 op=nft_register_rule pid=3251 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.093000 audit[3251]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=528 a0=3 a1=ffffda7b7a30 a2=0 a3=1 items=0 ppid=3113 pid=3251 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.093000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 Aug 13 00:02:35.096000 audit[3252]: NETFILTER_CFG table=filter:75 family=10 entries=1 op=nft_register_chain pid=3252 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.096000 audit[3252]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=104 a0=3 a1=fffff1e6b730 a2=0 a3=1 items=0 ppid=3113 pid=3252 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.096000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 Aug 13 00:02:35.102000 audit[3254]: NETFILTER_CFG table=filter:76 family=10 entries=1 op=nft_register_rule pid=3254 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.102000 audit[3254]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=748 a0=3 a1=fffffcef2030 a2=0 a3=1 items=0 ppid=3113 pid=3254 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.102000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A Aug 13 00:02:35.110000 audit[3257]: NETFILTER_CFG table=filter:77 family=10 entries=1 op=nft_register_rule pid=3257 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.110000 audit[3257]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=748 a0=3 a1=ffffe4a89dd0 a2=0 a3=1 items=0 ppid=3113 pid=3257 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.110000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D Aug 13 00:02:35.119000 audit[3260]: NETFILTER_CFG table=filter:78 family=10 entries=1 op=nft_register_rule pid=3260 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.119000 audit[3260]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=748 a0=3 a1=fffff4f66c30 a2=0 a3=1 items=0 ppid=3113 pid=3260 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.119000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C Aug 13 00:02:35.121000 audit[3261]: NETFILTER_CFG table=nat:79 family=10 entries=1 op=nft_register_chain pid=3261 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.121000 audit[3261]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=96 a0=3 a1=ffffdb0a68e0 a2=0 a3=1 items=0 ppid=3113 pid=3261 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.121000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D74006E6174 Aug 13 00:02:35.126000 audit[3263]: NETFILTER_CFG table=nat:80 family=10 entries=2 op=nft_register_chain pid=3263 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.126000 audit[3263]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=600 a0=3 a1=fffffe2de1a0 a2=0 a3=1 items=0 ppid=3113 pid=3263 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.126000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Aug 13 00:02:35.133000 audit[3266]: NETFILTER_CFG table=nat:81 family=10 entries=2 op=nft_register_chain pid=3266 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.133000 audit[3266]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=608 a0=3 a1=fffff7cc0d10 a2=0 a3=1 items=0 ppid=3113 pid=3266 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.133000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 Aug 13 00:02:35.136000 audit[3267]: NETFILTER_CFG table=nat:82 family=10 entries=1 op=nft_register_chain pid=3267 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.136000 audit[3267]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffeb05c170 a2=0 a3=1 items=0 ppid=3113 pid=3267 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.136000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D504F5354524F5554494E47002D74006E6174 Aug 13 00:02:35.141000 audit[3269]: NETFILTER_CFG table=nat:83 family=10 entries=2 op=nft_register_chain pid=3269 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.141000 audit[3269]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=612 a0=3 a1=ffffee45e320 a2=0 a3=1 items=0 ppid=3113 pid=3269 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.141000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 Aug 13 00:02:35.144000 audit[3270]: NETFILTER_CFG table=filter:84 family=10 entries=1 op=nft_register_chain pid=3270 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.144000 audit[3270]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=100 a0=3 a1=ffffeb849410 a2=0 a3=1 items=0 ppid=3113 pid=3270 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.144000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 Aug 13 00:02:35.150000 audit[3272]: NETFILTER_CFG table=filter:85 family=10 entries=1 op=nft_register_rule pid=3272 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.150000 audit[3272]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=228 a0=3 a1=ffffeb8d1190 a2=0 a3=1 items=0 ppid=3113 pid=3272 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.150000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C Aug 13 00:02:35.157000 audit[3275]: NETFILTER_CFG table=filter:86 family=10 entries=1 op=nft_register_rule pid=3275 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" Aug 13 00:02:35.157000 audit[3275]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=228 a0=3 a1=ffffddb239d0 a2=0 a3=1 items=0 ppid=3113 pid=3275 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.157000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C Aug 13 00:02:35.163000 audit[3277]: NETFILTER_CFG table=filter:87 family=10 entries=3 op=nft_register_rule pid=3277 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" Aug 13 00:02:35.163000 audit[3277]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2088 a0=3 a1=ffffe78e4670 a2=0 a3=1 items=0 ppid=3113 pid=3277 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.163000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:02:35.164000 audit[3277]: NETFILTER_CFG table=nat:88 family=10 entries=7 op=nft_register_chain pid=3277 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" Aug 13 00:02:35.164000 audit[3277]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2056 a0=3 a1=ffffe78e4670 a2=0 a3=1 items=0 ppid=3113 pid=3277 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:02:35.164000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:02:35.491809 env[1913]: time="2025-08-13T00:02:35.491542295Z" level=error msg="PullImage \"quay.io/tigera/operator:v1.38.3\" failed" error="failed to pull and unpack image \"quay.io/tigera/operator:v1.38.3\": failed to copy: httpReadSeeker: failed open: failed to do request: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/7f/7f8a5b1dba618e907d5f7804e42b3bd7cd5766bc3b0a66da25ff2c687e356bb0?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTCKFFWO5C%2F20250813%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20250813T000235Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=0edd7ea5cce13c68b0ccfaf4b2ab701973f78644f704884c99c15118d6be062f®ion=us-east-1&namespace=tigera&repo_name=operator&akamai_signature=exp=1755044255~hmac=4ced7d2002c0dcc0fd33772ac99b3d1403a8f46a0a51daf693f087f674400761\": dial tcp: lookup cdn01.quay.io: no such host" Aug 13 00:02:35.492975 kubelet[2984]: E0813 00:02:35.492820 2984 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = failed to pull and unpack image \"quay.io/tigera/operator:v1.38.3\": failed to copy: httpReadSeeker: failed open: failed to do request: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/7f/7f8a5b1dba618e907d5f7804e42b3bd7cd5766bc3b0a66da25ff2c687e356bb0?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTCKFFWO5C%2F20250813%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20250813T000235Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=0edd7ea5cce13c68b0ccfaf4b2ab701973f78644f704884c99c15118d6be062f®ion=us-east-1&namespace=tigera&repo_name=operator&akamai_signature=exp=1755044255~hmac=4ced7d2002c0dcc0fd33772ac99b3d1403a8f46a0a51daf693f087f674400761\": dial tcp: lookup cdn01.quay.io: no such host" image="quay.io/tigera/operator:v1.38.3" Aug 13 00:02:35.493878 kubelet[2984]: E0813 00:02:35.493018 2984 kuberuntime_image.go:55] "Failed to pull image" err="failed to pull and unpack image \"quay.io/tigera/operator:v1.38.3\": failed to copy: httpReadSeeker: failed open: failed to do request: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/7f/7f8a5b1dba618e907d5f7804e42b3bd7cd5766bc3b0a66da25ff2c687e356bb0?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTCKFFWO5C%2F20250813%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20250813T000235Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=0edd7ea5cce13c68b0ccfaf4b2ab701973f78644f704884c99c15118d6be062f®ion=us-east-1&namespace=tigera&repo_name=operator&akamai_signature=exp=1755044255~hmac=4ced7d2002c0dcc0fd33772ac99b3d1403a8f46a0a51daf693f087f674400761\": dial tcp: lookup cdn01.quay.io: no such host" image="quay.io/tigera/operator:v1.38.3" Aug 13 00:02:35.494055 kubelet[2984]: E0813 00:02:35.493464 2984 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tigera-operator,Image:quay.io/tigera/operator:v1.38.3,Command:[operator],Args:[-manage-crds=true],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:tigera-operator,ValueFrom:nil,},EnvVar{Name:TIGERA_OPERATOR_INIT_IMAGE_VERSION,Value:v1.38.3,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-lib-calico,ReadOnly:true,MountPath:/var/lib/calico,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gp7wj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:kubernetes-services-endpoint,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tigera-operator-5bf8dfcb4-6q4vb_tigera-operator(3bbe2390-6aa1-4e54-89d1-ba802bea3845): ErrImagePull: failed to pull and unpack image \"quay.io/tigera/operator:v1.38.3\": failed to copy: httpReadSeeker: failed open: failed to do request: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/7f/7f8a5b1dba618e907d5f7804e42b3bd7cd5766bc3b0a66da25ff2c687e356bb0?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTCKFFWO5C%2F20250813%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20250813T000235Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=0edd7ea5cce13c68b0ccfaf4b2ab701973f78644f704884c99c15118d6be062f®ion=us-east-1&namespace=tigera&repo_name=operator&akamai_signature=exp=1755044255~hmac=4ced7d2002c0dcc0fd33772ac99b3d1403a8f46a0a51daf693f087f674400761\": dial tcp: lookup cdn01.quay.io: no such host" logger="UnhandledError" Aug 13 00:02:35.496568 kubelet[2984]: E0813 00:02:35.496440 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tigera-operator\" with ErrImagePull: \"failed to pull and unpack image \\\"quay.io/tigera/operator:v1.38.3\\\": failed to copy: httpReadSeeker: failed open: failed to do request: Get \\\"https://cdn01.quay.io/quayio-production-s3/sha256/7f/7f8a5b1dba618e907d5f7804e42b3bd7cd5766bc3b0a66da25ff2c687e356bb0?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTCKFFWO5C%2F20250813%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20250813T000235Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=0edd7ea5cce13c68b0ccfaf4b2ab701973f78644f704884c99c15118d6be062f®ion=us-east-1&namespace=tigera&repo_name=operator&akamai_signature=exp=1755044255~hmac=4ced7d2002c0dcc0fd33772ac99b3d1403a8f46a0a51daf693f087f674400761\\\": dial tcp: lookup cdn01.quay.io: no such host\"" pod="tigera-operator/tigera-operator-5bf8dfcb4-6q4vb" podUID="3bbe2390-6aa1-4e54-89d1-ba802bea3845" Aug 13 00:02:35.684460 kubelet[2984]: E0813 00:02:35.683983 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tigera-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/tigera/operator:v1.38.3\\\"\"" pod="tigera-operator/tigera-operator-5bf8dfcb4-6q4vb" podUID="3bbe2390-6aa1-4e54-89d1-ba802bea3845" Aug 13 00:02:35.703638 kubelet[2984]: I0813 00:02:35.703528 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-g2wxc" podStartSLOduration=2.70347647 podStartE2EDuration="2.70347647s" podCreationTimestamp="2025-08-13 00:02:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:02:34.701724897 +0000 UTC m=+7.448903841" watchObservedRunningTime="2025-08-13 00:02:35.70347647 +0000 UTC m=+8.450655378" Aug 13 00:02:48.479152 env[1913]: time="2025-08-13T00:02:48.478990953Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\"" Aug 13 00:02:49.738780 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount913714293.mount: Deactivated successfully. Aug 13 00:02:50.837159 env[1913]: time="2025-08-13T00:02:50.837035530Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/tigera/operator:v1.38.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:50.841651 env[1913]: time="2025-08-13T00:02:50.841568885Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:7f8a5b1dba618e907d5f7804e42b3bd7cd5766bc3b0a66da25ff2c687e356bb0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:50.845560 env[1913]: time="2025-08-13T00:02:50.845489484Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/tigera/operator:v1.38.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:50.849911 env[1913]: time="2025-08-13T00:02:50.849859155Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:02:50.853170 env[1913]: time="2025-08-13T00:02:50.851871092Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\" returns image reference \"sha256:7f8a5b1dba618e907d5f7804e42b3bd7cd5766bc3b0a66da25ff2c687e356bb0\"" Aug 13 00:02:50.859789 env[1913]: time="2025-08-13T00:02:50.859732534Z" level=info msg="CreateContainer within sandbox \"046e44291ace527a4c6139cecc0cda76979228ae7a44565e87bc672c8ee2bd9f\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Aug 13 00:02:50.894834 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2891740310.mount: Deactivated successfully. Aug 13 00:02:50.908164 env[1913]: time="2025-08-13T00:02:50.908065833Z" level=info msg="CreateContainer within sandbox \"046e44291ace527a4c6139cecc0cda76979228ae7a44565e87bc672c8ee2bd9f\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"8df57597aa0a4b83cbc5e0108269881d4b8fd0461af10d1c21edbd6c7e0f625f\"" Aug 13 00:02:50.911369 env[1913]: time="2025-08-13T00:02:50.911132303Z" level=info msg="StartContainer for \"8df57597aa0a4b83cbc5e0108269881d4b8fd0461af10d1c21edbd6c7e0f625f\"" Aug 13 00:02:51.045784 env[1913]: time="2025-08-13T00:02:51.045720192Z" level=info msg="StartContainer for \"8df57597aa0a4b83cbc5e0108269881d4b8fd0461af10d1c21edbd6c7e0f625f\" returns successfully" Aug 13 00:02:58.089736 sudo[2223]: pam_unix(sudo:session): session closed for user root Aug 13 00:02:58.089000 audit[2223]: USER_END pid=2223 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Aug 13 00:02:58.091894 kernel: kauditd_printk_skb: 143 callbacks suppressed Aug 13 00:02:58.092043 kernel: audit: type=1106 audit(1755043378.089:276): pid=2223 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Aug 13 00:02:58.101000 audit[2223]: CRED_DISP pid=2223 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Aug 13 00:02:58.110003 kernel: audit: type=1104 audit(1755043378.101:277): pid=2223 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' Aug 13 00:02:58.135416 sshd[2219]: pam_unix(sshd:session): session closed for user core Aug 13 00:02:58.137000 audit[2219]: USER_END pid=2219 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:02:58.152065 systemd[1]: sshd@6-172.31.27.151:22-139.178.89.65:42846.service: Deactivated successfully. Aug 13 00:02:58.154006 systemd[1]: session-7.scope: Deactivated successfully. Aug 13 00:02:58.154201 systemd-logind[1901]: Session 7 logged out. Waiting for processes to exit. Aug 13 00:02:58.156240 systemd-logind[1901]: Removed session 7. Aug 13 00:02:58.137000 audit[2219]: CRED_DISP pid=2219 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:02:58.176329 kernel: audit: type=1106 audit(1755043378.137:278): pid=2219 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:02:58.176513 kernel: audit: type=1104 audit(1755043378.137:279): pid=2219 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:02:58.152000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-172.31.27.151:22-139.178.89.65:42846 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:02:58.186458 kernel: audit: type=1131 audit(1755043378.152:280): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-172.31.27.151:22-139.178.89.65:42846 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:03:02.260000 audit[3357]: NETFILTER_CFG table=filter:89 family=2 entries=15 op=nft_register_rule pid=3357 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:02.260000 audit[3357]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=5992 a0=3 a1=ffffedf7ecd0 a2=0 a3=1 items=0 ppid=3113 pid=3357 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:02.281931 kernel: audit: type=1325 audit(1755043382.260:281): table=filter:89 family=2 entries=15 op=nft_register_rule pid=3357 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:02.282155 kernel: audit: type=1300 audit(1755043382.260:281): arch=c00000b7 syscall=211 success=yes exit=5992 a0=3 a1=ffffedf7ecd0 a2=0 a3=1 items=0 ppid=3113 pid=3357 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:02.260000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:02.288341 kernel: audit: type=1327 audit(1755043382.260:281): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:02.275000 audit[3357]: NETFILTER_CFG table=nat:90 family=2 entries=12 op=nft_register_rule pid=3357 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:02.294486 kernel: audit: type=1325 audit(1755043382.275:282): table=nat:90 family=2 entries=12 op=nft_register_rule pid=3357 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:02.275000 audit[3357]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2700 a0=3 a1=ffffedf7ecd0 a2=0 a3=1 items=0 ppid=3113 pid=3357 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:02.309309 kernel: audit: type=1300 audit(1755043382.275:282): arch=c00000b7 syscall=211 success=yes exit=2700 a0=3 a1=ffffedf7ecd0 a2=0 a3=1 items=0 ppid=3113 pid=3357 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:02.275000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:02.345000 audit[3359]: NETFILTER_CFG table=filter:91 family=2 entries=16 op=nft_register_rule pid=3359 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:02.345000 audit[3359]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=5992 a0=3 a1=ffffed30e2f0 a2=0 a3=1 items=0 ppid=3113 pid=3359 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:02.345000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:02.351000 audit[3359]: NETFILTER_CFG table=nat:92 family=2 entries=12 op=nft_register_rule pid=3359 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:02.351000 audit[3359]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2700 a0=3 a1=ffffed30e2f0 a2=0 a3=1 items=0 ppid=3113 pid=3359 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:02.351000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:09.775000 audit[3363]: NETFILTER_CFG table=filter:93 family=2 entries=17 op=nft_register_rule pid=3363 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:09.780386 kernel: kauditd_printk_skb: 7 callbacks suppressed Aug 13 00:03:09.780632 kernel: audit: type=1325 audit(1755043389.775:285): table=filter:93 family=2 entries=17 op=nft_register_rule pid=3363 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:09.775000 audit[3363]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=6736 a0=3 a1=ffffd1e68770 a2=0 a3=1 items=0 ppid=3113 pid=3363 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:09.775000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:09.830190 kernel: audit: type=1300 audit(1755043389.775:285): arch=c00000b7 syscall=211 success=yes exit=6736 a0=3 a1=ffffd1e68770 a2=0 a3=1 items=0 ppid=3113 pid=3363 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:09.830355 kernel: audit: type=1327 audit(1755043389.775:285): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:09.832000 audit[3363]: NETFILTER_CFG table=nat:94 family=2 entries=12 op=nft_register_rule pid=3363 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:09.839938 kernel: audit: type=1325 audit(1755043389.832:286): table=nat:94 family=2 entries=12 op=nft_register_rule pid=3363 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:09.832000 audit[3363]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2700 a0=3 a1=ffffd1e68770 a2=0 a3=1 items=0 ppid=3113 pid=3363 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:09.859329 kernel: audit: type=1300 audit(1755043389.832:286): arch=c00000b7 syscall=211 success=yes exit=2700 a0=3 a1=ffffd1e68770 a2=0 a3=1 items=0 ppid=3113 pid=3363 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:09.832000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:09.875650 kernel: audit: type=1327 audit(1755043389.832:286): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:09.907000 audit[3365]: NETFILTER_CFG table=filter:95 family=2 entries=18 op=nft_register_rule pid=3365 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:09.907000 audit[3365]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=6736 a0=3 a1=ffffdfc63550 a2=0 a3=1 items=0 ppid=3113 pid=3365 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:09.928234 kernel: audit: type=1325 audit(1755043389.907:287): table=filter:95 family=2 entries=18 op=nft_register_rule pid=3365 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:09.928386 kernel: audit: type=1300 audit(1755043389.907:287): arch=c00000b7 syscall=211 success=yes exit=6736 a0=3 a1=ffffdfc63550 a2=0 a3=1 items=0 ppid=3113 pid=3365 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:09.907000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:09.946146 kernel: audit: type=1327 audit(1755043389.907:287): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:09.961000 audit[3365]: NETFILTER_CFG table=nat:96 family=2 entries=12 op=nft_register_rule pid=3365 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:09.969142 kernel: audit: type=1325 audit(1755043389.961:288): table=nat:96 family=2 entries=12 op=nft_register_rule pid=3365 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:09.961000 audit[3365]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2700 a0=3 a1=ffffdfc63550 a2=0 a3=1 items=0 ppid=3113 pid=3365 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:09.961000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:10.096816 kubelet[2984]: I0813 00:03:10.096494 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-5bf8dfcb4-6q4vb" podStartSLOduration=20.728360826 podStartE2EDuration="37.096448843s" podCreationTimestamp="2025-08-13 00:02:33 +0000 UTC" firstStartedPulling="2025-08-13 00:02:34.486976531 +0000 UTC m=+7.234155439" lastFinishedPulling="2025-08-13 00:02:50.855064548 +0000 UTC m=+23.602243456" observedRunningTime="2025-08-13 00:02:51.743342248 +0000 UTC m=+24.490521168" watchObservedRunningTime="2025-08-13 00:03:10.096448843 +0000 UTC m=+42.843627739" Aug 13 00:03:10.194721 kubelet[2984]: I0813 00:03:10.194621 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb4cd17c-d008-48e1-bdd8-356e96855153-tigera-ca-bundle\") pod \"calico-typha-568ddcdc8c-wkc6k\" (UID: \"eb4cd17c-d008-48e1-bdd8-356e96855153\") " pod="calico-system/calico-typha-568ddcdc8c-wkc6k" Aug 13 00:03:10.194889 kubelet[2984]: I0813 00:03:10.194729 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/eb4cd17c-d008-48e1-bdd8-356e96855153-typha-certs\") pod \"calico-typha-568ddcdc8c-wkc6k\" (UID: \"eb4cd17c-d008-48e1-bdd8-356e96855153\") " pod="calico-system/calico-typha-568ddcdc8c-wkc6k" Aug 13 00:03:10.194889 kubelet[2984]: I0813 00:03:10.194801 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z64kg\" (UniqueName: \"kubernetes.io/projected/eb4cd17c-d008-48e1-bdd8-356e96855153-kube-api-access-z64kg\") pod \"calico-typha-568ddcdc8c-wkc6k\" (UID: \"eb4cd17c-d008-48e1-bdd8-356e96855153\") " pod="calico-system/calico-typha-568ddcdc8c-wkc6k" Aug 13 00:03:10.413334 env[1913]: time="2025-08-13T00:03:10.412677622Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-568ddcdc8c-wkc6k,Uid:eb4cd17c-d008-48e1-bdd8-356e96855153,Namespace:calico-system,Attempt:0,}" Aug 13 00:03:10.450706 env[1913]: time="2025-08-13T00:03:10.447509709Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:03:10.450706 env[1913]: time="2025-08-13T00:03:10.447763908Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:03:10.450706 env[1913]: time="2025-08-13T00:03:10.447876254Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:03:10.450706 env[1913]: time="2025-08-13T00:03:10.449069236Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/13c88e873c8a2be67cb3690299ca531f607dc5dfc169fc284a48f1ec29116dc7 pid=3375 runtime=io.containerd.runc.v2 Aug 13 00:03:10.497315 kubelet[2984]: I0813 00:03:10.496514 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/efec1a1f-91df-455b-97a6-fe740f6229c1-cni-net-dir\") pod \"calico-node-wvcf6\" (UID: \"efec1a1f-91df-455b-97a6-fe740f6229c1\") " pod="calico-system/calico-node-wvcf6" Aug 13 00:03:10.497315 kubelet[2984]: I0813 00:03:10.496580 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/efec1a1f-91df-455b-97a6-fe740f6229c1-tigera-ca-bundle\") pod \"calico-node-wvcf6\" (UID: \"efec1a1f-91df-455b-97a6-fe740f6229c1\") " pod="calico-system/calico-node-wvcf6" Aug 13 00:03:10.497315 kubelet[2984]: I0813 00:03:10.496624 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/efec1a1f-91df-455b-97a6-fe740f6229c1-var-lib-calico\") pod \"calico-node-wvcf6\" (UID: \"efec1a1f-91df-455b-97a6-fe740f6229c1\") " pod="calico-system/calico-node-wvcf6" Aug 13 00:03:10.497315 kubelet[2984]: I0813 00:03:10.496668 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/efec1a1f-91df-455b-97a6-fe740f6229c1-lib-modules\") pod \"calico-node-wvcf6\" (UID: \"efec1a1f-91df-455b-97a6-fe740f6229c1\") " pod="calico-system/calico-node-wvcf6" Aug 13 00:03:10.497315 kubelet[2984]: I0813 00:03:10.496704 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/efec1a1f-91df-455b-97a6-fe740f6229c1-flexvol-driver-host\") pod \"calico-node-wvcf6\" (UID: \"efec1a1f-91df-455b-97a6-fe740f6229c1\") " pod="calico-system/calico-node-wvcf6" Aug 13 00:03:10.497315 kubelet[2984]: I0813 00:03:10.496748 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/efec1a1f-91df-455b-97a6-fe740f6229c1-node-certs\") pod \"calico-node-wvcf6\" (UID: \"efec1a1f-91df-455b-97a6-fe740f6229c1\") " pod="calico-system/calico-node-wvcf6" Aug 13 00:03:10.497315 kubelet[2984]: I0813 00:03:10.496786 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/efec1a1f-91df-455b-97a6-fe740f6229c1-cni-log-dir\") pod \"calico-node-wvcf6\" (UID: \"efec1a1f-91df-455b-97a6-fe740f6229c1\") " pod="calico-system/calico-node-wvcf6" Aug 13 00:03:10.497315 kubelet[2984]: I0813 00:03:10.496820 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/efec1a1f-91df-455b-97a6-fe740f6229c1-var-run-calico\") pod \"calico-node-wvcf6\" (UID: \"efec1a1f-91df-455b-97a6-fe740f6229c1\") " pod="calico-system/calico-node-wvcf6" Aug 13 00:03:10.497315 kubelet[2984]: I0813 00:03:10.496860 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/efec1a1f-91df-455b-97a6-fe740f6229c1-cni-bin-dir\") pod \"calico-node-wvcf6\" (UID: \"efec1a1f-91df-455b-97a6-fe740f6229c1\") " pod="calico-system/calico-node-wvcf6" Aug 13 00:03:10.497315 kubelet[2984]: I0813 00:03:10.496897 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/efec1a1f-91df-455b-97a6-fe740f6229c1-policysync\") pod \"calico-node-wvcf6\" (UID: \"efec1a1f-91df-455b-97a6-fe740f6229c1\") " pod="calico-system/calico-node-wvcf6" Aug 13 00:03:10.497315 kubelet[2984]: I0813 00:03:10.496932 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/efec1a1f-91df-455b-97a6-fe740f6229c1-xtables-lock\") pod \"calico-node-wvcf6\" (UID: \"efec1a1f-91df-455b-97a6-fe740f6229c1\") " pod="calico-system/calico-node-wvcf6" Aug 13 00:03:10.497315 kubelet[2984]: I0813 00:03:10.496976 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpfmw\" (UniqueName: \"kubernetes.io/projected/efec1a1f-91df-455b-97a6-fe740f6229c1-kube-api-access-zpfmw\") pod \"calico-node-wvcf6\" (UID: \"efec1a1f-91df-455b-97a6-fe740f6229c1\") " pod="calico-system/calico-node-wvcf6" Aug 13 00:03:10.594988 kubelet[2984]: E0813 00:03:10.594605 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-n4g7q" podUID="21d84c26-2961-4407-a494-a46bbc34cc12" Aug 13 00:03:10.611921 kubelet[2984]: E0813 00:03:10.611885 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.612129 kubelet[2984]: W0813 00:03:10.612080 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.612331 kubelet[2984]: E0813 00:03:10.612303 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.623848 kubelet[2984]: E0813 00:03:10.623803 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.624130 kubelet[2984]: W0813 00:03:10.624074 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.624293 kubelet[2984]: E0813 00:03:10.624257 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.630188 kubelet[2984]: E0813 00:03:10.628415 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.630188 kubelet[2984]: W0813 00:03:10.628457 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.630188 kubelet[2984]: E0813 00:03:10.628493 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.670978 kubelet[2984]: E0813 00:03:10.670833 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.670978 kubelet[2984]: W0813 00:03:10.670883 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.670978 kubelet[2984]: E0813 00:03:10.670918 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.672751 kubelet[2984]: E0813 00:03:10.671390 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.672751 kubelet[2984]: W0813 00:03:10.671421 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.672751 kubelet[2984]: E0813 00:03:10.671448 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.672751 kubelet[2984]: E0813 00:03:10.671804 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.672751 kubelet[2984]: W0813 00:03:10.671821 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.672751 kubelet[2984]: E0813 00:03:10.671840 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.672751 kubelet[2984]: E0813 00:03:10.672168 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.672751 kubelet[2984]: W0813 00:03:10.672183 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.672751 kubelet[2984]: E0813 00:03:10.672203 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.672751 kubelet[2984]: E0813 00:03:10.672542 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.672751 kubelet[2984]: W0813 00:03:10.672561 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.672751 kubelet[2984]: E0813 00:03:10.672582 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.673659 kubelet[2984]: E0813 00:03:10.672943 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.673659 kubelet[2984]: W0813 00:03:10.672976 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.673659 kubelet[2984]: E0813 00:03:10.672999 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.673659 kubelet[2984]: E0813 00:03:10.673406 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.673659 kubelet[2984]: W0813 00:03:10.673440 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.673659 kubelet[2984]: E0813 00:03:10.673463 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.675333 kubelet[2984]: E0813 00:03:10.675278 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.675333 kubelet[2984]: W0813 00:03:10.675320 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.675571 kubelet[2984]: E0813 00:03:10.675355 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.677559 kubelet[2984]: E0813 00:03:10.675831 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.677559 kubelet[2984]: W0813 00:03:10.675867 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.677559 kubelet[2984]: E0813 00:03:10.675893 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.677559 kubelet[2984]: E0813 00:03:10.677309 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.677559 kubelet[2984]: W0813 00:03:10.677339 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.677559 kubelet[2984]: E0813 00:03:10.677376 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.680211 kubelet[2984]: E0813 00:03:10.680096 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.680211 kubelet[2984]: W0813 00:03:10.680199 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.680461 kubelet[2984]: E0813 00:03:10.680235 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.682066 kubelet[2984]: E0813 00:03:10.682009 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.682066 kubelet[2984]: W0813 00:03:10.682052 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.682364 kubelet[2984]: E0813 00:03:10.682090 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.683374 kubelet[2984]: E0813 00:03:10.683312 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.683374 kubelet[2984]: W0813 00:03:10.683351 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.683648 kubelet[2984]: E0813 00:03:10.683386 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.683879 kubelet[2984]: E0813 00:03:10.683848 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.684001 kubelet[2984]: W0813 00:03:10.683878 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.684001 kubelet[2984]: E0813 00:03:10.683904 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.685410 kubelet[2984]: E0813 00:03:10.685357 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.685410 kubelet[2984]: W0813 00:03:10.685401 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.685635 kubelet[2984]: E0813 00:03:10.685435 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.687391 kubelet[2984]: E0813 00:03:10.687340 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.687391 kubelet[2984]: W0813 00:03:10.687379 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.687608 kubelet[2984]: E0813 00:03:10.687414 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.688743 env[1913]: time="2025-08-13T00:03:10.688238571Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-wvcf6,Uid:efec1a1f-91df-455b-97a6-fe740f6229c1,Namespace:calico-system,Attempt:0,}" Aug 13 00:03:10.691155 kubelet[2984]: E0813 00:03:10.689839 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.691155 kubelet[2984]: W0813 00:03:10.689883 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.691155 kubelet[2984]: E0813 00:03:10.689944 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.691155 kubelet[2984]: E0813 00:03:10.690577 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.691155 kubelet[2984]: W0813 00:03:10.690603 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.691155 kubelet[2984]: E0813 00:03:10.690630 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.691155 kubelet[2984]: E0813 00:03:10.691161 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.691634 kubelet[2984]: W0813 00:03:10.691182 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.691634 kubelet[2984]: E0813 00:03:10.691206 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.691754 kubelet[2984]: E0813 00:03:10.691635 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.691754 kubelet[2984]: W0813 00:03:10.691653 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.691754 kubelet[2984]: E0813 00:03:10.691703 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.704840 kubelet[2984]: E0813 00:03:10.704800 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.705218 kubelet[2984]: W0813 00:03:10.705027 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.708284 kubelet[2984]: E0813 00:03:10.705348 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.708284 kubelet[2984]: I0813 00:03:10.705416 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/21d84c26-2961-4407-a494-a46bbc34cc12-kubelet-dir\") pod \"csi-node-driver-n4g7q\" (UID: \"21d84c26-2961-4407-a494-a46bbc34cc12\") " pod="calico-system/csi-node-driver-n4g7q" Aug 13 00:03:10.708826 kubelet[2984]: E0813 00:03:10.708791 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.709006 kubelet[2984]: W0813 00:03:10.708947 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.709734 kubelet[2984]: E0813 00:03:10.709689 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.710022 kubelet[2984]: I0813 00:03:10.709976 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/21d84c26-2961-4407-a494-a46bbc34cc12-registration-dir\") pod \"csi-node-driver-n4g7q\" (UID: \"21d84c26-2961-4407-a494-a46bbc34cc12\") " pod="calico-system/csi-node-driver-n4g7q" Aug 13 00:03:10.710280 kubelet[2984]: E0813 00:03:10.710192 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.710487 kubelet[2984]: W0813 00:03:10.710458 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.710648 kubelet[2984]: E0813 00:03:10.710622 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.711086 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.720148 kubelet[2984]: W0813 00:03:10.715197 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.715254 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.720148 kubelet[2984]: I0813 00:03:10.715298 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/21d84c26-2961-4407-a494-a46bbc34cc12-socket-dir\") pod \"csi-node-driver-n4g7q\" (UID: \"21d84c26-2961-4407-a494-a46bbc34cc12\") " pod="calico-system/csi-node-driver-n4g7q" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.715695 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.720148 kubelet[2984]: W0813 00:03:10.715715 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.715741 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.716042 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.720148 kubelet[2984]: W0813 00:03:10.716060 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.716079 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.716528 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.720148 kubelet[2984]: W0813 00:03:10.716549 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.716572 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.717018 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.720148 kubelet[2984]: W0813 00:03:10.717050 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.717077 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.717555 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.720148 kubelet[2984]: W0813 00:03:10.717578 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.717604 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.718002 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.720148 kubelet[2984]: W0813 00:03:10.718021 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.718046 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.720148 kubelet[2984]: I0813 00:03:10.718083 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wflrz\" (UniqueName: \"kubernetes.io/projected/21d84c26-2961-4407-a494-a46bbc34cc12-kube-api-access-wflrz\") pod \"csi-node-driver-n4g7q\" (UID: \"21d84c26-2961-4407-a494-a46bbc34cc12\") " pod="calico-system/csi-node-driver-n4g7q" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.718494 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.720148 kubelet[2984]: W0813 00:03:10.718518 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.718543 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.720148 kubelet[2984]: I0813 00:03:10.718575 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/21d84c26-2961-4407-a494-a46bbc34cc12-varrun\") pod \"csi-node-driver-n4g7q\" (UID: \"21d84c26-2961-4407-a494-a46bbc34cc12\") " pod="calico-system/csi-node-driver-n4g7q" Aug 13 00:03:10.720148 kubelet[2984]: E0813 00:03:10.719719 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.720148 kubelet[2984]: W0813 00:03:10.719751 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.721906 kubelet[2984]: E0813 00:03:10.719790 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.721906 kubelet[2984]: E0813 00:03:10.720201 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.721906 kubelet[2984]: W0813 00:03:10.720220 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.721906 kubelet[2984]: E0813 00:03:10.720243 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.721906 kubelet[2984]: E0813 00:03:10.720627 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.721906 kubelet[2984]: W0813 00:03:10.720645 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.721906 kubelet[2984]: E0813 00:03:10.720665 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.721906 kubelet[2984]: E0813 00:03:10.720927 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.721906 kubelet[2984]: W0813 00:03:10.720943 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.721906 kubelet[2984]: E0813 00:03:10.720962 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.745593 env[1913]: time="2025-08-13T00:03:10.745460324Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:03:10.745746 env[1913]: time="2025-08-13T00:03:10.745610626Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:03:10.745746 env[1913]: time="2025-08-13T00:03:10.745674695Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:03:10.746118 env[1913]: time="2025-08-13T00:03:10.746011851Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/d322b14568ffd3c041d38cb88942ebe7375d11f13c0964901f9cff59f92871ce pid=3455 runtime=io.containerd.runc.v2 Aug 13 00:03:10.820007 kubelet[2984]: E0813 00:03:10.819605 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.820007 kubelet[2984]: W0813 00:03:10.819654 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.820007 kubelet[2984]: E0813 00:03:10.819688 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.820695 kubelet[2984]: E0813 00:03:10.820650 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.820695 kubelet[2984]: W0813 00:03:10.820685 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.820874 kubelet[2984]: E0813 00:03:10.820725 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.823133 kubelet[2984]: E0813 00:03:10.821261 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.823133 kubelet[2984]: W0813 00:03:10.821296 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.823133 kubelet[2984]: E0813 00:03:10.821332 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.823133 kubelet[2984]: E0813 00:03:10.821839 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.823133 kubelet[2984]: W0813 00:03:10.821863 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.823133 kubelet[2984]: E0813 00:03:10.822018 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.823133 kubelet[2984]: E0813 00:03:10.822339 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.823133 kubelet[2984]: W0813 00:03:10.822356 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.823133 kubelet[2984]: E0813 00:03:10.822503 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.823133 kubelet[2984]: E0813 00:03:10.822762 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.823133 kubelet[2984]: W0813 00:03:10.822778 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.823133 kubelet[2984]: E0813 00:03:10.822916 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.830139 kubelet[2984]: E0813 00:03:10.825295 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.830139 kubelet[2984]: W0813 00:03:10.825336 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.830139 kubelet[2984]: E0813 00:03:10.825562 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.830139 kubelet[2984]: E0813 00:03:10.825901 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.830139 kubelet[2984]: W0813 00:03:10.825923 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.830139 kubelet[2984]: E0813 00:03:10.826128 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.830139 kubelet[2984]: E0813 00:03:10.826489 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.830139 kubelet[2984]: W0813 00:03:10.826510 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.830139 kubelet[2984]: E0813 00:03:10.826681 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.830139 kubelet[2984]: E0813 00:03:10.828526 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.830139 kubelet[2984]: W0813 00:03:10.828551 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.830139 kubelet[2984]: E0813 00:03:10.828736 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.830909 env[1913]: time="2025-08-13T00:03:10.827899085Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-568ddcdc8c-wkc6k,Uid:eb4cd17c-d008-48e1-bdd8-356e96855153,Namespace:calico-system,Attempt:0,} returns sandbox id \"13c88e873c8a2be67cb3690299ca531f607dc5dfc169fc284a48f1ec29116dc7\"" Aug 13 00:03:10.833053 env[1913]: time="2025-08-13T00:03:10.832970478Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\"" Aug 13 00:03:10.835136 kubelet[2984]: E0813 00:03:10.834178 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.835136 kubelet[2984]: W0813 00:03:10.834215 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.835136 kubelet[2984]: E0813 00:03:10.834400 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.840304 kubelet[2984]: E0813 00:03:10.840211 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.840304 kubelet[2984]: W0813 00:03:10.840291 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.840610 kubelet[2984]: E0813 00:03:10.840453 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.844342 kubelet[2984]: E0813 00:03:10.844275 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.844342 kubelet[2984]: W0813 00:03:10.844325 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.844596 kubelet[2984]: E0813 00:03:10.844544 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.847133 kubelet[2984]: E0813 00:03:10.844881 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.847133 kubelet[2984]: W0813 00:03:10.844914 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.847133 kubelet[2984]: E0813 00:03:10.845125 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.847133 kubelet[2984]: E0813 00:03:10.845464 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.847133 kubelet[2984]: W0813 00:03:10.845489 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.847133 kubelet[2984]: E0813 00:03:10.845685 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.847133 kubelet[2984]: E0813 00:03:10.846214 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.847133 kubelet[2984]: W0813 00:03:10.846272 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.847133 kubelet[2984]: E0813 00:03:10.846464 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.847133 kubelet[2984]: E0813 00:03:10.846771 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.847133 kubelet[2984]: W0813 00:03:10.846805 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.847133 kubelet[2984]: E0813 00:03:10.846957 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.847925 kubelet[2984]: E0813 00:03:10.847895 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.847925 kubelet[2984]: W0813 00:03:10.847918 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.850161 kubelet[2984]: E0813 00:03:10.848085 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.850161 kubelet[2984]: E0813 00:03:10.848397 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.850161 kubelet[2984]: W0813 00:03:10.848432 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.850161 kubelet[2984]: E0813 00:03:10.848639 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.850161 kubelet[2984]: E0813 00:03:10.848952 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.850161 kubelet[2984]: W0813 00:03:10.848969 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.850161 kubelet[2984]: E0813 00:03:10.849155 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.850814 kubelet[2984]: E0813 00:03:10.850767 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.850942 kubelet[2984]: W0813 00:03:10.850915 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.851314 kubelet[2984]: E0813 00:03:10.851286 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.852051 kubelet[2984]: E0813 00:03:10.852024 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.852244 kubelet[2984]: W0813 00:03:10.852217 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.852525 kubelet[2984]: E0813 00:03:10.852501 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.853448 kubelet[2984]: E0813 00:03:10.853414 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.853671 kubelet[2984]: W0813 00:03:10.853640 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.854340 kubelet[2984]: E0813 00:03:10.854306 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.854943 kubelet[2984]: E0813 00:03:10.854913 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.855196 kubelet[2984]: W0813 00:03:10.855165 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.855335 kubelet[2984]: E0813 00:03:10.855310 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.856436 kubelet[2984]: E0813 00:03:10.856397 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.856643 kubelet[2984]: W0813 00:03:10.856614 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.856775 kubelet[2984]: E0813 00:03:10.856748 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.931366 kubelet[2984]: E0813 00:03:10.931219 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.931579 kubelet[2984]: W0813 00:03:10.931545 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.931749 kubelet[2984]: E0813 00:03:10.931722 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.932714 kubelet[2984]: E0813 00:03:10.932681 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:10.932893 kubelet[2984]: W0813 00:03:10.932864 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:10.933044 kubelet[2984]: E0813 00:03:10.933015 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:10.947186 env[1913]: time="2025-08-13T00:03:10.947059943Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-wvcf6,Uid:efec1a1f-91df-455b-97a6-fe740f6229c1,Namespace:calico-system,Attempt:0,} returns sandbox id \"d322b14568ffd3c041d38cb88942ebe7375d11f13c0964901f9cff59f92871ce\"" Aug 13 00:03:11.045000 audit[3525]: NETFILTER_CFG table=filter:97 family=2 entries=20 op=nft_register_rule pid=3525 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:11.045000 audit[3525]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=8224 a0=3 a1=ffffd41672e0 a2=0 a3=1 items=0 ppid=3113 pid=3525 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:11.045000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:11.057000 audit[3525]: NETFILTER_CFG table=nat:98 family=2 entries=12 op=nft_register_rule pid=3525 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:11.057000 audit[3525]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=2700 a0=3 a1=ffffd41672e0 a2=0 a3=1 items=0 ppid=3113 pid=3525 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:11.057000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:11.325170 systemd[1]: run-containerd-runc-k8s.io-13c88e873c8a2be67cb3690299ca531f607dc5dfc169fc284a48f1ec29116dc7-runc.XMOviD.mount: Deactivated successfully. Aug 13 00:03:11.907063 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount624920054.mount: Deactivated successfully. Aug 13 00:03:12.474876 kubelet[2984]: E0813 00:03:12.474822 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-n4g7q" podUID="21d84c26-2961-4407-a494-a46bbc34cc12" Aug 13 00:03:13.309819 env[1913]: time="2025-08-13T00:03:13.309753992Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/typha:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:13.313710 env[1913]: time="2025-08-13T00:03:13.313660600Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:bd819526ff844d29b60cd75e846a1f55306016ff269d881d52a9b6c7b2eef0b2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:13.316572 env[1913]: time="2025-08-13T00:03:13.316519308Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/typha:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:13.319596 env[1913]: time="2025-08-13T00:03:13.319545238Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:13.320584 env[1913]: time="2025-08-13T00:03:13.320540709Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\" returns image reference \"sha256:bd819526ff844d29b60cd75e846a1f55306016ff269d881d52a9b6c7b2eef0b2\"" Aug 13 00:03:13.326752 env[1913]: time="2025-08-13T00:03:13.326683602Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\"" Aug 13 00:03:13.355696 env[1913]: time="2025-08-13T00:03:13.355623570Z" level=info msg="CreateContainer within sandbox \"13c88e873c8a2be67cb3690299ca531f607dc5dfc169fc284a48f1ec29116dc7\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Aug 13 00:03:13.385982 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4219442187.mount: Deactivated successfully. Aug 13 00:03:13.394943 env[1913]: time="2025-08-13T00:03:13.394855970Z" level=info msg="CreateContainer within sandbox \"13c88e873c8a2be67cb3690299ca531f607dc5dfc169fc284a48f1ec29116dc7\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"feb014d527407888715b11424cad39a2a0325103cf40d2bec92e95ffff0ea407\"" Aug 13 00:03:13.405352 env[1913]: time="2025-08-13T00:03:13.397799915Z" level=info msg="StartContainer for \"feb014d527407888715b11424cad39a2a0325103cf40d2bec92e95ffff0ea407\"" Aug 13 00:03:13.545786 env[1913]: time="2025-08-13T00:03:13.545719898Z" level=info msg="StartContainer for \"feb014d527407888715b11424cad39a2a0325103cf40d2bec92e95ffff0ea407\" returns successfully" Aug 13 00:03:13.812872 kubelet[2984]: E0813 00:03:13.812812 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.812872 kubelet[2984]: W0813 00:03:13.812855 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.813581 kubelet[2984]: E0813 00:03:13.812911 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.814433 kubelet[2984]: E0813 00:03:13.814353 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.814433 kubelet[2984]: W0813 00:03:13.814424 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.814633 kubelet[2984]: E0813 00:03:13.814484 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.816423 kubelet[2984]: E0813 00:03:13.816362 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.816585 kubelet[2984]: W0813 00:03:13.816408 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.816585 kubelet[2984]: E0813 00:03:13.816469 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.818455 kubelet[2984]: E0813 00:03:13.818393 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.818621 kubelet[2984]: W0813 00:03:13.818440 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.818621 kubelet[2984]: E0813 00:03:13.818500 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.821524 kubelet[2984]: E0813 00:03:13.821469 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.821689 kubelet[2984]: W0813 00:03:13.821531 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.821689 kubelet[2984]: E0813 00:03:13.821570 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.823377 kubelet[2984]: E0813 00:03:13.823310 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.823377 kubelet[2984]: W0813 00:03:13.823355 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.823639 kubelet[2984]: E0813 00:03:13.823416 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.827358 kubelet[2984]: E0813 00:03:13.827296 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.827358 kubelet[2984]: W0813 00:03:13.827342 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.827576 kubelet[2984]: E0813 00:03:13.827379 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.829364 kubelet[2984]: E0813 00:03:13.829302 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.829364 kubelet[2984]: W0813 00:03:13.829349 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.829606 kubelet[2984]: E0813 00:03:13.829385 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.830047 kubelet[2984]: E0813 00:03:13.830000 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.830047 kubelet[2984]: W0813 00:03:13.830038 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.830289 kubelet[2984]: E0813 00:03:13.830070 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.831351 kubelet[2984]: E0813 00:03:13.831297 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.831351 kubelet[2984]: W0813 00:03:13.831339 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.831567 kubelet[2984]: E0813 00:03:13.831377 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.834380 kubelet[2984]: E0813 00:03:13.834325 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.834380 kubelet[2984]: W0813 00:03:13.834365 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.834655 kubelet[2984]: E0813 00:03:13.834402 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.836416 kubelet[2984]: E0813 00:03:13.836350 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.836416 kubelet[2984]: W0813 00:03:13.836406 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.836650 kubelet[2984]: E0813 00:03:13.836442 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.837042 kubelet[2984]: E0813 00:03:13.837003 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.837042 kubelet[2984]: W0813 00:03:13.837036 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.837307 kubelet[2984]: E0813 00:03:13.837064 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.837737 kubelet[2984]: E0813 00:03:13.837695 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.837737 kubelet[2984]: W0813 00:03:13.837732 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.837929 kubelet[2984]: E0813 00:03:13.837760 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.839361 kubelet[2984]: E0813 00:03:13.839295 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.839361 kubelet[2984]: W0813 00:03:13.839335 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.843251 kubelet[2984]: E0813 00:03:13.843186 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.857418 kubelet[2984]: E0813 00:03:13.857376 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.857418 kubelet[2984]: W0813 00:03:13.857413 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.857640 kubelet[2984]: E0813 00:03:13.857446 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.858323 kubelet[2984]: E0813 00:03:13.858279 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.858323 kubelet[2984]: W0813 00:03:13.858315 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.858581 kubelet[2984]: E0813 00:03:13.858355 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.859641 kubelet[2984]: E0813 00:03:13.859266 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.859641 kubelet[2984]: W0813 00:03:13.859298 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.859641 kubelet[2984]: E0813 00:03:13.859332 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.860274 kubelet[2984]: E0813 00:03:13.859966 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.860274 kubelet[2984]: W0813 00:03:13.859991 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.860274 kubelet[2984]: E0813 00:03:13.860017 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.863082 kubelet[2984]: E0813 00:03:13.862207 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.863082 kubelet[2984]: W0813 00:03:13.862242 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.863082 kubelet[2984]: E0813 00:03:13.862448 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.863787 kubelet[2984]: E0813 00:03:13.863754 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.863980 kubelet[2984]: W0813 00:03:13.863949 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.864168 kubelet[2984]: E0813 00:03:13.864139 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.866158 kubelet[2984]: E0813 00:03:13.864609 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.866158 kubelet[2984]: W0813 00:03:13.864643 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.866158 kubelet[2984]: E0813 00:03:13.864686 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.866158 kubelet[2984]: E0813 00:03:13.865079 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.866158 kubelet[2984]: W0813 00:03:13.865132 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.866158 kubelet[2984]: E0813 00:03:13.865166 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.866158 kubelet[2984]: E0813 00:03:13.865598 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.866158 kubelet[2984]: W0813 00:03:13.865621 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.866158 kubelet[2984]: E0813 00:03:13.865646 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.866158 kubelet[2984]: E0813 00:03:13.865934 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.866158 kubelet[2984]: W0813 00:03:13.865953 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.866158 kubelet[2984]: E0813 00:03:13.865976 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.866916 kubelet[2984]: E0813 00:03:13.866357 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.866916 kubelet[2984]: W0813 00:03:13.866383 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.866916 kubelet[2984]: E0813 00:03:13.866406 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.866916 kubelet[2984]: E0813 00:03:13.866739 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.866916 kubelet[2984]: W0813 00:03:13.866756 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.866916 kubelet[2984]: E0813 00:03:13.866776 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.869138 kubelet[2984]: E0813 00:03:13.867475 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.869138 kubelet[2984]: W0813 00:03:13.867507 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.869138 kubelet[2984]: E0813 00:03:13.867798 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.869138 kubelet[2984]: W0813 00:03:13.867814 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.869138 kubelet[2984]: E0813 00:03:13.867835 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.869707 kubelet[2984]: E0813 00:03:13.869610 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.871350 kubelet[2984]: E0813 00:03:13.871292 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.871350 kubelet[2984]: W0813 00:03:13.871337 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.871588 kubelet[2984]: E0813 00:03:13.871372 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.874174 kubelet[2984]: E0813 00:03:13.871763 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.874174 kubelet[2984]: W0813 00:03:13.871794 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.874174 kubelet[2984]: E0813 00:03:13.871819 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.874174 kubelet[2984]: E0813 00:03:13.872300 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.874174 kubelet[2984]: W0813 00:03:13.872322 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.874174 kubelet[2984]: E0813 00:03:13.872351 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:13.874635 kubelet[2984]: E0813 00:03:13.874265 2984 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:03:13.874635 kubelet[2984]: W0813 00:03:13.874295 2984 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:03:13.874635 kubelet[2984]: E0813 00:03:13.874327 2984 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:03:14.475166 kubelet[2984]: E0813 00:03:14.475059 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-n4g7q" podUID="21d84c26-2961-4407-a494-a46bbc34cc12" Aug 13 00:03:14.532876 env[1913]: time="2025-08-13T00:03:14.532814937Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:14.539040 env[1913]: time="2025-08-13T00:03:14.538958656Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:53f638101e3d73f7dd5e42dc42fb3d94ae1978e8958677222c3de6ec1d8c3d4f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:14.547630 env[1913]: time="2025-08-13T00:03:14.547553678Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:14.554172 env[1913]: time="2025-08-13T00:03:14.554048017Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" returns image reference \"sha256:53f638101e3d73f7dd5e42dc42fb3d94ae1978e8958677222c3de6ec1d8c3d4f\"" Aug 13 00:03:14.554518 env[1913]: time="2025-08-13T00:03:14.554474118Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:14.562043 env[1913]: time="2025-08-13T00:03:14.561967036Z" level=info msg="CreateContainer within sandbox \"d322b14568ffd3c041d38cb88942ebe7375d11f13c0964901f9cff59f92871ce\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Aug 13 00:03:14.591251 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3670182628.mount: Deactivated successfully. Aug 13 00:03:14.604200 env[1913]: time="2025-08-13T00:03:14.604129186Z" level=info msg="CreateContainer within sandbox \"d322b14568ffd3c041d38cb88942ebe7375d11f13c0964901f9cff59f92871ce\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"f60855b6009909f3480ec16459e94a737bb08df4f93df6056c14fb652360c0a1\"" Aug 13 00:03:14.607641 env[1913]: time="2025-08-13T00:03:14.607555175Z" level=info msg="StartContainer for \"f60855b6009909f3480ec16459e94a737bb08df4f93df6056c14fb652360c0a1\"" Aug 13 00:03:14.748848 env[1913]: time="2025-08-13T00:03:14.746991063Z" level=info msg="StartContainer for \"f60855b6009909f3480ec16459e94a737bb08df4f93df6056c14fb652360c0a1\" returns successfully" Aug 13 00:03:14.836637 kubelet[2984]: I0813 00:03:14.836415 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-568ddcdc8c-wkc6k" podStartSLOduration=2.345917261 podStartE2EDuration="4.836279457s" podCreationTimestamp="2025-08-13 00:03:10 +0000 UTC" firstStartedPulling="2025-08-13 00:03:10.832036578 +0000 UTC m=+43.579215474" lastFinishedPulling="2025-08-13 00:03:13.322398774 +0000 UTC m=+46.069577670" observedRunningTime="2025-08-13 00:03:13.889664659 +0000 UTC m=+46.636843591" watchObservedRunningTime="2025-08-13 00:03:14.836279457 +0000 UTC m=+47.583458353" Aug 13 00:03:14.910000 audit[3664]: NETFILTER_CFG table=filter:99 family=2 entries=21 op=nft_register_rule pid=3664 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:14.915176 kernel: kauditd_printk_skb: 8 callbacks suppressed Aug 13 00:03:14.915292 kernel: audit: type=1325 audit(1755043394.910:291): table=filter:99 family=2 entries=21 op=nft_register_rule pid=3664 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:14.910000 audit[3664]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=7480 a0=3 a1=ffffc0e4c790 a2=0 a3=1 items=0 ppid=3113 pid=3664 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:14.937264 kernel: audit: type=1300 audit(1755043394.910:291): arch=c00000b7 syscall=211 success=yes exit=7480 a0=3 a1=ffffc0e4c790 a2=0 a3=1 items=0 ppid=3113 pid=3664 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:14.910000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:14.950938 kernel: audit: type=1327 audit(1755043394.910:291): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:14.926000 audit[3664]: NETFILTER_CFG table=nat:100 family=2 entries=19 op=nft_register_chain pid=3664 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:14.926000 audit[3664]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=6276 a0=3 a1=ffffc0e4c790 a2=0 a3=1 items=0 ppid=3113 pid=3664 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:14.975143 kernel: audit: type=1325 audit(1755043394.926:292): table=nat:100 family=2 entries=19 op=nft_register_chain pid=3664 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:14.975312 kernel: audit: type=1300 audit(1755043394.926:292): arch=c00000b7 syscall=211 success=yes exit=6276 a0=3 a1=ffffc0e4c790 a2=0 a3=1 items=0 ppid=3113 pid=3664 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:14.926000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:14.983405 kernel: audit: type=1327 audit(1755043394.926:292): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:15.223069 env[1913]: time="2025-08-13T00:03:15.222652506Z" level=info msg="shim disconnected" id=f60855b6009909f3480ec16459e94a737bb08df4f93df6056c14fb652360c0a1 Aug 13 00:03:15.223069 env[1913]: time="2025-08-13T00:03:15.222719239Z" level=warning msg="cleaning up after shim disconnected" id=f60855b6009909f3480ec16459e94a737bb08df4f93df6056c14fb652360c0a1 namespace=k8s.io Aug 13 00:03:15.223069 env[1913]: time="2025-08-13T00:03:15.222740143Z" level=info msg="cleaning up dead shim" Aug 13 00:03:15.238386 env[1913]: time="2025-08-13T00:03:15.238327854Z" level=warning msg="cleanup warnings time=\"2025-08-13T00:03:15Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3665 runtime=io.containerd.runc.v2\n" Aug 13 00:03:15.335054 systemd[1]: run-containerd-runc-k8s.io-f60855b6009909f3480ec16459e94a737bb08df4f93df6056c14fb652360c0a1-runc.02lnbg.mount: Deactivated successfully. Aug 13 00:03:15.335353 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f60855b6009909f3480ec16459e94a737bb08df4f93df6056c14fb652360c0a1-rootfs.mount: Deactivated successfully. Aug 13 00:03:15.816822 env[1913]: time="2025-08-13T00:03:15.816751794Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\"" Aug 13 00:03:16.474498 kubelet[2984]: E0813 00:03:16.474409 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-n4g7q" podUID="21d84c26-2961-4407-a494-a46bbc34cc12" Aug 13 00:03:18.474897 kubelet[2984]: E0813 00:03:18.474817 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-n4g7q" podUID="21d84c26-2961-4407-a494-a46bbc34cc12" Aug 13 00:03:19.625887 env[1913]: time="2025-08-13T00:03:19.625809290Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/cni:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:19.630144 env[1913]: time="2025-08-13T00:03:19.630066272Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:f6e344d58b3c5524e767c7d1dd4cb29c85ce820b0f3005a603532b6a22db5588,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:19.633682 env[1913]: time="2025-08-13T00:03:19.633618751Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/cni:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:19.637185 env[1913]: time="2025-08-13T00:03:19.637134725Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:19.638567 env[1913]: time="2025-08-13T00:03:19.638506038Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\" returns image reference \"sha256:f6e344d58b3c5524e767c7d1dd4cb29c85ce820b0f3005a603532b6a22db5588\"" Aug 13 00:03:19.646281 env[1913]: time="2025-08-13T00:03:19.646224298Z" level=info msg="CreateContainer within sandbox \"d322b14568ffd3c041d38cb88942ebe7375d11f13c0964901f9cff59f92871ce\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Aug 13 00:03:19.674087 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1546437348.mount: Deactivated successfully. Aug 13 00:03:19.685710 env[1913]: time="2025-08-13T00:03:19.685646543Z" level=info msg="CreateContainer within sandbox \"d322b14568ffd3c041d38cb88942ebe7375d11f13c0964901f9cff59f92871ce\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"53436430922fcd21cefe2c3fdd4e7a1886e0f8c8be304d48baaebe68523335a8\"" Aug 13 00:03:19.688943 env[1913]: time="2025-08-13T00:03:19.688722077Z" level=info msg="StartContainer for \"53436430922fcd21cefe2c3fdd4e7a1886e0f8c8be304d48baaebe68523335a8\"" Aug 13 00:03:19.744262 systemd[1]: run-containerd-runc-k8s.io-53436430922fcd21cefe2c3fdd4e7a1886e0f8c8be304d48baaebe68523335a8-runc.lhurAE.mount: Deactivated successfully. Aug 13 00:03:19.830372 env[1913]: time="2025-08-13T00:03:19.830305822Z" level=info msg="StartContainer for \"53436430922fcd21cefe2c3fdd4e7a1886e0f8c8be304d48baaebe68523335a8\" returns successfully" Aug 13 00:03:20.474951 kubelet[2984]: E0813 00:03:20.474887 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-n4g7q" podUID="21d84c26-2961-4407-a494-a46bbc34cc12" Aug 13 00:03:20.920493 env[1913]: time="2025-08-13T00:03:20.920380110Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/calico-kubeconfig\": WRITE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 00:03:20.966923 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-53436430922fcd21cefe2c3fdd4e7a1886e0f8c8be304d48baaebe68523335a8-rootfs.mount: Deactivated successfully. Aug 13 00:03:21.001520 kubelet[2984]: I0813 00:03:21.001474 2984 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Aug 13 00:03:21.079999 kubelet[2984]: W0813 00:03:21.079951 2984 reflector.go:561] object-"kube-system"/"coredns": failed to list *v1.ConfigMap: configmaps "coredns" is forbidden: User "system:node:ip-172-31-27-151" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ip-172-31-27-151' and this object Aug 13 00:03:21.096715 kubelet[2984]: E0813 00:03:21.096636 2984 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"coredns\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"coredns\" is forbidden: User \"system:node:ip-172-31-27-151\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ip-172-31-27-151' and this object" logger="UnhandledError" Aug 13 00:03:21.134557 kubelet[2984]: I0813 00:03:21.134494 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6d611ea-61c9-46e6-9d86-1c24bd310146-whisker-ca-bundle\") pod \"whisker-755dcdc665-9mnng\" (UID: \"f6d611ea-61c9-46e6-9d86-1c24bd310146\") " pod="calico-system/whisker-755dcdc665-9mnng" Aug 13 00:03:21.134744 kubelet[2984]: I0813 00:03:21.134619 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48c5f0fe-d3a5-4b24-a4f6-448275db6b4d-config-volume\") pod \"coredns-7c65d6cfc9-kqx4k\" (UID: \"48c5f0fe-d3a5-4b24-a4f6-448275db6b4d\") " pod="kube-system/coredns-7c65d6cfc9-kqx4k" Aug 13 00:03:21.134744 kubelet[2984]: I0813 00:03:21.134714 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7mmh\" (UniqueName: \"kubernetes.io/projected/48c5f0fe-d3a5-4b24-a4f6-448275db6b4d-kube-api-access-m7mmh\") pod \"coredns-7c65d6cfc9-kqx4k\" (UID: \"48c5f0fe-d3a5-4b24-a4f6-448275db6b4d\") " pod="kube-system/coredns-7c65d6cfc9-kqx4k" Aug 13 00:03:21.134884 kubelet[2984]: I0813 00:03:21.134798 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/f6d611ea-61c9-46e6-9d86-1c24bd310146-whisker-backend-key-pair\") pod \"whisker-755dcdc665-9mnng\" (UID: \"f6d611ea-61c9-46e6-9d86-1c24bd310146\") " pod="calico-system/whisker-755dcdc665-9mnng" Aug 13 00:03:21.134970 kubelet[2984]: I0813 00:03:21.134887 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nlvr\" (UniqueName: \"kubernetes.io/projected/f6d611ea-61c9-46e6-9d86-1c24bd310146-kube-api-access-6nlvr\") pod \"whisker-755dcdc665-9mnng\" (UID: \"f6d611ea-61c9-46e6-9d86-1c24bd310146\") " pod="calico-system/whisker-755dcdc665-9mnng" Aug 13 00:03:21.135036 kubelet[2984]: I0813 00:03:21.134966 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjcl9\" (UniqueName: \"kubernetes.io/projected/ba5590d9-af5c-484d-9d4d-2e76ac5327b7-kube-api-access-xjcl9\") pod \"coredns-7c65d6cfc9-v6d44\" (UID: \"ba5590d9-af5c-484d-9d4d-2e76ac5327b7\") " pod="kube-system/coredns-7c65d6cfc9-v6d44" Aug 13 00:03:21.135130 kubelet[2984]: I0813 00:03:21.135056 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ba5590d9-af5c-484d-9d4d-2e76ac5327b7-config-volume\") pod \"coredns-7c65d6cfc9-v6d44\" (UID: \"ba5590d9-af5c-484d-9d4d-2e76ac5327b7\") " pod="kube-system/coredns-7c65d6cfc9-v6d44" Aug 13 00:03:21.237538 kubelet[2984]: I0813 00:03:21.237321 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-254lc\" (UniqueName: \"kubernetes.io/projected/71cf2b5c-a9b2-457e-a496-c93a3a2d8126-kube-api-access-254lc\") pod \"calico-apiserver-6ff96b959-s8g8j\" (UID: \"71cf2b5c-a9b2-457e-a496-c93a3a2d8126\") " pod="calico-apiserver/calico-apiserver-6ff96b959-s8g8j" Aug 13 00:03:21.237538 kubelet[2984]: I0813 00:03:21.237441 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce695f2d-ac34-4846-848c-9adafe443946-tigera-ca-bundle\") pod \"calico-kube-controllers-d75f5fd6d-27n9j\" (UID: \"ce695f2d-ac34-4846-848c-9adafe443946\") " pod="calico-system/calico-kube-controllers-d75f5fd6d-27n9j" Aug 13 00:03:21.238517 kubelet[2984]: I0813 00:03:21.237483 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/71cf2b5c-a9b2-457e-a496-c93a3a2d8126-calico-apiserver-certs\") pod \"calico-apiserver-6ff96b959-s8g8j\" (UID: \"71cf2b5c-a9b2-457e-a496-c93a3a2d8126\") " pod="calico-apiserver/calico-apiserver-6ff96b959-s8g8j" Aug 13 00:03:21.238517 kubelet[2984]: I0813 00:03:21.238097 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf609856-35b0-4517-9fba-45b28ef2be55-config\") pod \"goldmane-58fd7646b9-6tzfz\" (UID: \"cf609856-35b0-4517-9fba-45b28ef2be55\") " pod="calico-system/goldmane-58fd7646b9-6tzfz" Aug 13 00:03:21.238517 kubelet[2984]: I0813 00:03:21.238184 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkwlq\" (UniqueName: \"kubernetes.io/projected/ce695f2d-ac34-4846-848c-9adafe443946-kube-api-access-mkwlq\") pod \"calico-kube-controllers-d75f5fd6d-27n9j\" (UID: \"ce695f2d-ac34-4846-848c-9adafe443946\") " pod="calico-system/calico-kube-controllers-d75f5fd6d-27n9j" Aug 13 00:03:21.238517 kubelet[2984]: I0813 00:03:21.238250 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd4j5\" (UniqueName: \"kubernetes.io/projected/cf609856-35b0-4517-9fba-45b28ef2be55-kube-api-access-hd4j5\") pod \"goldmane-58fd7646b9-6tzfz\" (UID: \"cf609856-35b0-4517-9fba-45b28ef2be55\") " pod="calico-system/goldmane-58fd7646b9-6tzfz" Aug 13 00:03:21.238517 kubelet[2984]: I0813 00:03:21.238302 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e-calico-apiserver-certs\") pod \"calico-apiserver-6ff96b959-jggcm\" (UID: \"6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e\") " pod="calico-apiserver/calico-apiserver-6ff96b959-jggcm" Aug 13 00:03:21.238517 kubelet[2984]: I0813 00:03:21.238338 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/cf609856-35b0-4517-9fba-45b28ef2be55-goldmane-key-pair\") pod \"goldmane-58fd7646b9-6tzfz\" (UID: \"cf609856-35b0-4517-9fba-45b28ef2be55\") " pod="calico-system/goldmane-58fd7646b9-6tzfz" Aug 13 00:03:21.238517 kubelet[2984]: I0813 00:03:21.238386 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7bkr\" (UniqueName: \"kubernetes.io/projected/6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e-kube-api-access-v7bkr\") pod \"calico-apiserver-6ff96b959-jggcm\" (UID: \"6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e\") " pod="calico-apiserver/calico-apiserver-6ff96b959-jggcm" Aug 13 00:03:21.238517 kubelet[2984]: I0813 00:03:21.238449 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf609856-35b0-4517-9fba-45b28ef2be55-goldmane-ca-bundle\") pod \"goldmane-58fd7646b9-6tzfz\" (UID: \"cf609856-35b0-4517-9fba-45b28ef2be55\") " pod="calico-system/goldmane-58fd7646b9-6tzfz" Aug 13 00:03:21.422162 env[1913]: time="2025-08-13T00:03:21.422054710Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-755dcdc665-9mnng,Uid:f6d611ea-61c9-46e6-9d86-1c24bd310146,Namespace:calico-system,Attempt:0,}" Aug 13 00:03:21.465387 env[1913]: time="2025-08-13T00:03:21.465316027Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6ff96b959-s8g8j,Uid:71cf2b5c-a9b2-457e-a496-c93a3a2d8126,Namespace:calico-apiserver,Attempt:0,}" Aug 13 00:03:21.499179 env[1913]: time="2025-08-13T00:03:21.498861105Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6ff96b959-jggcm,Uid:6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e,Namespace:calico-apiserver,Attempt:0,}" Aug 13 00:03:21.505338 env[1913]: time="2025-08-13T00:03:21.505284861Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-6tzfz,Uid:cf609856-35b0-4517-9fba-45b28ef2be55,Namespace:calico-system,Attempt:0,}" Aug 13 00:03:21.509712 env[1913]: time="2025-08-13T00:03:21.509643350Z" level=info msg="shim disconnected" id=53436430922fcd21cefe2c3fdd4e7a1886e0f8c8be304d48baaebe68523335a8 Aug 13 00:03:21.510197 env[1913]: time="2025-08-13T00:03:21.509992829Z" level=warning msg="cleaning up after shim disconnected" id=53436430922fcd21cefe2c3fdd4e7a1886e0f8c8be304d48baaebe68523335a8 namespace=k8s.io Aug 13 00:03:21.510778 env[1913]: time="2025-08-13T00:03:21.510293900Z" level=info msg="cleaning up dead shim" Aug 13 00:03:21.511846 env[1913]: time="2025-08-13T00:03:21.511791478Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-d75f5fd6d-27n9j,Uid:ce695f2d-ac34-4846-848c-9adafe443946,Namespace:calico-system,Attempt:0,}" Aug 13 00:03:21.540666 env[1913]: time="2025-08-13T00:03:21.540607108Z" level=warning msg="cleanup warnings time=\"2025-08-13T00:03:21Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3738 runtime=io.containerd.runc.v2\n" Aug 13 00:03:21.807201 env[1913]: time="2025-08-13T00:03:21.807010441Z" level=error msg="Failed to destroy network for sandbox \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.808313 env[1913]: time="2025-08-13T00:03:21.808237057Z" level=error msg="encountered an error cleaning up failed sandbox \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.808580 env[1913]: time="2025-08-13T00:03:21.808525852Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-d75f5fd6d-27n9j,Uid:ce695f2d-ac34-4846-848c-9adafe443946,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.809136 kubelet[2984]: E0813 00:03:21.809046 2984 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.809805 kubelet[2984]: E0813 00:03:21.809182 2984 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-d75f5fd6d-27n9j" Aug 13 00:03:21.809805 kubelet[2984]: E0813 00:03:21.809235 2984 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-d75f5fd6d-27n9j" Aug 13 00:03:21.811694 kubelet[2984]: E0813 00:03:21.809322 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-d75f5fd6d-27n9j_calico-system(ce695f2d-ac34-4846-848c-9adafe443946)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-d75f5fd6d-27n9j_calico-system(ce695f2d-ac34-4846-848c-9adafe443946)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-d75f5fd6d-27n9j" podUID="ce695f2d-ac34-4846-848c-9adafe443946" Aug 13 00:03:21.855612 env[1913]: time="2025-08-13T00:03:21.855559044Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\"" Aug 13 00:03:21.860898 kubelet[2984]: I0813 00:03:21.860276 2984 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Aug 13 00:03:21.865820 env[1913]: time="2025-08-13T00:03:21.865746203Z" level=info msg="StopPodSandbox for \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\"" Aug 13 00:03:21.906318 env[1913]: time="2025-08-13T00:03:21.906085697Z" level=error msg="Failed to destroy network for sandbox \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.907059 env[1913]: time="2025-08-13T00:03:21.906991801Z" level=error msg="encountered an error cleaning up failed sandbox \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.907238 env[1913]: time="2025-08-13T00:03:21.907091462Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-6tzfz,Uid:cf609856-35b0-4517-9fba-45b28ef2be55,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.908056 kubelet[2984]: E0813 00:03:21.907532 2984 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.908056 kubelet[2984]: E0813 00:03:21.907613 2984 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-58fd7646b9-6tzfz" Aug 13 00:03:21.908056 kubelet[2984]: E0813 00:03:21.907649 2984 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-58fd7646b9-6tzfz" Aug 13 00:03:21.908056 kubelet[2984]: E0813 00:03:21.907727 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-58fd7646b9-6tzfz_calico-system(cf609856-35b0-4517-9fba-45b28ef2be55)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-58fd7646b9-6tzfz_calico-system(cf609856-35b0-4517-9fba-45b28ef2be55)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-58fd7646b9-6tzfz" podUID="cf609856-35b0-4517-9fba-45b28ef2be55" Aug 13 00:03:21.908485 env[1913]: time="2025-08-13T00:03:21.907656943Z" level=error msg="Failed to destroy network for sandbox \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.909827 env[1913]: time="2025-08-13T00:03:21.909748971Z" level=error msg="encountered an error cleaning up failed sandbox \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.910315 env[1913]: time="2025-08-13T00:03:21.910253240Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6ff96b959-jggcm,Uid:6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.910994 kubelet[2984]: E0813 00:03:21.910922 2984 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.911189 kubelet[2984]: E0813 00:03:21.911004 2984 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6ff96b959-jggcm" Aug 13 00:03:21.911189 kubelet[2984]: E0813 00:03:21.911040 2984 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6ff96b959-jggcm" Aug 13 00:03:21.911189 kubelet[2984]: E0813 00:03:21.911144 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6ff96b959-jggcm_calico-apiserver(6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6ff96b959-jggcm_calico-apiserver(6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6ff96b959-jggcm" podUID="6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e" Aug 13 00:03:21.922578 env[1913]: time="2025-08-13T00:03:21.922425590Z" level=error msg="Failed to destroy network for sandbox \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.924407 env[1913]: time="2025-08-13T00:03:21.924325831Z" level=error msg="encountered an error cleaning up failed sandbox \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.924773 env[1913]: time="2025-08-13T00:03:21.924719603Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6ff96b959-s8g8j,Uid:71cf2b5c-a9b2-457e-a496-c93a3a2d8126,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.925481 kubelet[2984]: E0813 00:03:21.925401 2984 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:21.925637 kubelet[2984]: E0813 00:03:21.925486 2984 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6ff96b959-s8g8j" Aug 13 00:03:21.925637 kubelet[2984]: E0813 00:03:21.925529 2984 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6ff96b959-s8g8j" Aug 13 00:03:21.925637 kubelet[2984]: E0813 00:03:21.925602 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6ff96b959-s8g8j_calico-apiserver(71cf2b5c-a9b2-457e-a496-c93a3a2d8126)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6ff96b959-s8g8j_calico-apiserver(71cf2b5c-a9b2-457e-a496-c93a3a2d8126)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6ff96b959-s8g8j" podUID="71cf2b5c-a9b2-457e-a496-c93a3a2d8126" Aug 13 00:03:22.010811 env[1913]: time="2025-08-13T00:03:22.010724862Z" level=error msg="Failed to destroy network for sandbox \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:22.015343 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309-shm.mount: Deactivated successfully. Aug 13 00:03:22.016268 env[1913]: time="2025-08-13T00:03:22.016165544Z" level=error msg="encountered an error cleaning up failed sandbox \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:22.016369 env[1913]: time="2025-08-13T00:03:22.016291161Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-755dcdc665-9mnng,Uid:f6d611ea-61c9-46e6-9d86-1c24bd310146,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:22.018505 kubelet[2984]: E0813 00:03:22.016651 2984 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:22.018505 kubelet[2984]: E0813 00:03:22.016733 2984 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-755dcdc665-9mnng" Aug 13 00:03:22.018505 kubelet[2984]: E0813 00:03:22.016768 2984 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-755dcdc665-9mnng" Aug 13 00:03:22.018505 kubelet[2984]: E0813 00:03:22.016827 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-755dcdc665-9mnng_calico-system(f6d611ea-61c9-46e6-9d86-1c24bd310146)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-755dcdc665-9mnng_calico-system(f6d611ea-61c9-46e6-9d86-1c24bd310146)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-755dcdc665-9mnng" podUID="f6d611ea-61c9-46e6-9d86-1c24bd310146" Aug 13 00:03:22.036518 env[1913]: time="2025-08-13T00:03:22.036443438Z" level=error msg="StopPodSandbox for \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\" failed" error="failed to destroy network for sandbox \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:22.037295 kubelet[2984]: E0813 00:03:22.036949 2984 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Aug 13 00:03:22.037295 kubelet[2984]: E0813 00:03:22.037034 2984 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403"} Aug 13 00:03:22.037295 kubelet[2984]: E0813 00:03:22.037150 2984 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"ce695f2d-ac34-4846-848c-9adafe443946\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:03:22.037295 kubelet[2984]: E0813 00:03:22.037192 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"ce695f2d-ac34-4846-848c-9adafe443946\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-d75f5fd6d-27n9j" podUID="ce695f2d-ac34-4846-848c-9adafe443946" Aug 13 00:03:22.253879 kubelet[2984]: E0813 00:03:22.253815 2984 configmap.go:193] Couldn't get configMap kube-system/coredns: failed to sync configmap cache: timed out waiting for the condition Aug 13 00:03:22.254148 kubelet[2984]: E0813 00:03:22.253949 2984 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ba5590d9-af5c-484d-9d4d-2e76ac5327b7-config-volume podName:ba5590d9-af5c-484d-9d4d-2e76ac5327b7 nodeName:}" failed. No retries permitted until 2025-08-13 00:03:22.753920122 +0000 UTC m=+55.501099030 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/ba5590d9-af5c-484d-9d4d-2e76ac5327b7-config-volume") pod "coredns-7c65d6cfc9-v6d44" (UID: "ba5590d9-af5c-484d-9d4d-2e76ac5327b7") : failed to sync configmap cache: timed out waiting for the condition Aug 13 00:03:22.255555 kubelet[2984]: E0813 00:03:22.255400 2984 configmap.go:193] Couldn't get configMap kube-system/coredns: failed to sync configmap cache: timed out waiting for the condition Aug 13 00:03:22.255555 kubelet[2984]: E0813 00:03:22.255506 2984 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/48c5f0fe-d3a5-4b24-a4f6-448275db6b4d-config-volume podName:48c5f0fe-d3a5-4b24-a4f6-448275db6b4d nodeName:}" failed. No retries permitted until 2025-08-13 00:03:22.755480316 +0000 UTC m=+55.502659224 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/48c5f0fe-d3a5-4b24-a4f6-448275db6b4d-config-volume") pod "coredns-7c65d6cfc9-kqx4k" (UID: "48c5f0fe-d3a5-4b24-a4f6-448275db6b4d") : failed to sync configmap cache: timed out waiting for the condition Aug 13 00:03:22.480901 env[1913]: time="2025-08-13T00:03:22.480843844Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-n4g7q,Uid:21d84c26-2961-4407-a494-a46bbc34cc12,Namespace:calico-system,Attempt:0,}" Aug 13 00:03:22.604210 env[1913]: time="2025-08-13T00:03:22.599429866Z" level=error msg="Failed to destroy network for sandbox \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:22.604442 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678-shm.mount: Deactivated successfully. Aug 13 00:03:22.605714 env[1913]: time="2025-08-13T00:03:22.605640955Z" level=error msg="encountered an error cleaning up failed sandbox \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:22.605971 env[1913]: time="2025-08-13T00:03:22.605898213Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-n4g7q,Uid:21d84c26-2961-4407-a494-a46bbc34cc12,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:22.606947 kubelet[2984]: E0813 00:03:22.606883 2984 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:22.607096 kubelet[2984]: E0813 00:03:22.606986 2984 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-n4g7q" Aug 13 00:03:22.607096 kubelet[2984]: E0813 00:03:22.607050 2984 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-n4g7q" Aug 13 00:03:22.607371 kubelet[2984]: E0813 00:03:22.607255 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-n4g7q_calico-system(21d84c26-2961-4407-a494-a46bbc34cc12)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-n4g7q_calico-system(21d84c26-2961-4407-a494-a46bbc34cc12)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-n4g7q" podUID="21d84c26-2961-4407-a494-a46bbc34cc12" Aug 13 00:03:22.865275 kubelet[2984]: I0813 00:03:22.864979 2984 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Aug 13 00:03:22.869284 env[1913]: time="2025-08-13T00:03:22.866975389Z" level=info msg="StopPodSandbox for \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\"" Aug 13 00:03:22.872984 kubelet[2984]: I0813 00:03:22.872273 2984 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Aug 13 00:03:22.874897 env[1913]: time="2025-08-13T00:03:22.873241114Z" level=info msg="StopPodSandbox for \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\"" Aug 13 00:03:22.895571 env[1913]: time="2025-08-13T00:03:22.895288520Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-kqx4k,Uid:48c5f0fe-d3a5-4b24-a4f6-448275db6b4d,Namespace:kube-system,Attempt:0,}" Aug 13 00:03:22.907807 kubelet[2984]: I0813 00:03:22.907755 2984 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Aug 13 00:03:22.915863 env[1913]: time="2025-08-13T00:03:22.915804280Z" level=info msg="StopPodSandbox for \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\"" Aug 13 00:03:22.919571 kubelet[2984]: I0813 00:03:22.918829 2984 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Aug 13 00:03:22.923728 env[1913]: time="2025-08-13T00:03:22.923658424Z" level=info msg="StopPodSandbox for \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\"" Aug 13 00:03:22.930158 kubelet[2984]: I0813 00:03:22.929567 2984 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Aug 13 00:03:22.933494 env[1913]: time="2025-08-13T00:03:22.933249752Z" level=info msg="StopPodSandbox for \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\"" Aug 13 00:03:22.979051 env[1913]: time="2025-08-13T00:03:22.978994411Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-v6d44,Uid:ba5590d9-af5c-484d-9d4d-2e76ac5327b7,Namespace:kube-system,Attempt:0,}" Aug 13 00:03:23.078817 env[1913]: time="2025-08-13T00:03:23.078720973Z" level=error msg="StopPodSandbox for \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\" failed" error="failed to destroy network for sandbox \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:23.083647 kubelet[2984]: E0813 00:03:23.083415 2984 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Aug 13 00:03:23.083647 kubelet[2984]: E0813 00:03:23.083493 2984 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678"} Aug 13 00:03:23.083647 kubelet[2984]: E0813 00:03:23.083552 2984 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"21d84c26-2961-4407-a494-a46bbc34cc12\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:03:23.083647 kubelet[2984]: E0813 00:03:23.083593 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"21d84c26-2961-4407-a494-a46bbc34cc12\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-n4g7q" podUID="21d84c26-2961-4407-a494-a46bbc34cc12" Aug 13 00:03:23.233524 env[1913]: time="2025-08-13T00:03:23.233411840Z" level=error msg="StopPodSandbox for \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\" failed" error="failed to destroy network for sandbox \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:23.233901 kubelet[2984]: E0813 00:03:23.233810 2984 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Aug 13 00:03:23.234021 kubelet[2984]: E0813 00:03:23.233935 2984 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802"} Aug 13 00:03:23.234097 kubelet[2984]: E0813 00:03:23.233992 2984 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"cf609856-35b0-4517-9fba-45b28ef2be55\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:03:23.234097 kubelet[2984]: E0813 00:03:23.234058 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"cf609856-35b0-4517-9fba-45b28ef2be55\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-58fd7646b9-6tzfz" podUID="cf609856-35b0-4517-9fba-45b28ef2be55" Aug 13 00:03:23.248586 env[1913]: time="2025-08-13T00:03:23.248491991Z" level=error msg="StopPodSandbox for \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\" failed" error="failed to destroy network for sandbox \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:23.249019 kubelet[2984]: E0813 00:03:23.248949 2984 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Aug 13 00:03:23.249185 kubelet[2984]: E0813 00:03:23.249050 2984 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c"} Aug 13 00:03:23.249185 kubelet[2984]: E0813 00:03:23.249142 2984 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"71cf2b5c-a9b2-457e-a496-c93a3a2d8126\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:03:23.249398 kubelet[2984]: E0813 00:03:23.249187 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"71cf2b5c-a9b2-457e-a496-c93a3a2d8126\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6ff96b959-s8g8j" podUID="71cf2b5c-a9b2-457e-a496-c93a3a2d8126" Aug 13 00:03:23.260712 env[1913]: time="2025-08-13T00:03:23.260591855Z" level=error msg="StopPodSandbox for \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\" failed" error="failed to destroy network for sandbox \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:23.261299 kubelet[2984]: E0813 00:03:23.261190 2984 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Aug 13 00:03:23.261448 kubelet[2984]: E0813 00:03:23.261311 2984 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014"} Aug 13 00:03:23.261448 kubelet[2984]: E0813 00:03:23.261391 2984 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:03:23.261623 kubelet[2984]: E0813 00:03:23.261458 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6ff96b959-jggcm" podUID="6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e" Aug 13 00:03:23.276474 env[1913]: time="2025-08-13T00:03:23.276178267Z" level=error msg="StopPodSandbox for \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\" failed" error="failed to destroy network for sandbox \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:23.276908 kubelet[2984]: E0813 00:03:23.276832 2984 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Aug 13 00:03:23.277025 kubelet[2984]: E0813 00:03:23.276931 2984 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309"} Aug 13 00:03:23.277092 kubelet[2984]: E0813 00:03:23.277019 2984 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"f6d611ea-61c9-46e6-9d86-1c24bd310146\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:03:23.277267 kubelet[2984]: E0813 00:03:23.277093 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"f6d611ea-61c9-46e6-9d86-1c24bd310146\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-755dcdc665-9mnng" podUID="f6d611ea-61c9-46e6-9d86-1c24bd310146" Aug 13 00:03:23.315012 env[1913]: time="2025-08-13T00:03:23.314938111Z" level=error msg="Failed to destroy network for sandbox \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:23.320129 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf-shm.mount: Deactivated successfully. Aug 13 00:03:23.324592 env[1913]: time="2025-08-13T00:03:23.324515300Z" level=error msg="encountered an error cleaning up failed sandbox \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:23.324867 env[1913]: time="2025-08-13T00:03:23.324813539Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-kqx4k,Uid:48c5f0fe-d3a5-4b24-a4f6-448275db6b4d,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:23.326711 kubelet[2984]: E0813 00:03:23.325349 2984 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:23.326711 kubelet[2984]: E0813 00:03:23.325454 2984 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-kqx4k" Aug 13 00:03:23.326711 kubelet[2984]: E0813 00:03:23.325492 2984 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-kqx4k" Aug 13 00:03:23.328342 kubelet[2984]: E0813 00:03:23.328246 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-kqx4k_kube-system(48c5f0fe-d3a5-4b24-a4f6-448275db6b4d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-kqx4k_kube-system(48c5f0fe-d3a5-4b24-a4f6-448275db6b4d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-kqx4k" podUID="48c5f0fe-d3a5-4b24-a4f6-448275db6b4d" Aug 13 00:03:23.360307 env[1913]: time="2025-08-13T00:03:23.360234821Z" level=error msg="Failed to destroy network for sandbox \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:23.366696 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a-shm.mount: Deactivated successfully. Aug 13 00:03:23.370308 env[1913]: time="2025-08-13T00:03:23.370235902Z" level=error msg="encountered an error cleaning up failed sandbox \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:23.370552 env[1913]: time="2025-08-13T00:03:23.370501681Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-v6d44,Uid:ba5590d9-af5c-484d-9d4d-2e76ac5327b7,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:23.372948 kubelet[2984]: E0813 00:03:23.371634 2984 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:23.372948 kubelet[2984]: E0813 00:03:23.371710 2984 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-v6d44" Aug 13 00:03:23.372948 kubelet[2984]: E0813 00:03:23.371744 2984 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-v6d44" Aug 13 00:03:23.372948 kubelet[2984]: E0813 00:03:23.371804 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-v6d44_kube-system(ba5590d9-af5c-484d-9d4d-2e76ac5327b7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-v6d44_kube-system(ba5590d9-af5c-484d-9d4d-2e76ac5327b7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-v6d44" podUID="ba5590d9-af5c-484d-9d4d-2e76ac5327b7" Aug 13 00:03:23.933483 kubelet[2984]: I0813 00:03:23.933298 2984 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Aug 13 00:03:23.938207 env[1913]: time="2025-08-13T00:03:23.938155154Z" level=info msg="StopPodSandbox for \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\"" Aug 13 00:03:23.946873 kubelet[2984]: I0813 00:03:23.945021 2984 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Aug 13 00:03:23.947153 env[1913]: time="2025-08-13T00:03:23.945880415Z" level=info msg="StopPodSandbox for \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\"" Aug 13 00:03:24.038199 env[1913]: time="2025-08-13T00:03:24.038090027Z" level=error msg="StopPodSandbox for \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\" failed" error="failed to destroy network for sandbox \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:24.038556 kubelet[2984]: E0813 00:03:24.038463 2984 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Aug 13 00:03:24.038658 kubelet[2984]: E0813 00:03:24.038590 2984 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf"} Aug 13 00:03:24.038729 kubelet[2984]: E0813 00:03:24.038648 2984 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"48c5f0fe-d3a5-4b24-a4f6-448275db6b4d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:03:24.038952 kubelet[2984]: E0813 00:03:24.038716 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"48c5f0fe-d3a5-4b24-a4f6-448275db6b4d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-kqx4k" podUID="48c5f0fe-d3a5-4b24-a4f6-448275db6b4d" Aug 13 00:03:24.045067 env[1913]: time="2025-08-13T00:03:24.043424242Z" level=error msg="StopPodSandbox for \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\" failed" error="failed to destroy network for sandbox \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:03:24.045755 kubelet[2984]: E0813 00:03:24.045680 2984 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Aug 13 00:03:24.045872 kubelet[2984]: E0813 00:03:24.045775 2984 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a"} Aug 13 00:03:24.045937 kubelet[2984]: E0813 00:03:24.045884 2984 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"ba5590d9-af5c-484d-9d4d-2e76ac5327b7\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:03:24.046170 kubelet[2984]: E0813 00:03:24.045950 2984 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"ba5590d9-af5c-484d-9d4d-2e76ac5327b7\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-v6d44" podUID="ba5590d9-af5c-484d-9d4d-2e76ac5327b7" Aug 13 00:03:30.176700 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2565520714.mount: Deactivated successfully. Aug 13 00:03:30.248697 env[1913]: time="2025-08-13T00:03:30.248618522Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/node:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:30.254528 env[1913]: time="2025-08-13T00:03:30.254461116Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:1c6ddca599ddd18c061e797a7830b0aea985f8b023c5e43d815a9ed1088893a9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:30.258986 env[1913]: time="2025-08-13T00:03:30.258924406Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/node:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:30.262506 env[1913]: time="2025-08-13T00:03:30.262445426Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:30.263672 env[1913]: time="2025-08-13T00:03:30.263621195Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\" returns image reference \"sha256:1c6ddca599ddd18c061e797a7830b0aea985f8b023c5e43d815a9ed1088893a9\"" Aug 13 00:03:30.297875 env[1913]: time="2025-08-13T00:03:30.297808644Z" level=info msg="CreateContainer within sandbox \"d322b14568ffd3c041d38cb88942ebe7375d11f13c0964901f9cff59f92871ce\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Aug 13 00:03:30.326658 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2387638352.mount: Deactivated successfully. Aug 13 00:03:30.339235 env[1913]: time="2025-08-13T00:03:30.339170962Z" level=info msg="CreateContainer within sandbox \"d322b14568ffd3c041d38cb88942ebe7375d11f13c0964901f9cff59f92871ce\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"be0ed1e29b588d67e52fae942c6b3420ba4d8c77c6e0d8347717d2a5797a2a37\"" Aug 13 00:03:30.340977 env[1913]: time="2025-08-13T00:03:30.340496672Z" level=info msg="StartContainer for \"be0ed1e29b588d67e52fae942c6b3420ba4d8c77c6e0d8347717d2a5797a2a37\"" Aug 13 00:03:30.467383 env[1913]: time="2025-08-13T00:03:30.466787177Z" level=info msg="StartContainer for \"be0ed1e29b588d67e52fae942c6b3420ba4d8c77c6e0d8347717d2a5797a2a37\" returns successfully" Aug 13 00:03:30.738132 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Aug 13 00:03:30.738317 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Aug 13 00:03:31.059800 kubelet[2984]: I0813 00:03:31.059674 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-wvcf6" podStartSLOduration=1.748207824 podStartE2EDuration="21.059629752s" podCreationTimestamp="2025-08-13 00:03:10 +0000 UTC" firstStartedPulling="2025-08-13 00:03:10.954628146 +0000 UTC m=+43.701807066" lastFinishedPulling="2025-08-13 00:03:30.266050098 +0000 UTC m=+63.013228994" observedRunningTime="2025-08-13 00:03:31.057268194 +0000 UTC m=+63.804447126" watchObservedRunningTime="2025-08-13 00:03:31.059629752 +0000 UTC m=+63.806808672" Aug 13 00:03:31.219787 env[1913]: time="2025-08-13T00:03:31.219709663Z" level=info msg="StopPodSandbox for \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\"" Aug 13 00:03:31.587568 env[1913]: 2025-08-13 00:03:31.427 [INFO][4168] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Aug 13 00:03:31.587568 env[1913]: 2025-08-13 00:03:31.428 [INFO][4168] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" iface="eth0" netns="/var/run/netns/cni-dc67cee5-61d8-ed1a-371a-ef4af928e74b" Aug 13 00:03:31.587568 env[1913]: 2025-08-13 00:03:31.429 [INFO][4168] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" iface="eth0" netns="/var/run/netns/cni-dc67cee5-61d8-ed1a-371a-ef4af928e74b" Aug 13 00:03:31.587568 env[1913]: 2025-08-13 00:03:31.431 [INFO][4168] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" iface="eth0" netns="/var/run/netns/cni-dc67cee5-61d8-ed1a-371a-ef4af928e74b" Aug 13 00:03:31.587568 env[1913]: 2025-08-13 00:03:31.431 [INFO][4168] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Aug 13 00:03:31.587568 env[1913]: 2025-08-13 00:03:31.431 [INFO][4168] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Aug 13 00:03:31.587568 env[1913]: 2025-08-13 00:03:31.560 [INFO][4179] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" HandleID="k8s-pod-network.12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Workload="ip--172--31--27--151-k8s-whisker--755dcdc665--9mnng-eth0" Aug 13 00:03:31.587568 env[1913]: 2025-08-13 00:03:31.561 [INFO][4179] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:31.587568 env[1913]: 2025-08-13 00:03:31.561 [INFO][4179] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:31.587568 env[1913]: 2025-08-13 00:03:31.575 [WARNING][4179] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" HandleID="k8s-pod-network.12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Workload="ip--172--31--27--151-k8s-whisker--755dcdc665--9mnng-eth0" Aug 13 00:03:31.587568 env[1913]: 2025-08-13 00:03:31.575 [INFO][4179] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" HandleID="k8s-pod-network.12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Workload="ip--172--31--27--151-k8s-whisker--755dcdc665--9mnng-eth0" Aug 13 00:03:31.587568 env[1913]: 2025-08-13 00:03:31.579 [INFO][4179] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:31.587568 env[1913]: 2025-08-13 00:03:31.584 [INFO][4168] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Aug 13 00:03:31.589002 env[1913]: time="2025-08-13T00:03:31.587803263Z" level=info msg="TearDown network for sandbox \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\" successfully" Aug 13 00:03:31.589002 env[1913]: time="2025-08-13T00:03:31.587851203Z" level=info msg="StopPodSandbox for \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\" returns successfully" Aug 13 00:03:31.598040 systemd[1]: run-netns-cni\x2ddc67cee5\x2d61d8\x2ded1a\x2d371a\x2def4af928e74b.mount: Deactivated successfully. Aug 13 00:03:31.645131 kubelet[2984]: I0813 00:03:31.645061 2984 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/f6d611ea-61c9-46e6-9d86-1c24bd310146-whisker-backend-key-pair\") pod \"f6d611ea-61c9-46e6-9d86-1c24bd310146\" (UID: \"f6d611ea-61c9-46e6-9d86-1c24bd310146\") " Aug 13 00:03:31.645681 kubelet[2984]: I0813 00:03:31.645557 2984 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nlvr\" (UniqueName: \"kubernetes.io/projected/f6d611ea-61c9-46e6-9d86-1c24bd310146-kube-api-access-6nlvr\") pod \"f6d611ea-61c9-46e6-9d86-1c24bd310146\" (UID: \"f6d611ea-61c9-46e6-9d86-1c24bd310146\") " Aug 13 00:03:31.645795 kubelet[2984]: I0813 00:03:31.645724 2984 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6d611ea-61c9-46e6-9d86-1c24bd310146-whisker-ca-bundle\") pod \"f6d611ea-61c9-46e6-9d86-1c24bd310146\" (UID: \"f6d611ea-61c9-46e6-9d86-1c24bd310146\") " Aug 13 00:03:31.649434 kubelet[2984]: I0813 00:03:31.649338 2984 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6d611ea-61c9-46e6-9d86-1c24bd310146-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "f6d611ea-61c9-46e6-9d86-1c24bd310146" (UID: "f6d611ea-61c9-46e6-9d86-1c24bd310146"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Aug 13 00:03:31.667407 systemd[1]: var-lib-kubelet-pods-f6d611ea\x2d61c9\x2d46e6\x2d9d86\x2d1c24bd310146-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d6nlvr.mount: Deactivated successfully. Aug 13 00:03:31.671529 kubelet[2984]: I0813 00:03:31.671481 2984 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6d611ea-61c9-46e6-9d86-1c24bd310146-kube-api-access-6nlvr" (OuterVolumeSpecName: "kube-api-access-6nlvr") pod "f6d611ea-61c9-46e6-9d86-1c24bd310146" (UID: "f6d611ea-61c9-46e6-9d86-1c24bd310146"). InnerVolumeSpecName "kube-api-access-6nlvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Aug 13 00:03:31.686339 systemd[1]: var-lib-kubelet-pods-f6d611ea\x2d61c9\x2d46e6\x2d9d86\x2d1c24bd310146-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Aug 13 00:03:31.691415 kubelet[2984]: I0813 00:03:31.691317 2984 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6d611ea-61c9-46e6-9d86-1c24bd310146-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "f6d611ea-61c9-46e6-9d86-1c24bd310146" (UID: "f6d611ea-61c9-46e6-9d86-1c24bd310146"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGidValue "" Aug 13 00:03:31.746792 kubelet[2984]: I0813 00:03:31.746731 2984 reconciler_common.go:293] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/f6d611ea-61c9-46e6-9d86-1c24bd310146-whisker-backend-key-pair\") on node \"ip-172-31-27-151\" DevicePath \"\"" Aug 13 00:03:31.746950 kubelet[2984]: I0813 00:03:31.746800 2984 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nlvr\" (UniqueName: \"kubernetes.io/projected/f6d611ea-61c9-46e6-9d86-1c24bd310146-kube-api-access-6nlvr\") on node \"ip-172-31-27-151\" DevicePath \"\"" Aug 13 00:03:31.746950 kubelet[2984]: I0813 00:03:31.746827 2984 reconciler_common.go:293] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6d611ea-61c9-46e6-9d86-1c24bd310146-whisker-ca-bundle\") on node \"ip-172-31-27-151\" DevicePath \"\"" Aug 13 00:03:32.148551 kubelet[2984]: I0813 00:03:32.148487 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/fe1fd988-a91d-488e-9bc2-95e088424fc2-whisker-backend-key-pair\") pod \"whisker-66f7b8977f-mqv9d\" (UID: \"fe1fd988-a91d-488e-9bc2-95e088424fc2\") " pod="calico-system/whisker-66f7b8977f-mqv9d" Aug 13 00:03:32.149440 kubelet[2984]: I0813 00:03:32.149386 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6nlh\" (UniqueName: \"kubernetes.io/projected/fe1fd988-a91d-488e-9bc2-95e088424fc2-kube-api-access-r6nlh\") pod \"whisker-66f7b8977f-mqv9d\" (UID: \"fe1fd988-a91d-488e-9bc2-95e088424fc2\") " pod="calico-system/whisker-66f7b8977f-mqv9d" Aug 13 00:03:32.149682 kubelet[2984]: I0813 00:03:32.149633 2984 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fe1fd988-a91d-488e-9bc2-95e088424fc2-whisker-ca-bundle\") pod \"whisker-66f7b8977f-mqv9d\" (UID: \"fe1fd988-a91d-488e-9bc2-95e088424fc2\") " pod="calico-system/whisker-66f7b8977f-mqv9d" Aug 13 00:03:32.182874 systemd[1]: run-containerd-runc-k8s.io-be0ed1e29b588d67e52fae942c6b3420ba4d8c77c6e0d8347717d2a5797a2a37-runc.CPUY2t.mount: Deactivated successfully. Aug 13 00:03:32.417207 env[1913]: time="2025-08-13T00:03:32.417087722Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-66f7b8977f-mqv9d,Uid:fe1fd988-a91d-488e-9bc2-95e088424fc2,Namespace:calico-system,Attempt:0,}" Aug 13 00:03:32.639790 systemd-networkd[1587]: cali6498075399c: Link UP Aug 13 00:03:32.642252 (udev-worker)[4243]: Network interface NamePolicy= disabled on kernel command line. Aug 13 00:03:32.650199 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Aug 13 00:03:32.650342 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cali6498075399c: link becomes ready Aug 13 00:03:32.648647 systemd-networkd[1587]: cali6498075399c: Gained carrier Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.485 [INFO][4223] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.509 [INFO][4223] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--27--151-k8s-whisker--66f7b8977f--mqv9d-eth0 whisker-66f7b8977f- calico-system fe1fd988-a91d-488e-9bc2-95e088424fc2 963 0 2025-08-13 00:03:32 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:66f7b8977f projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ip-172-31-27-151 whisker-66f7b8977f-mqv9d eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali6498075399c [] [] }} ContainerID="e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" Namespace="calico-system" Pod="whisker-66f7b8977f-mqv9d" WorkloadEndpoint="ip--172--31--27--151-k8s-whisker--66f7b8977f--mqv9d-" Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.509 [INFO][4223] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" Namespace="calico-system" Pod="whisker-66f7b8977f-mqv9d" WorkloadEndpoint="ip--172--31--27--151-k8s-whisker--66f7b8977f--mqv9d-eth0" Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.557 [INFO][4236] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" HandleID="k8s-pod-network.e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" Workload="ip--172--31--27--151-k8s-whisker--66f7b8977f--mqv9d-eth0" Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.557 [INFO][4236] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" HandleID="k8s-pod-network.e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" Workload="ip--172--31--27--151-k8s-whisker--66f7b8977f--mqv9d-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002c10a0), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-27-151", "pod":"whisker-66f7b8977f-mqv9d", "timestamp":"2025-08-13 00:03:32.557226665 +0000 UTC"}, Hostname:"ip-172-31-27-151", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.557 [INFO][4236] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.557 [INFO][4236] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.557 [INFO][4236] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-27-151' Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.571 [INFO][4236] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" host="ip-172-31-27-151" Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.580 [INFO][4236] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-27-151" Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.588 [INFO][4236] ipam/ipam.go 511: Trying affinity for 192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.591 [INFO][4236] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.596 [INFO][4236] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.596 [INFO][4236] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.64/26 handle="k8s-pod-network.e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" host="ip-172-31-27-151" Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.600 [INFO][4236] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.608 [INFO][4236] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.64/26 handle="k8s-pod-network.e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" host="ip-172-31-27-151" Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.617 [INFO][4236] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.65/26] block=192.168.1.64/26 handle="k8s-pod-network.e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" host="ip-172-31-27-151" Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.619 [INFO][4236] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.65/26] handle="k8s-pod-network.e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" host="ip-172-31-27-151" Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.619 [INFO][4236] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:32.677787 env[1913]: 2025-08-13 00:03:32.619 [INFO][4236] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.65/26] IPv6=[] ContainerID="e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" HandleID="k8s-pod-network.e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" Workload="ip--172--31--27--151-k8s-whisker--66f7b8977f--mqv9d-eth0" Aug 13 00:03:32.679782 env[1913]: 2025-08-13 00:03:32.623 [INFO][4223] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" Namespace="calico-system" Pod="whisker-66f7b8977f-mqv9d" WorkloadEndpoint="ip--172--31--27--151-k8s-whisker--66f7b8977f--mqv9d-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-whisker--66f7b8977f--mqv9d-eth0", GenerateName:"whisker-66f7b8977f-", Namespace:"calico-system", SelfLink:"", UID:"fe1fd988-a91d-488e-9bc2-95e088424fc2", ResourceVersion:"963", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 32, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"66f7b8977f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"", Pod:"whisker-66f7b8977f-mqv9d", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.1.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali6498075399c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:32.679782 env[1913]: 2025-08-13 00:03:32.623 [INFO][4223] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.65/32] ContainerID="e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" Namespace="calico-system" Pod="whisker-66f7b8977f-mqv9d" WorkloadEndpoint="ip--172--31--27--151-k8s-whisker--66f7b8977f--mqv9d-eth0" Aug 13 00:03:32.679782 env[1913]: 2025-08-13 00:03:32.623 [INFO][4223] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6498075399c ContainerID="e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" Namespace="calico-system" Pod="whisker-66f7b8977f-mqv9d" WorkloadEndpoint="ip--172--31--27--151-k8s-whisker--66f7b8977f--mqv9d-eth0" Aug 13 00:03:32.679782 env[1913]: 2025-08-13 00:03:32.649 [INFO][4223] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" Namespace="calico-system" Pod="whisker-66f7b8977f-mqv9d" WorkloadEndpoint="ip--172--31--27--151-k8s-whisker--66f7b8977f--mqv9d-eth0" Aug 13 00:03:32.679782 env[1913]: 2025-08-13 00:03:32.650 [INFO][4223] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" Namespace="calico-system" Pod="whisker-66f7b8977f-mqv9d" WorkloadEndpoint="ip--172--31--27--151-k8s-whisker--66f7b8977f--mqv9d-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-whisker--66f7b8977f--mqv9d-eth0", GenerateName:"whisker-66f7b8977f-", Namespace:"calico-system", SelfLink:"", UID:"fe1fd988-a91d-488e-9bc2-95e088424fc2", ResourceVersion:"963", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 32, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"66f7b8977f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e", Pod:"whisker-66f7b8977f-mqv9d", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.1.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali6498075399c", MAC:"2e:ec:51:a2:f1:13", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:32.679782 env[1913]: 2025-08-13 00:03:32.671 [INFO][4223] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e" Namespace="calico-system" Pod="whisker-66f7b8977f-mqv9d" WorkloadEndpoint="ip--172--31--27--151-k8s-whisker--66f7b8977f--mqv9d-eth0" Aug 13 00:03:32.700188 env[1913]: time="2025-08-13T00:03:32.700045803Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:03:32.700476 env[1913]: time="2025-08-13T00:03:32.700416114Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:03:32.700647 env[1913]: time="2025-08-13T00:03:32.700592515Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:03:32.701278 env[1913]: time="2025-08-13T00:03:32.701211120Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e pid=4258 runtime=io.containerd.runc.v2 Aug 13 00:03:32.861798 env[1913]: time="2025-08-13T00:03:32.861623302Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-66f7b8977f-mqv9d,Uid:fe1fd988-a91d-488e-9bc2-95e088424fc2,Namespace:calico-system,Attempt:0,} returns sandbox id \"e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e\"" Aug 13 00:03:32.862000 audit[4311]: AVC avc: denied { write } for pid=4311 comm="tee" name="fd" dev="proc" ino=22185 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 Aug 13 00:03:32.874299 env[1913]: time="2025-08-13T00:03:32.874236477Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\"" Aug 13 00:03:32.862000 audit[4311]: SYSCALL arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=ffffecc297c0 a2=241 a3=1b6 items=1 ppid=4298 pid=4311 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:32.890345 kernel: audit: type=1400 audit(1755043412.862:293): avc: denied { write } for pid=4311 comm="tee" name="fd" dev="proc" ino=22185 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 Aug 13 00:03:32.890561 kernel: audit: type=1300 audit(1755043412.862:293): arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=ffffecc297c0 a2=241 a3=1b6 items=1 ppid=4298 pid=4311 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:32.862000 audit: CWD cwd="/etc/service/enabled/node-status-reporter/log" Aug 13 00:03:32.897836 kernel: audit: type=1307 audit(1755043412.862:293): cwd="/etc/service/enabled/node-status-reporter/log" Aug 13 00:03:32.862000 audit: PATH item=0 name="/dev/fd/63" inode=22851 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 00:03:32.916923 kernel: audit: type=1302 audit(1755043412.862:293): item=0 name="/dev/fd/63" inode=22851 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 00:03:32.862000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 Aug 13 00:03:32.939757 kernel: audit: type=1327 audit(1755043412.862:293): proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 Aug 13 00:03:32.969000 audit[4358]: AVC avc: denied { write } for pid=4358 comm="tee" name="fd" dev="proc" ino=22217 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 Aug 13 00:03:32.969000 audit[4358]: SYSCALL arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=fffff93be7d0 a2=241 a3=1b6 items=1 ppid=4302 pid=4358 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:32.994893 kernel: audit: type=1400 audit(1755043412.969:294): avc: denied { write } for pid=4358 comm="tee" name="fd" dev="proc" ino=22217 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 Aug 13 00:03:32.995051 kernel: audit: type=1300 audit(1755043412.969:294): arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=fffff93be7d0 a2=241 a3=1b6 items=1 ppid=4302 pid=4358 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:32.969000 audit: CWD cwd="/etc/service/enabled/bird/log" Aug 13 00:03:33.000077 kernel: audit: type=1307 audit(1755043412.969:294): cwd="/etc/service/enabled/bird/log" Aug 13 00:03:32.969000 audit: PATH item=0 name="/dev/fd/63" inode=22195 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 00:03:33.008877 kernel: audit: type=1302 audit(1755043412.969:294): item=0 name="/dev/fd/63" inode=22195 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 00:03:32.969000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 Aug 13 00:03:33.018209 kernel: audit: type=1327 audit(1755043412.969:294): proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 Aug 13 00:03:33.018000 audit[4364]: AVC avc: denied { write } for pid=4364 comm="tee" name="fd" dev="proc" ino=22221 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 Aug 13 00:03:33.018000 audit[4364]: SYSCALL arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=ffffcd50f7cf a2=241 a3=1b6 items=1 ppid=4306 pid=4364 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.018000 audit: CWD cwd="/etc/service/enabled/confd/log" Aug 13 00:03:33.018000 audit: PATH item=0 name="/dev/fd/63" inode=22211 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 00:03:33.018000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 Aug 13 00:03:33.023000 audit[4362]: AVC avc: denied { write } for pid=4362 comm="tee" name="fd" dev="proc" ino=22225 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 Aug 13 00:03:33.023000 audit[4362]: SYSCALL arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=ffffe4aca7d1 a2=241 a3=1b6 items=1 ppid=4304 pid=4362 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.023000 audit: CWD cwd="/etc/service/enabled/cni/log" Aug 13 00:03:33.023000 audit: PATH item=0 name="/dev/fd/63" inode=22210 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 00:03:33.023000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 Aug 13 00:03:33.048000 audit[4368]: AVC avc: denied { write } for pid=4368 comm="tee" name="fd" dev="proc" ino=22915 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 Aug 13 00:03:33.048000 audit[4368]: SYSCALL arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=ffffd105e7cf a2=241 a3=1b6 items=1 ppid=4300 pid=4368 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.048000 audit: CWD cwd="/etc/service/enabled/bird6/log" Aug 13 00:03:33.048000 audit: PATH item=0 name="/dev/fd/63" inode=22213 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 00:03:33.048000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 Aug 13 00:03:33.051000 audit[4366]: AVC avc: denied { write } for pid=4366 comm="tee" name="fd" dev="proc" ino=22917 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 Aug 13 00:03:33.051000 audit[4366]: SYSCALL arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=ffffe926e7cf a2=241 a3=1b6 items=1 ppid=4296 pid=4366 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.051000 audit: CWD cwd="/etc/service/enabled/felix/log" Aug 13 00:03:33.051000 audit: PATH item=0 name="/dev/fd/63" inode=22212 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 00:03:33.051000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 Aug 13 00:03:33.056000 audit[4370]: AVC avc: denied { write } for pid=4370 comm="tee" name="fd" dev="proc" ino=22229 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 Aug 13 00:03:33.056000 audit[4370]: SYSCALL arch=c00000b7 syscall=56 success=yes exit=3 a0=ffffffffffffff9c a1=ffffc0a1a7bf a2=241 a3=1b6 items=1 ppid=4305 pid=4370 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.056000 audit: CWD cwd="/etc/service/enabled/allocate-tunnel-addrs/log" Aug 13 00:03:33.056000 audit: PATH item=0 name="/dev/fd/63" inode=22214 dev=00:0b mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Aug 13 00:03:33.056000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 Aug 13 00:03:33.480091 kubelet[2984]: I0813 00:03:33.480020 2984 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6d611ea-61c9-46e6-9d86-1c24bd310146" path="/var/lib/kubelet/pods/f6d611ea-61c9-46e6-9d86-1c24bd310146/volumes" Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { bpf } for pid=4412 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { bpf } for pid=4412 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { bpf } for pid=4412 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { bpf } for pid=4412 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit: BPF prog-id=10 op=LOAD Aug 13 00:03:33.719000 audit[4412]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=ffffe0f6e868 a2=98 a3=ffffe0f6e858 items=0 ppid=4310 pid=4412 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.719000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Aug 13 00:03:33.719000 audit: BPF prog-id=10 op=UNLOAD Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { bpf } for pid=4412 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { bpf } for pid=4412 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { bpf } for pid=4412 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { bpf } for pid=4412 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit: BPF prog-id=11 op=LOAD Aug 13 00:03:33.719000 audit[4412]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=ffffe0f6e718 a2=74 a3=95 items=0 ppid=4310 pid=4412 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.719000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Aug 13 00:03:33.719000 audit: BPF prog-id=11 op=UNLOAD Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { bpf } for pid=4412 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { bpf } for pid=4412 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { bpf } for pid=4412 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit[4412]: AVC avc: denied { bpf } for pid=4412 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.719000 audit: BPF prog-id=12 op=LOAD Aug 13 00:03:33.719000 audit[4412]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=ffffe0f6e748 a2=40 a3=ffffe0f6e778 items=0 ppid=4310 pid=4412 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.719000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Aug 13 00:03:33.720000 audit: BPF prog-id=12 op=UNLOAD Aug 13 00:03:33.720000 audit[4412]: AVC avc: denied { perfmon } for pid=4412 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.720000 audit[4412]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=0 a1=ffffe0f6e860 a2=50 a3=0 items=0 ppid=4310 pid=4412 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.720000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F74632F676C6F62616C732F63616C695F63746C625F70726F677300747970650070726F675F6172726179006B657900340076616C7565003400656E74726965730033006E616D650063616C695F63746C625F70726F677300666C6167730030 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit: BPF prog-id=13 op=LOAD Aug 13 00:03:33.725000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=ffffe4523678 a2=98 a3=ffffe4523668 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.725000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.725000 audit: BPF prog-id=13 op=UNLOAD Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit: BPF prog-id=14 op=LOAD Aug 13 00:03:33.725000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=5 a1=ffffe4523308 a2=74 a3=95 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.725000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.725000 audit: BPF prog-id=14 op=UNLOAD Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.725000 audit: BPF prog-id=15 op=LOAD Aug 13 00:03:33.725000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=5 a1=ffffe4523368 a2=94 a3=2 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.725000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.725000 audit: BPF prog-id=15 op=UNLOAD Aug 13 00:03:33.762382 systemd-networkd[1587]: cali6498075399c: Gained IPv6LL Aug 13 00:03:33.921000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.921000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.921000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.921000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.921000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.921000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.921000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.921000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.921000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.921000 audit: BPF prog-id=16 op=LOAD Aug 13 00:03:33.921000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=5 a1=ffffe4523328 a2=40 a3=ffffe4523358 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.921000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.921000 audit: BPF prog-id=16 op=UNLOAD Aug 13 00:03:33.921000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.921000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=0 a1=ffffe4523440 a2=50 a3=0 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.921000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.956000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.956000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=ffffe4523398 a2=28 a3=ffffe45234c8 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.956000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.956000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.956000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=ffffe45233c8 a2=28 a3=ffffe45234f8 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.956000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.957000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.957000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=ffffe4523278 a2=28 a3=ffffe45233a8 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.957000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.957000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.957000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=ffffe45233e8 a2=28 a3=ffffe4523518 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.957000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=ffffe45233c8 a2=28 a3=ffffe45234f8 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.958000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=ffffe45233b8 a2=28 a3=ffffe45234e8 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.958000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=ffffe45233e8 a2=28 a3=ffffe4523518 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.958000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=ffffe45233c8 a2=28 a3=ffffe45234f8 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.958000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=ffffe45233e8 a2=28 a3=ffffe4523518 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.958000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=ffffe45233b8 a2=28 a3=ffffe45234e8 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.958000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=ffffe4523438 a2=28 a3=ffffe4523578 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.958000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=5 a0=0 a1=ffffe4523170 a2=50 a3=0 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.958000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.958000 audit: BPF prog-id=17 op=LOAD Aug 13 00:03:33.958000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=6 a0=5 a1=ffffe4523178 a2=94 a3=5 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.958000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.959000 audit: BPF prog-id=17 op=UNLOAD Aug 13 00:03:33.959000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.959000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=5 a0=0 a1=ffffe4523280 a2=50 a3=0 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.959000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.959000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.959000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=0 a0=16 a1=ffffe45233c8 a2=4 a3=3 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.959000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.959000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.959000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.959000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.959000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.959000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.959000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.959000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.959000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.959000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.959000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.959000 audit[4413]: AVC avc: denied { confidentiality } for pid=4413 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 Aug 13 00:03:33.959000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=5 a1=ffffe45233a8 a2=94 a3=6 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.959000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.960000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.960000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.960000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.960000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.960000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.960000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.960000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.960000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.960000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.960000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.960000 audit[4413]: AVC avc: denied { confidentiality } for pid=4413 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 Aug 13 00:03:33.960000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=5 a1=ffffe4522b78 a2=94 a3=83 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.960000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.965000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.965000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.965000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.965000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.965000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.965000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.965000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.965000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.965000 audit[4413]: AVC avc: denied { perfmon } for pid=4413 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.965000 audit[4413]: AVC avc: denied { bpf } for pid=4413 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.965000 audit[4413]: AVC avc: denied { confidentiality } for pid=4413 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 Aug 13 00:03:33.965000 audit[4413]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=5 a1=ffffe4522b78 a2=94 a3=83 items=0 ppid=4310 pid=4413 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.965000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E Aug 13 00:03:33.988000 audit[4418]: AVC avc: denied { bpf } for pid=4418 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.988000 audit[4418]: AVC avc: denied { bpf } for pid=4418 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.988000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.988000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.988000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.988000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.988000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.988000 audit[4418]: AVC avc: denied { bpf } for pid=4418 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.988000 audit[4418]: AVC avc: denied { bpf } for pid=4418 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.988000 audit: BPF prog-id=18 op=LOAD Aug 13 00:03:33.988000 audit[4418]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=ffffe1514ab8 a2=98 a3=ffffe1514aa8 items=0 ppid=4310 pid=4418 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.988000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Aug 13 00:03:33.990000 audit: BPF prog-id=18 op=UNLOAD Aug 13 00:03:33.990000 audit[4418]: AVC avc: denied { bpf } for pid=4418 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.990000 audit[4418]: AVC avc: denied { bpf } for pid=4418 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.990000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.990000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.990000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.990000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.990000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.990000 audit[4418]: AVC avc: denied { bpf } for pid=4418 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.990000 audit[4418]: AVC avc: denied { bpf } for pid=4418 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.990000 audit: BPF prog-id=19 op=LOAD Aug 13 00:03:33.990000 audit[4418]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=ffffe1514968 a2=74 a3=95 items=0 ppid=4310 pid=4418 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.990000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Aug 13 00:03:33.991000 audit: BPF prog-id=19 op=UNLOAD Aug 13 00:03:33.991000 audit[4418]: AVC avc: denied { bpf } for pid=4418 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.991000 audit[4418]: AVC avc: denied { bpf } for pid=4418 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.991000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.991000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.991000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.991000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.991000 audit[4418]: AVC avc: denied { perfmon } for pid=4418 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.991000 audit[4418]: AVC avc: denied { bpf } for pid=4418 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.991000 audit[4418]: AVC avc: denied { bpf } for pid=4418 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:33.991000 audit: BPF prog-id=20 op=LOAD Aug 13 00:03:33.991000 audit[4418]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=ffffe1514998 a2=40 a3=ffffe15149c8 items=0 ppid=4310 pid=4418 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:33.991000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F Aug 13 00:03:33.994000 audit: BPF prog-id=20 op=UNLOAD Aug 13 00:03:34.159851 systemd-networkd[1587]: vxlan.calico: Link UP Aug 13 00:03:34.159876 systemd-networkd[1587]: vxlan.calico: Gained carrier Aug 13 00:03:34.220792 (udev-worker)[4132]: Network interface NamePolicy= disabled on kernel command line. Aug 13 00:03:34.224000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.224000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.224000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.224000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.224000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.224000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.224000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.224000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.224000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.224000 audit: BPF prog-id=21 op=LOAD Aug 13 00:03:34.224000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=fffff1806c68 a2=98 a3=fffff1806c58 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.224000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.225000 audit: BPF prog-id=21 op=UNLOAD Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit: BPF prog-id=22 op=LOAD Aug 13 00:03:34.225000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=fffff1806948 a2=74 a3=95 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.225000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.225000 audit: BPF prog-id=22 op=UNLOAD Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit: BPF prog-id=23 op=LOAD Aug 13 00:03:34.225000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=fffff18069a8 a2=94 a3=2 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.225000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.225000 audit: BPF prog-id=23 op=UNLOAD Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=12 a1=fffff18069d8 a2=28 a3=fffff1806b08 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.225000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff1806a08 a2=28 a3=fffff1806b38 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.225000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff18068b8 a2=28 a3=fffff18069e8 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.225000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=12 a1=fffff1806a28 a2=28 a3=fffff1806b58 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.225000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=12 a1=fffff1806a08 a2=28 a3=fffff1806b38 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.225000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=12 a1=fffff18069f8 a2=28 a3=fffff1806b28 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.225000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=12 a1=fffff1806a28 a2=28 a3=fffff1806b58 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.225000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.225000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.225000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff1806a08 a2=28 a3=fffff1806b38 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.225000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.226000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.226000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff1806a28 a2=28 a3=fffff1806b58 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.226000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.226000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.226000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=fffff18069f8 a2=28 a3=fffff1806b28 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.226000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.226000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.226000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=12 a1=fffff1806a78 a2=28 a3=fffff1806bb8 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.226000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.226000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.226000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.226000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.226000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.226000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.226000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.226000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.226000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.226000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.226000 audit: BPF prog-id=24 op=LOAD Aug 13 00:03:34.226000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=6 a0=5 a1=fffff1806898 a2=40 a3=fffff18068c8 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.226000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.226000 audit: BPF prog-id=24 op=UNLOAD Aug 13 00:03:34.226000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.226000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=0 a1=fffff18068c0 a2=50 a3=0 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.226000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=6 a0=0 a1=fffff18068c0 a2=50 a3=0 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.231000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit: BPF prog-id=25 op=LOAD Aug 13 00:03:34.231000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=6 a0=5 a1=fffff1806028 a2=94 a3=2 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.231000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.231000 audit: BPF prog-id=25 op=UNLOAD Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { perfmon } for pid=4442 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit[4442]: AVC avc: denied { bpf } for pid=4442 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.231000 audit: BPF prog-id=26 op=LOAD Aug 13 00:03:34.231000 audit[4442]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=6 a0=5 a1=fffff18061b8 a2=94 a3=30 items=0 ppid=4310 pid=4442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.231000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit: BPF prog-id=27 op=LOAD Aug 13 00:03:34.243000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=3 a0=5 a1=ffffd560a0d8 a2=98 a3=ffffd560a0c8 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.243000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.243000 audit: BPF prog-id=27 op=UNLOAD Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit: BPF prog-id=28 op=LOAD Aug 13 00:03:34.243000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=5 a1=ffffd5609d68 a2=74 a3=95 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.243000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.243000 audit: BPF prog-id=28 op=UNLOAD Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.243000 audit: BPF prog-id=29 op=LOAD Aug 13 00:03:34.243000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=5 a1=ffffd5609dc8 a2=94 a3=2 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.243000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.244000 audit: BPF prog-id=29 op=UNLOAD Aug 13 00:03:34.291060 env[1913]: time="2025-08-13T00:03:34.290962338Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/whisker:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:34.297053 env[1913]: time="2025-08-13T00:03:34.295900589Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:309942601a9ca6c4e92bcd09162824fef1c137a5c5d92fbbb45be0f29bfd1817,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:34.301365 env[1913]: time="2025-08-13T00:03:34.301286696Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/whisker:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:34.305915 env[1913]: time="2025-08-13T00:03:34.305852117Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:34.309265 env[1913]: time="2025-08-13T00:03:34.307740486Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\" returns image reference \"sha256:309942601a9ca6c4e92bcd09162824fef1c137a5c5d92fbbb45be0f29bfd1817\"" Aug 13 00:03:34.318446 env[1913]: time="2025-08-13T00:03:34.317856751Z" level=info msg="CreateContainer within sandbox \"e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Aug 13 00:03:34.352348 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3010791050.mount: Deactivated successfully. Aug 13 00:03:34.361049 env[1913]: time="2025-08-13T00:03:34.360971957Z" level=info msg="CreateContainer within sandbox \"e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"fb422fcb0f4cd1c1295e0b5c8639e9bf1133cb6a4425f7e45138e805b239b07d\"" Aug 13 00:03:34.364529 env[1913]: time="2025-08-13T00:03:34.364463658Z" level=info msg="StartContainer for \"fb422fcb0f4cd1c1295e0b5c8639e9bf1133cb6a4425f7e45138e805b239b07d\"" Aug 13 00:03:34.476910 env[1913]: time="2025-08-13T00:03:34.476758654Z" level=info msg="StopPodSandbox for \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\"" Aug 13 00:03:34.477820 env[1913]: time="2025-08-13T00:03:34.476781058Z" level=info msg="StopPodSandbox for \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\"" Aug 13 00:03:34.514000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.514000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.514000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.514000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.514000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.514000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.514000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.514000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.514000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.514000 audit: BPF prog-id=30 op=LOAD Aug 13 00:03:34.514000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=5 a1=ffffd5609d88 a2=40 a3=ffffd5609db8 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.514000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.515000 audit: BPF prog-id=30 op=UNLOAD Aug 13 00:03:34.515000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.515000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=0 a1=ffffd5609ea0 a2=50 a3=0 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.515000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=ffffd5609df8 a2=28 a3=ffffd5609f28 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.533000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=ffffd5609e28 a2=28 a3=ffffd5609f58 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.533000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=ffffd5609cd8 a2=28 a3=ffffd5609e08 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.533000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=ffffd5609e48 a2=28 a3=ffffd5609f78 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.533000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=ffffd5609e28 a2=28 a3=ffffd5609f58 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.533000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=ffffd5609e18 a2=28 a3=ffffd5609f48 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.533000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=ffffd5609e48 a2=28 a3=ffffd5609f78 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.533000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=ffffd5609e28 a2=28 a3=ffffd5609f58 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.533000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=ffffd5609e48 a2=28 a3=ffffd5609f78 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.533000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=12 a1=ffffd5609e18 a2=28 a3=ffffd5609f48 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.533000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=4 a0=12 a1=ffffd5609e98 a2=28 a3=ffffd5609fd8 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.533000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=5 a0=0 a1=ffffd5609bd0 a2=50 a3=0 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.533000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.533000 audit: BPF prog-id=31 op=LOAD Aug 13 00:03:34.533000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=6 a0=5 a1=ffffd5609bd8 a2=94 a3=5 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.533000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.534000 audit: BPF prog-id=31 op=UNLOAD Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=5 a0=0 a1=ffffd5609ce0 a2=50 a3=0 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.534000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=0 a0=16 a1=ffffd5609e28 a2=4 a3=3 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.534000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { confidentiality } for pid=4446 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 Aug 13 00:03:34.534000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=5 a1=ffffd5609e08 a2=94 a3=6 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.534000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.534000 audit[4446]: AVC avc: denied { confidentiality } for pid=4446 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 Aug 13 00:03:34.534000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=5 a1=ffffd56095d8 a2=94 a3=83 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.534000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.535000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.535000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.535000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.535000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.535000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.535000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.535000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.535000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.535000 audit[4446]: AVC avc: denied { perfmon } for pid=4446 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.535000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.535000 audit[4446]: AVC avc: denied { confidentiality } for pid=4446 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 Aug 13 00:03:34.535000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=no exit=-22 a0=5 a1=ffffd56095d8 a2=94 a3=83 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.535000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.535000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.535000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=0 a0=f a1=ffffd560b018 a2=10 a3=ffffd560b108 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.535000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.536000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.536000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=0 a0=f a1=ffffd560aed8 a2=10 a3=ffffd560afc8 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.536000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.536000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.536000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=0 a0=f a1=ffffd560ae48 a2=10 a3=ffffd560afc8 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.536000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.536000 audit[4446]: AVC avc: denied { bpf } for pid=4446 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 Aug 13 00:03:34.536000 audit[4446]: SYSCALL arch=c00000b7 syscall=280 success=yes exit=0 a0=f a1=ffffd560ae48 a2=10 a3=ffffd560afc8 items=0 ppid=4310 pid=4446 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.536000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 Aug 13 00:03:34.548000 audit: BPF prog-id=26 op=UNLOAD Aug 13 00:03:34.723791 env[1913]: time="2025-08-13T00:03:34.723532833Z" level=info msg="StartContainer for \"fb422fcb0f4cd1c1295e0b5c8639e9bf1133cb6a4425f7e45138e805b239b07d\" returns successfully" Aug 13 00:03:34.734929 env[1913]: time="2025-08-13T00:03:34.734771682Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\"" Aug 13 00:03:34.920000 audit[4550]: NETFILTER_CFG table=mangle:101 family=2 entries=16 op=nft_register_chain pid=4550 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Aug 13 00:03:34.920000 audit[4550]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=6868 a0=3 a1=ffffe5f9af80 a2=0 a3=ffffa06f1fa8 items=0 ppid=4310 pid=4550 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.920000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Aug 13 00:03:34.961000 audit[4551]: NETFILTER_CFG table=nat:102 family=2 entries=15 op=nft_register_chain pid=4551 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Aug 13 00:03:34.965870 env[1913]: 2025-08-13 00:03:34.808 [INFO][4488] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Aug 13 00:03:34.965870 env[1913]: 2025-08-13 00:03:34.809 [INFO][4488] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" iface="eth0" netns="/var/run/netns/cni-b570b55c-ca42-0042-8b52-4c219d1eea55" Aug 13 00:03:34.965870 env[1913]: 2025-08-13 00:03:34.809 [INFO][4488] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" iface="eth0" netns="/var/run/netns/cni-b570b55c-ca42-0042-8b52-4c219d1eea55" Aug 13 00:03:34.965870 env[1913]: 2025-08-13 00:03:34.812 [INFO][4488] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" iface="eth0" netns="/var/run/netns/cni-b570b55c-ca42-0042-8b52-4c219d1eea55" Aug 13 00:03:34.965870 env[1913]: 2025-08-13 00:03:34.812 [INFO][4488] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Aug 13 00:03:34.965870 env[1913]: 2025-08-13 00:03:34.812 [INFO][4488] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Aug 13 00:03:34.965870 env[1913]: 2025-08-13 00:03:34.924 [INFO][4534] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" HandleID="k8s-pod-network.b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Workload="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:03:34.965870 env[1913]: 2025-08-13 00:03:34.925 [INFO][4534] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:34.965870 env[1913]: 2025-08-13 00:03:34.925 [INFO][4534] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:34.965870 env[1913]: 2025-08-13 00:03:34.948 [WARNING][4534] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" HandleID="k8s-pod-network.b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Workload="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:03:34.965870 env[1913]: 2025-08-13 00:03:34.948 [INFO][4534] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" HandleID="k8s-pod-network.b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Workload="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:03:34.965870 env[1913]: 2025-08-13 00:03:34.952 [INFO][4534] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:34.965870 env[1913]: 2025-08-13 00:03:34.959 [INFO][4488] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Aug 13 00:03:34.975976 systemd[1]: run-netns-cni\x2db570b55c\x2dca42\x2d0042\x2d8b52\x2d4c219d1eea55.mount: Deactivated successfully. Aug 13 00:03:34.978000 audit[4549]: NETFILTER_CFG table=raw:103 family=2 entries=21 op=nft_register_chain pid=4549 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Aug 13 00:03:34.981464 env[1913]: time="2025-08-13T00:03:34.981406503Z" level=info msg="TearDown network for sandbox \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\" successfully" Aug 13 00:03:34.981632 env[1913]: time="2025-08-13T00:03:34.981597185Z" level=info msg="StopPodSandbox for \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\" returns successfully" Aug 13 00:03:34.983047 env[1913]: time="2025-08-13T00:03:34.982995723Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-6tzfz,Uid:cf609856-35b0-4517-9fba-45b28ef2be55,Namespace:calico-system,Attempt:1,}" Aug 13 00:03:34.978000 audit[4549]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=8452 a0=3 a1=ffffd46bff30 a2=0 a3=ffff85626fa8 items=0 ppid=4310 pid=4549 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.978000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Aug 13 00:03:34.961000 audit[4551]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=5084 a0=3 a1=ffffee963e60 a2=0 a3=ffffa29fafa8 items=0 ppid=4310 pid=4551 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:34.961000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Aug 13 00:03:35.021000 audit[4559]: NETFILTER_CFG table=filter:104 family=2 entries=94 op=nft_register_chain pid=4559 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Aug 13 00:03:35.021000 audit[4559]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=53116 a0=3 a1=ffffea7910b0 a2=0 a3=ffffa83f9fa8 items=0 ppid=4310 pid=4559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:35.021000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Aug 13 00:03:35.138997 env[1913]: 2025-08-13 00:03:34.871 [INFO][4498] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Aug 13 00:03:35.138997 env[1913]: 2025-08-13 00:03:34.871 [INFO][4498] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" iface="eth0" netns="/var/run/netns/cni-cc3b90e4-bb2f-d749-716b-eeeee04b0b44" Aug 13 00:03:35.138997 env[1913]: 2025-08-13 00:03:34.871 [INFO][4498] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" iface="eth0" netns="/var/run/netns/cni-cc3b90e4-bb2f-d749-716b-eeeee04b0b44" Aug 13 00:03:35.138997 env[1913]: 2025-08-13 00:03:34.872 [INFO][4498] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" iface="eth0" netns="/var/run/netns/cni-cc3b90e4-bb2f-d749-716b-eeeee04b0b44" Aug 13 00:03:35.138997 env[1913]: 2025-08-13 00:03:34.872 [INFO][4498] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Aug 13 00:03:35.138997 env[1913]: 2025-08-13 00:03:34.872 [INFO][4498] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Aug 13 00:03:35.138997 env[1913]: 2025-08-13 00:03:35.102 [INFO][4544] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" HandleID="k8s-pod-network.3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Workload="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:03:35.138997 env[1913]: 2025-08-13 00:03:35.111 [INFO][4544] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:35.138997 env[1913]: 2025-08-13 00:03:35.111 [INFO][4544] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:35.138997 env[1913]: 2025-08-13 00:03:35.130 [WARNING][4544] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" HandleID="k8s-pod-network.3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Workload="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:03:35.138997 env[1913]: 2025-08-13 00:03:35.130 [INFO][4544] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" HandleID="k8s-pod-network.3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Workload="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:03:35.138997 env[1913]: 2025-08-13 00:03:35.132 [INFO][4544] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:35.138997 env[1913]: 2025-08-13 00:03:35.135 [INFO][4498] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Aug 13 00:03:35.140206 env[1913]: time="2025-08-13T00:03:35.139291381Z" level=info msg="TearDown network for sandbox \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\" successfully" Aug 13 00:03:35.140206 env[1913]: time="2025-08-13T00:03:35.139389841Z" level=info msg="StopPodSandbox for \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\" returns successfully" Aug 13 00:03:35.140742 env[1913]: time="2025-08-13T00:03:35.140680642Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-d75f5fd6d-27n9j,Uid:ce695f2d-ac34-4846-848c-9adafe443946,Namespace:calico-system,Attempt:1,}" Aug 13 00:03:35.366292 systemd[1]: run-netns-cni\x2dcc3b90e4\x2dbb2f\x2dd749\x2d716b\x2deeeee04b0b44.mount: Deactivated successfully. Aug 13 00:03:35.377500 (udev-worker)[4448]: Network interface NamePolicy= disabled on kernel command line. Aug 13 00:03:35.391868 systemd-networkd[1587]: cali4658d01368a: Link UP Aug 13 00:03:35.407270 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Aug 13 00:03:35.407428 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cali4658d01368a: link becomes ready Aug 13 00:03:35.400581 systemd-networkd[1587]: cali4658d01368a: Gained carrier Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.198 [INFO][4564] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0 goldmane-58fd7646b9- calico-system cf609856-35b0-4517-9fba-45b28ef2be55 980 0 2025-08-13 00:03:09 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:58fd7646b9 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ip-172-31-27-151 goldmane-58fd7646b9-6tzfz eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali4658d01368a [] [] }} ContainerID="bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" Namespace="calico-system" Pod="goldmane-58fd7646b9-6tzfz" WorkloadEndpoint="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-" Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.198 [INFO][4564] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" Namespace="calico-system" Pod="goldmane-58fd7646b9-6tzfz" WorkloadEndpoint="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.273 [INFO][4590] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" HandleID="k8s-pod-network.bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" Workload="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.273 [INFO][4590] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" HandleID="k8s-pod-network.bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" Workload="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002c3710), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-27-151", "pod":"goldmane-58fd7646b9-6tzfz", "timestamp":"2025-08-13 00:03:35.273226426 +0000 UTC"}, Hostname:"ip-172-31-27-151", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.273 [INFO][4590] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.273 [INFO][4590] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.273 [INFO][4590] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-27-151' Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.296 [INFO][4590] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" host="ip-172-31-27-151" Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.305 [INFO][4590] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-27-151" Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.314 [INFO][4590] ipam/ipam.go 511: Trying affinity for 192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.318 [INFO][4590] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.322 [INFO][4590] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.322 [INFO][4590] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.64/26 handle="k8s-pod-network.bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" host="ip-172-31-27-151" Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.326 [INFO][4590] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.346 [INFO][4590] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.64/26 handle="k8s-pod-network.bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" host="ip-172-31-27-151" Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.360 [INFO][4590] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.66/26] block=192.168.1.64/26 handle="k8s-pod-network.bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" host="ip-172-31-27-151" Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.360 [INFO][4590] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.66/26] handle="k8s-pod-network.bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" host="ip-172-31-27-151" Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.360 [INFO][4590] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:35.480964 env[1913]: 2025-08-13 00:03:35.360 [INFO][4590] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.66/26] IPv6=[] ContainerID="bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" HandleID="k8s-pod-network.bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" Workload="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:03:35.482720 env[1913]: 2025-08-13 00:03:35.371 [INFO][4564] cni-plugin/k8s.go 418: Populated endpoint ContainerID="bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" Namespace="calico-system" Pod="goldmane-58fd7646b9-6tzfz" WorkloadEndpoint="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"cf609856-35b0-4517-9fba-45b28ef2be55", ResourceVersion:"980", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"", Pod:"goldmane-58fd7646b9-6tzfz", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.1.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali4658d01368a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:35.482720 env[1913]: 2025-08-13 00:03:35.372 [INFO][4564] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.66/32] ContainerID="bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" Namespace="calico-system" Pod="goldmane-58fd7646b9-6tzfz" WorkloadEndpoint="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:03:35.482720 env[1913]: 2025-08-13 00:03:35.372 [INFO][4564] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4658d01368a ContainerID="bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" Namespace="calico-system" Pod="goldmane-58fd7646b9-6tzfz" WorkloadEndpoint="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:03:35.482720 env[1913]: 2025-08-13 00:03:35.421 [INFO][4564] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" Namespace="calico-system" Pod="goldmane-58fd7646b9-6tzfz" WorkloadEndpoint="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:03:35.482720 env[1913]: 2025-08-13 00:03:35.423 [INFO][4564] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" Namespace="calico-system" Pod="goldmane-58fd7646b9-6tzfz" WorkloadEndpoint="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"cf609856-35b0-4517-9fba-45b28ef2be55", ResourceVersion:"980", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb", Pod:"goldmane-58fd7646b9-6tzfz", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.1.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali4658d01368a", MAC:"b6:88:26:d8:fa:df", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:35.482720 env[1913]: 2025-08-13 00:03:35.450 [INFO][4564] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb" Namespace="calico-system" Pod="goldmane-58fd7646b9-6tzfz" WorkloadEndpoint="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:03:35.494502 env[1913]: time="2025-08-13T00:03:35.493680846Z" level=info msg="StopPodSandbox for \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\"" Aug 13 00:03:35.570311 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cali03a672766d6: link becomes ready Aug 13 00:03:35.575957 systemd-networkd[1587]: cali03a672766d6: Link UP Aug 13 00:03:35.576571 systemd-networkd[1587]: cali03a672766d6: Gained carrier Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.256 [INFO][4579] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0 calico-kube-controllers-d75f5fd6d- calico-system ce695f2d-ac34-4846-848c-9adafe443946 981 0 2025-08-13 00:03:10 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:d75f5fd6d projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ip-172-31-27-151 calico-kube-controllers-d75f5fd6d-27n9j eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali03a672766d6 [] [] }} ContainerID="00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" Namespace="calico-system" Pod="calico-kube-controllers-d75f5fd6d-27n9j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-" Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.256 [INFO][4579] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" Namespace="calico-system" Pod="calico-kube-controllers-d75f5fd6d-27n9j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.395 [INFO][4599] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" HandleID="k8s-pod-network.00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" Workload="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.396 [INFO][4599] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" HandleID="k8s-pod-network.00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" Workload="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002d3820), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-27-151", "pod":"calico-kube-controllers-d75f5fd6d-27n9j", "timestamp":"2025-08-13 00:03:35.395938236 +0000 UTC"}, Hostname:"ip-172-31-27-151", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.396 [INFO][4599] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.396 [INFO][4599] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.397 [INFO][4599] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-27-151' Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.423 [INFO][4599] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" host="ip-172-31-27-151" Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.438 [INFO][4599] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-27-151" Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.454 [INFO][4599] ipam/ipam.go 511: Trying affinity for 192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.459 [INFO][4599] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.464 [INFO][4599] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.465 [INFO][4599] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.64/26 handle="k8s-pod-network.00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" host="ip-172-31-27-151" Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.468 [INFO][4599] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0 Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.497 [INFO][4599] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.64/26 handle="k8s-pod-network.00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" host="ip-172-31-27-151" Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.527 [INFO][4599] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.67/26] block=192.168.1.64/26 handle="k8s-pod-network.00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" host="ip-172-31-27-151" Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.527 [INFO][4599] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.67/26] handle="k8s-pod-network.00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" host="ip-172-31-27-151" Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.527 [INFO][4599] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:35.607759 env[1913]: 2025-08-13 00:03:35.527 [INFO][4599] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.67/26] IPv6=[] ContainerID="00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" HandleID="k8s-pod-network.00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" Workload="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:03:35.609005 env[1913]: 2025-08-13 00:03:35.536 [INFO][4579] cni-plugin/k8s.go 418: Populated endpoint ContainerID="00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" Namespace="calico-system" Pod="calico-kube-controllers-d75f5fd6d-27n9j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0", GenerateName:"calico-kube-controllers-d75f5fd6d-", Namespace:"calico-system", SelfLink:"", UID:"ce695f2d-ac34-4846-848c-9adafe443946", ResourceVersion:"981", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"d75f5fd6d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"", Pod:"calico-kube-controllers-d75f5fd6d-27n9j", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.1.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali03a672766d6", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:35.609005 env[1913]: 2025-08-13 00:03:35.536 [INFO][4579] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.67/32] ContainerID="00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" Namespace="calico-system" Pod="calico-kube-controllers-d75f5fd6d-27n9j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:03:35.609005 env[1913]: 2025-08-13 00:03:35.536 [INFO][4579] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali03a672766d6 ContainerID="00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" Namespace="calico-system" Pod="calico-kube-controllers-d75f5fd6d-27n9j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:03:35.609005 env[1913]: 2025-08-13 00:03:35.574 [INFO][4579] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" Namespace="calico-system" Pod="calico-kube-controllers-d75f5fd6d-27n9j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:03:35.609005 env[1913]: 2025-08-13 00:03:35.583 [INFO][4579] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" Namespace="calico-system" Pod="calico-kube-controllers-d75f5fd6d-27n9j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0", GenerateName:"calico-kube-controllers-d75f5fd6d-", Namespace:"calico-system", SelfLink:"", UID:"ce695f2d-ac34-4846-848c-9adafe443946", ResourceVersion:"981", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"d75f5fd6d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0", Pod:"calico-kube-controllers-d75f5fd6d-27n9j", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.1.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali03a672766d6", MAC:"46:c6:7f:d9:65:64", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:35.609005 env[1913]: 2025-08-13 00:03:35.604 [INFO][4579] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0" Namespace="calico-system" Pod="calico-kube-controllers-d75f5fd6d-27n9j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:03:35.615000 audit[4632]: NETFILTER_CFG table=filter:105 family=2 entries=44 op=nft_register_chain pid=4632 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Aug 13 00:03:35.615000 audit[4632]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=25180 a0=3 a1=fffff5164630 a2=0 a3=ffff89dd2fa8 items=0 ppid=4310 pid=4632 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:35.615000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Aug 13 00:03:35.639750 env[1913]: time="2025-08-13T00:03:35.639476283Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:03:35.640095 env[1913]: time="2025-08-13T00:03:35.640027231Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:03:35.640394 env[1913]: time="2025-08-13T00:03:35.640332622Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:03:35.644972 env[1913]: time="2025-08-13T00:03:35.643073177Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb pid=4639 runtime=io.containerd.runc.v2 Aug 13 00:03:35.667000 audit[4660]: NETFILTER_CFG table=filter:106 family=2 entries=40 op=nft_register_chain pid=4660 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Aug 13 00:03:35.667000 audit[4660]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=20764 a0=3 a1=fffffb3b3cd0 a2=0 a3=ffffab777fa8 items=0 ppid=4310 pid=4660 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:35.667000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Aug 13 00:03:35.681450 env[1913]: time="2025-08-13T00:03:35.681053853Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:03:35.681763 env[1913]: time="2025-08-13T00:03:35.681695402Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:03:35.682005 env[1913]: time="2025-08-13T00:03:35.681892059Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:03:35.683713 env[1913]: time="2025-08-13T00:03:35.683591583Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0 pid=4663 runtime=io.containerd.runc.v2 Aug 13 00:03:35.777913 systemd[1]: run-containerd-runc-k8s.io-00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0-runc.Pl8ptY.mount: Deactivated successfully. Aug 13 00:03:35.779432 systemd-networkd[1587]: vxlan.calico: Gained IPv6LL Aug 13 00:03:36.058830 env[1913]: 2025-08-13 00:03:35.847 [INFO][4626] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Aug 13 00:03:36.058830 env[1913]: 2025-08-13 00:03:35.849 [INFO][4626] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" iface="eth0" netns="/var/run/netns/cni-12978413-259f-db37-aa3f-9952988b36dc" Aug 13 00:03:36.058830 env[1913]: 2025-08-13 00:03:35.851 [INFO][4626] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" iface="eth0" netns="/var/run/netns/cni-12978413-259f-db37-aa3f-9952988b36dc" Aug 13 00:03:36.058830 env[1913]: 2025-08-13 00:03:35.851 [INFO][4626] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" iface="eth0" netns="/var/run/netns/cni-12978413-259f-db37-aa3f-9952988b36dc" Aug 13 00:03:36.058830 env[1913]: 2025-08-13 00:03:35.851 [INFO][4626] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Aug 13 00:03:36.058830 env[1913]: 2025-08-13 00:03:35.851 [INFO][4626] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Aug 13 00:03:36.058830 env[1913]: 2025-08-13 00:03:36.014 [INFO][4708] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" HandleID="k8s-pod-network.df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:03:36.058830 env[1913]: 2025-08-13 00:03:36.018 [INFO][4708] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:36.058830 env[1913]: 2025-08-13 00:03:36.018 [INFO][4708] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:36.058830 env[1913]: 2025-08-13 00:03:36.034 [WARNING][4708] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" HandleID="k8s-pod-network.df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:03:36.058830 env[1913]: 2025-08-13 00:03:36.034 [INFO][4708] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" HandleID="k8s-pod-network.df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:03:36.058830 env[1913]: 2025-08-13 00:03:36.036 [INFO][4708] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:36.058830 env[1913]: 2025-08-13 00:03:36.049 [INFO][4626] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Aug 13 00:03:36.060840 env[1913]: time="2025-08-13T00:03:36.059152515Z" level=info msg="TearDown network for sandbox \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\" successfully" Aug 13 00:03:36.060840 env[1913]: time="2025-08-13T00:03:36.059210751Z" level=info msg="StopPodSandbox for \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\" returns successfully" Aug 13 00:03:36.062559 env[1913]: time="2025-08-13T00:03:36.062487530Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6ff96b959-s8g8j,Uid:71cf2b5c-a9b2-457e-a496-c93a3a2d8126,Namespace:calico-apiserver,Attempt:1,}" Aug 13 00:03:36.074554 env[1913]: time="2025-08-13T00:03:36.074490781Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-d75f5fd6d-27n9j,Uid:ce695f2d-ac34-4846-848c-9adafe443946,Namespace:calico-system,Attempt:1,} returns sandbox id \"00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0\"" Aug 13 00:03:36.095378 env[1913]: time="2025-08-13T00:03:36.094473799Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-6tzfz,Uid:cf609856-35b0-4517-9fba-45b28ef2be55,Namespace:calico-system,Attempt:1,} returns sandbox id \"bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb\"" Aug 13 00:03:36.330201 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cali9af93f385cf: link becomes ready Aug 13 00:03:36.327249 systemd-networkd[1587]: cali9af93f385cf: Link UP Aug 13 00:03:36.332254 systemd-networkd[1587]: cali9af93f385cf: Gained carrier Aug 13 00:03:36.350134 systemd[1]: run-netns-cni\x2d12978413\x2d259f\x2ddb37\x2daa3f\x2d9952988b36dc.mount: Deactivated successfully. Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.183 [INFO][4732] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0 calico-apiserver-6ff96b959- calico-apiserver 71cf2b5c-a9b2-457e-a496-c93a3a2d8126 991 0 2025-08-13 00:03:02 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6ff96b959 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ip-172-31-27-151 calico-apiserver-6ff96b959-s8g8j eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali9af93f385cf [] [] }} ContainerID="df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-s8g8j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-" Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.183 [INFO][4732] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-s8g8j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.241 [INFO][4740] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" HandleID="k8s-pod-network.df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.241 [INFO][4740] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" HandleID="k8s-pod-network.df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002c3010), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ip-172-31-27-151", "pod":"calico-apiserver-6ff96b959-s8g8j", "timestamp":"2025-08-13 00:03:36.241372134 +0000 UTC"}, Hostname:"ip-172-31-27-151", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.241 [INFO][4740] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.241 [INFO][4740] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.242 [INFO][4740] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-27-151' Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.255 [INFO][4740] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" host="ip-172-31-27-151" Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.262 [INFO][4740] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-27-151" Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.269 [INFO][4740] ipam/ipam.go 511: Trying affinity for 192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.272 [INFO][4740] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.284 [INFO][4740] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.284 [INFO][4740] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.64/26 handle="k8s-pod-network.df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" host="ip-172-31-27-151" Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.287 [INFO][4740] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19 Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.294 [INFO][4740] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.64/26 handle="k8s-pod-network.df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" host="ip-172-31-27-151" Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.309 [INFO][4740] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.68/26] block=192.168.1.64/26 handle="k8s-pod-network.df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" host="ip-172-31-27-151" Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.309 [INFO][4740] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.68/26] handle="k8s-pod-network.df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" host="ip-172-31-27-151" Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.309 [INFO][4740] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:36.383158 env[1913]: 2025-08-13 00:03:36.309 [INFO][4740] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.68/26] IPv6=[] ContainerID="df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" HandleID="k8s-pod-network.df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:03:36.384591 env[1913]: 2025-08-13 00:03:36.313 [INFO][4732] cni-plugin/k8s.go 418: Populated endpoint ContainerID="df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-s8g8j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0", GenerateName:"calico-apiserver-6ff96b959-", Namespace:"calico-apiserver", SelfLink:"", UID:"71cf2b5c-a9b2-457e-a496-c93a3a2d8126", ResourceVersion:"991", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6ff96b959", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"", Pod:"calico-apiserver-6ff96b959-s8g8j", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.1.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali9af93f385cf", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:36.384591 env[1913]: 2025-08-13 00:03:36.313 [INFO][4732] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.68/32] ContainerID="df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-s8g8j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:03:36.384591 env[1913]: 2025-08-13 00:03:36.313 [INFO][4732] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9af93f385cf ContainerID="df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-s8g8j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:03:36.384591 env[1913]: 2025-08-13 00:03:36.333 [INFO][4732] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-s8g8j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:03:36.384591 env[1913]: 2025-08-13 00:03:36.334 [INFO][4732] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-s8g8j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0", GenerateName:"calico-apiserver-6ff96b959-", Namespace:"calico-apiserver", SelfLink:"", UID:"71cf2b5c-a9b2-457e-a496-c93a3a2d8126", ResourceVersion:"991", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6ff96b959", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19", Pod:"calico-apiserver-6ff96b959-s8g8j", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.1.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali9af93f385cf", MAC:"c2:7e:a4:6b:63:8b", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:36.384591 env[1913]: 2025-08-13 00:03:36.375 [INFO][4732] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-s8g8j" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:03:36.401000 audit[4755]: NETFILTER_CFG table=filter:107 family=2 entries=58 op=nft_register_chain pid=4755 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Aug 13 00:03:36.401000 audit[4755]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=30584 a0=3 a1=ffffe7651c40 a2=0 a3=ffffa1a68fa8 items=0 ppid=4310 pid=4755 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:36.401000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Aug 13 00:03:36.432031 env[1913]: time="2025-08-13T00:03:36.431886738Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:03:36.432031 env[1913]: time="2025-08-13T00:03:36.431972215Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:03:36.432433 env[1913]: time="2025-08-13T00:03:36.431999395Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:03:36.433558 env[1913]: time="2025-08-13T00:03:36.433314688Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19 pid=4763 runtime=io.containerd.runc.v2 Aug 13 00:03:36.482141 env[1913]: time="2025-08-13T00:03:36.477352053Z" level=info msg="StopPodSandbox for \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\"" Aug 13 00:03:36.503401 env[1913]: time="2025-08-13T00:03:36.503345913Z" level=info msg="StopPodSandbox for \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\"" Aug 13 00:03:36.870729 env[1913]: time="2025-08-13T00:03:36.870654095Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6ff96b959-s8g8j,Uid:71cf2b5c-a9b2-457e-a496-c93a3a2d8126,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19\"" Aug 13 00:03:36.926350 env[1913]: 2025-08-13 00:03:36.781 [INFO][4806] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Aug 13 00:03:36.926350 env[1913]: 2025-08-13 00:03:36.782 [INFO][4806] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" iface="eth0" netns="/var/run/netns/cni-fe6028cf-3a84-36a2-8a97-29b061128be4" Aug 13 00:03:36.926350 env[1913]: 2025-08-13 00:03:36.782 [INFO][4806] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" iface="eth0" netns="/var/run/netns/cni-fe6028cf-3a84-36a2-8a97-29b061128be4" Aug 13 00:03:36.926350 env[1913]: 2025-08-13 00:03:36.782 [INFO][4806] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" iface="eth0" netns="/var/run/netns/cni-fe6028cf-3a84-36a2-8a97-29b061128be4" Aug 13 00:03:36.926350 env[1913]: 2025-08-13 00:03:36.782 [INFO][4806] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Aug 13 00:03:36.926350 env[1913]: 2025-08-13 00:03:36.782 [INFO][4806] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Aug 13 00:03:36.926350 env[1913]: 2025-08-13 00:03:36.892 [INFO][4829] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" HandleID="k8s-pod-network.a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:03:36.926350 env[1913]: 2025-08-13 00:03:36.893 [INFO][4829] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:36.926350 env[1913]: 2025-08-13 00:03:36.893 [INFO][4829] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:36.926350 env[1913]: 2025-08-13 00:03:36.908 [WARNING][4829] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" HandleID="k8s-pod-network.a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:03:36.926350 env[1913]: 2025-08-13 00:03:36.908 [INFO][4829] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" HandleID="k8s-pod-network.a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:03:36.926350 env[1913]: 2025-08-13 00:03:36.911 [INFO][4829] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:36.926350 env[1913]: 2025-08-13 00:03:36.923 [INFO][4806] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Aug 13 00:03:36.947912 systemd[1]: run-netns-cni\x2dfe6028cf\x2d3a84\x2d36a2\x2d8a97\x2d29b061128be4.mount: Deactivated successfully. Aug 13 00:03:36.953132 env[1913]: time="2025-08-13T00:03:36.949462421Z" level=info msg="TearDown network for sandbox \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\" successfully" Aug 13 00:03:36.953132 env[1913]: time="2025-08-13T00:03:36.949532418Z" level=info msg="StopPodSandbox for \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\" returns successfully" Aug 13 00:03:36.955972 env[1913]: time="2025-08-13T00:03:36.955911686Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-kqx4k,Uid:48c5f0fe-d3a5-4b24-a4f6-448275db6b4d,Namespace:kube-system,Attempt:1,}" Aug 13 00:03:36.965336 systemd-networkd[1587]: cali4658d01368a: Gained IPv6LL Aug 13 00:03:37.009960 env[1913]: 2025-08-13 00:03:36.729 [INFO][4810] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Aug 13 00:03:37.009960 env[1913]: 2025-08-13 00:03:36.730 [INFO][4810] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" iface="eth0" netns="/var/run/netns/cni-2a899caa-86b3-7d76-315c-2659dfec7958" Aug 13 00:03:37.009960 env[1913]: 2025-08-13 00:03:36.730 [INFO][4810] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" iface="eth0" netns="/var/run/netns/cni-2a899caa-86b3-7d76-315c-2659dfec7958" Aug 13 00:03:37.009960 env[1913]: 2025-08-13 00:03:36.731 [INFO][4810] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" iface="eth0" netns="/var/run/netns/cni-2a899caa-86b3-7d76-315c-2659dfec7958" Aug 13 00:03:37.009960 env[1913]: 2025-08-13 00:03:36.731 [INFO][4810] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Aug 13 00:03:37.009960 env[1913]: 2025-08-13 00:03:36.731 [INFO][4810] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Aug 13 00:03:37.009960 env[1913]: 2025-08-13 00:03:36.905 [INFO][4823] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" HandleID="k8s-pod-network.359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:03:37.009960 env[1913]: 2025-08-13 00:03:36.908 [INFO][4823] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:37.009960 env[1913]: 2025-08-13 00:03:36.911 [INFO][4823] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:37.009960 env[1913]: 2025-08-13 00:03:36.956 [WARNING][4823] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" HandleID="k8s-pod-network.359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:03:37.009960 env[1913]: 2025-08-13 00:03:36.956 [INFO][4823] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" HandleID="k8s-pod-network.359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:03:37.009960 env[1913]: 2025-08-13 00:03:36.962 [INFO][4823] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:37.009960 env[1913]: 2025-08-13 00:03:36.978 [INFO][4810] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Aug 13 00:03:37.027087 env[1913]: time="2025-08-13T00:03:37.027010064Z" level=info msg="TearDown network for sandbox \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\" successfully" Aug 13 00:03:37.027309 env[1913]: time="2025-08-13T00:03:37.027074840Z" level=info msg="StopPodSandbox for \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\" returns successfully" Aug 13 00:03:37.028618 env[1913]: time="2025-08-13T00:03:37.028552626Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6ff96b959-jggcm,Uid:6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e,Namespace:calico-apiserver,Attempt:1,}" Aug 13 00:03:37.219443 systemd-networkd[1587]: cali03a672766d6: Gained IPv6LL Aug 13 00:03:37.337008 systemd-networkd[1587]: cali28fc58a798b: Link UP Aug 13 00:03:37.348404 systemd[1]: run-netns-cni\x2d2a899caa\x2d86b3\x2d7d76\x2d315c\x2d2659dfec7958.mount: Deactivated successfully. Aug 13 00:03:37.359688 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Aug 13 00:03:37.361679 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cali28fc58a798b: link becomes ready Aug 13 00:03:37.360467 systemd-networkd[1587]: cali28fc58a798b: Gained carrier Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.161 [INFO][4843] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0 coredns-7c65d6cfc9- kube-system 48c5f0fe-d3a5-4b24-a4f6-448275db6b4d 1003 0 2025-08-13 00:02:33 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ip-172-31-27-151 coredns-7c65d6cfc9-kqx4k eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali28fc58a798b [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-kqx4k" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-" Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.161 [INFO][4843] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-kqx4k" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.246 [INFO][4867] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" HandleID="k8s-pod-network.dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.247 [INFO][4867] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" HandleID="k8s-pod-network.dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40003236e0), Attrs:map[string]string{"namespace":"kube-system", "node":"ip-172-31-27-151", "pod":"coredns-7c65d6cfc9-kqx4k", "timestamp":"2025-08-13 00:03:37.246088151 +0000 UTC"}, Hostname:"ip-172-31-27-151", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.247 [INFO][4867] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.248 [INFO][4867] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.248 [INFO][4867] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-27-151' Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.270 [INFO][4867] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" host="ip-172-31-27-151" Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.277 [INFO][4867] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-27-151" Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.285 [INFO][4867] ipam/ipam.go 511: Trying affinity for 192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.288 [INFO][4867] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.296 [INFO][4867] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.297 [INFO][4867] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.64/26 handle="k8s-pod-network.dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" host="ip-172-31-27-151" Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.299 [INFO][4867] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.306 [INFO][4867] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.64/26 handle="k8s-pod-network.dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" host="ip-172-31-27-151" Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.322 [INFO][4867] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.69/26] block=192.168.1.64/26 handle="k8s-pod-network.dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" host="ip-172-31-27-151" Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.322 [INFO][4867] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.69/26] handle="k8s-pod-network.dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" host="ip-172-31-27-151" Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.322 [INFO][4867] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:37.397949 env[1913]: 2025-08-13 00:03:37.322 [INFO][4867] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.69/26] IPv6=[] ContainerID="dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" HandleID="k8s-pod-network.dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:03:37.399644 env[1913]: 2025-08-13 00:03:37.326 [INFO][4843] cni-plugin/k8s.go 418: Populated endpoint ContainerID="dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-kqx4k" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"48c5f0fe-d3a5-4b24-a4f6-448275db6b4d", ResourceVersion:"1003", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 2, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"", Pod:"coredns-7c65d6cfc9-kqx4k", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.1.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali28fc58a798b", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:37.399644 env[1913]: 2025-08-13 00:03:37.326 [INFO][4843] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.69/32] ContainerID="dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-kqx4k" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:03:37.399644 env[1913]: 2025-08-13 00:03:37.327 [INFO][4843] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali28fc58a798b ContainerID="dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-kqx4k" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:03:37.399644 env[1913]: 2025-08-13 00:03:37.362 [INFO][4843] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-kqx4k" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:03:37.399644 env[1913]: 2025-08-13 00:03:37.363 [INFO][4843] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-kqx4k" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"48c5f0fe-d3a5-4b24-a4f6-448275db6b4d", ResourceVersion:"1003", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 2, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b", Pod:"coredns-7c65d6cfc9-kqx4k", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.1.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali28fc58a798b", MAC:"fa:6d:34:c1:3b:03", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:37.399644 env[1913]: 2025-08-13 00:03:37.390 [INFO][4843] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b" Namespace="kube-system" Pod="coredns-7c65d6cfc9-kqx4k" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:03:37.424000 audit[4891]: NETFILTER_CFG table=filter:108 family=2 entries=54 op=nft_register_chain pid=4891 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Aug 13 00:03:37.424000 audit[4891]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=26116 a0=3 a1=ffffc5a07710 a2=0 a3=ffff8378afa8 items=0 ppid=4310 pid=4891 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:37.424000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Aug 13 00:03:37.483453 env[1913]: time="2025-08-13T00:03:37.480457402Z" level=info msg="StopPodSandbox for \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\"" Aug 13 00:03:37.487072 env[1913]: time="2025-08-13T00:03:37.487007335Z" level=info msg="StopPodSandbox for \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\"" Aug 13 00:03:37.514942 env[1913]: time="2025-08-13T00:03:37.513616676Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:03:37.514942 env[1913]: time="2025-08-13T00:03:37.513688256Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:03:37.514942 env[1913]: time="2025-08-13T00:03:37.513713192Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:03:37.514942 env[1913]: time="2025-08-13T00:03:37.513968230Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b pid=4902 runtime=io.containerd.runc.v2 Aug 13 00:03:37.723193 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cali3abb69b4f11: link becomes ready Aug 13 00:03:37.723537 systemd-networkd[1587]: cali3abb69b4f11: Link UP Aug 13 00:03:37.723833 systemd-networkd[1587]: cali3abb69b4f11: Gained carrier Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.251 [INFO][4854] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0 calico-apiserver-6ff96b959- calico-apiserver 6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e 1002 0 2025-08-13 00:03:01 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6ff96b959 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ip-172-31-27-151 calico-apiserver-6ff96b959-jggcm eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali3abb69b4f11 [] [] }} ContainerID="c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-jggcm" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-" Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.251 [INFO][4854] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-jggcm" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.512 [INFO][4878] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" HandleID="k8s-pod-network.c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.514 [INFO][4878] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" HandleID="k8s-pod-network.c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002d3860), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ip-172-31-27-151", "pod":"calico-apiserver-6ff96b959-jggcm", "timestamp":"2025-08-13 00:03:37.512545536 +0000 UTC"}, Hostname:"ip-172-31-27-151", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.516 [INFO][4878] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.516 [INFO][4878] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.516 [INFO][4878] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-27-151' Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.556 [INFO][4878] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" host="ip-172-31-27-151" Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.570 [INFO][4878] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-27-151" Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.600 [INFO][4878] ipam/ipam.go 511: Trying affinity for 192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.625 [INFO][4878] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.632 [INFO][4878] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.636 [INFO][4878] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.64/26 handle="k8s-pod-network.c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" host="ip-172-31-27-151" Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.641 [INFO][4878] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9 Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.665 [INFO][4878] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.64/26 handle="k8s-pod-network.c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" host="ip-172-31-27-151" Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.685 [INFO][4878] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.70/26] block=192.168.1.64/26 handle="k8s-pod-network.c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" host="ip-172-31-27-151" Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.685 [INFO][4878] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.70/26] handle="k8s-pod-network.c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" host="ip-172-31-27-151" Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.686 [INFO][4878] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:37.770776 env[1913]: 2025-08-13 00:03:37.686 [INFO][4878] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.70/26] IPv6=[] ContainerID="c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" HandleID="k8s-pod-network.c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:03:37.772267 env[1913]: 2025-08-13 00:03:37.697 [INFO][4854] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-jggcm" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0", GenerateName:"calico-apiserver-6ff96b959-", Namespace:"calico-apiserver", SelfLink:"", UID:"6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e", ResourceVersion:"1002", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6ff96b959", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"", Pod:"calico-apiserver-6ff96b959-jggcm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.1.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3abb69b4f11", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:37.772267 env[1913]: 2025-08-13 00:03:37.697 [INFO][4854] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.70/32] ContainerID="c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-jggcm" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:03:37.772267 env[1913]: 2025-08-13 00:03:37.697 [INFO][4854] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3abb69b4f11 ContainerID="c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-jggcm" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:03:37.772267 env[1913]: 2025-08-13 00:03:37.726 [INFO][4854] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-jggcm" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:03:37.772267 env[1913]: 2025-08-13 00:03:37.726 [INFO][4854] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-jggcm" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0", GenerateName:"calico-apiserver-6ff96b959-", Namespace:"calico-apiserver", SelfLink:"", UID:"6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e", ResourceVersion:"1002", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6ff96b959", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9", Pod:"calico-apiserver-6ff96b959-jggcm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.1.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3abb69b4f11", MAC:"3e:d4:32:b2:57:26", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:37.772267 env[1913]: 2025-08-13 00:03:37.754 [INFO][4854] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9" Namespace="calico-apiserver" Pod="calico-apiserver-6ff96b959-jggcm" WorkloadEndpoint="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:03:37.798000 audit[4972]: NETFILTER_CFG table=filter:109 family=2 entries=53 op=nft_register_chain pid=4972 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Aug 13 00:03:37.798000 audit[4972]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=26640 a0=3 a1=ffffec03f5a0 a2=0 a3=ffff8ce9bfa8 items=0 ppid=4310 pid=4972 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:37.798000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Aug 13 00:03:37.802183 env[1913]: time="2025-08-13T00:03:37.801869450Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-kqx4k,Uid:48c5f0fe-d3a5-4b24-a4f6-448275db6b4d,Namespace:kube-system,Attempt:1,} returns sandbox id \"dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b\"" Aug 13 00:03:37.818870 env[1913]: time="2025-08-13T00:03:37.818792817Z" level=info msg="CreateContainer within sandbox \"dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 00:03:37.858424 systemd-networkd[1587]: cali9af93f385cf: Gained IPv6LL Aug 13 00:03:37.871808 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1199958288.mount: Deactivated successfully. Aug 13 00:03:37.897762 env[1913]: time="2025-08-13T00:03:37.897675534Z" level=info msg="CreateContainer within sandbox \"dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"3355537c13a080132a31b8512fd9a18637a7c1699ea305f85cc3514c988c3b1d\"" Aug 13 00:03:37.898700 env[1913]: time="2025-08-13T00:03:37.898622820Z" level=info msg="StartContainer for \"3355537c13a080132a31b8512fd9a18637a7c1699ea305f85cc3514c988c3b1d\"" Aug 13 00:03:37.908411 env[1913]: time="2025-08-13T00:03:37.907661390Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:03:37.908411 env[1913]: time="2025-08-13T00:03:37.907731962Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:03:37.908411 env[1913]: time="2025-08-13T00:03:37.907771431Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:03:37.909625 env[1913]: time="2025-08-13T00:03:37.909354181Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9 pid=4990 runtime=io.containerd.runc.v2 Aug 13 00:03:38.039442 env[1913]: 2025-08-13 00:03:37.820 [INFO][4932] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Aug 13 00:03:38.039442 env[1913]: 2025-08-13 00:03:37.820 [INFO][4932] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" iface="eth0" netns="/var/run/netns/cni-d1d5b105-b1af-29c0-5352-57c21cd80429" Aug 13 00:03:38.039442 env[1913]: 2025-08-13 00:03:37.824 [INFO][4932] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" iface="eth0" netns="/var/run/netns/cni-d1d5b105-b1af-29c0-5352-57c21cd80429" Aug 13 00:03:38.039442 env[1913]: 2025-08-13 00:03:37.825 [INFO][4932] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" iface="eth0" netns="/var/run/netns/cni-d1d5b105-b1af-29c0-5352-57c21cd80429" Aug 13 00:03:38.039442 env[1913]: 2025-08-13 00:03:37.825 [INFO][4932] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Aug 13 00:03:38.039442 env[1913]: 2025-08-13 00:03:37.825 [INFO][4932] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Aug 13 00:03:38.039442 env[1913]: 2025-08-13 00:03:37.990 [INFO][4979] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" HandleID="k8s-pod-network.6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:03:38.039442 env[1913]: 2025-08-13 00:03:37.993 [INFO][4979] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:38.039442 env[1913]: 2025-08-13 00:03:37.993 [INFO][4979] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:38.039442 env[1913]: 2025-08-13 00:03:38.017 [WARNING][4979] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" HandleID="k8s-pod-network.6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:03:38.039442 env[1913]: 2025-08-13 00:03:38.017 [INFO][4979] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" HandleID="k8s-pod-network.6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:03:38.039442 env[1913]: 2025-08-13 00:03:38.019 [INFO][4979] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:38.039442 env[1913]: 2025-08-13 00:03:38.026 [INFO][4932] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Aug 13 00:03:38.039442 env[1913]: time="2025-08-13T00:03:38.030123984Z" level=info msg="TearDown network for sandbox \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\" successfully" Aug 13 00:03:38.039442 env[1913]: time="2025-08-13T00:03:38.030171288Z" level=info msg="StopPodSandbox for \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\" returns successfully" Aug 13 00:03:38.041890 env[1913]: time="2025-08-13T00:03:38.041637881Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-v6d44,Uid:ba5590d9-af5c-484d-9d4d-2e76ac5327b7,Namespace:kube-system,Attempt:1,}" Aug 13 00:03:38.296260 env[1913]: time="2025-08-13T00:03:38.293543553Z" level=info msg="StartContainer for \"3355537c13a080132a31b8512fd9a18637a7c1699ea305f85cc3514c988c3b1d\" returns successfully" Aug 13 00:03:38.367525 env[1913]: 2025-08-13 00:03:37.974 [INFO][4941] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Aug 13 00:03:38.367525 env[1913]: 2025-08-13 00:03:37.975 [INFO][4941] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" iface="eth0" netns="/var/run/netns/cni-1eb12bdf-98d2-99eb-62e9-9ae7796f4b29" Aug 13 00:03:38.367525 env[1913]: 2025-08-13 00:03:37.975 [INFO][4941] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" iface="eth0" netns="/var/run/netns/cni-1eb12bdf-98d2-99eb-62e9-9ae7796f4b29" Aug 13 00:03:38.367525 env[1913]: 2025-08-13 00:03:37.976 [INFO][4941] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" iface="eth0" netns="/var/run/netns/cni-1eb12bdf-98d2-99eb-62e9-9ae7796f4b29" Aug 13 00:03:38.367525 env[1913]: 2025-08-13 00:03:37.976 [INFO][4941] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Aug 13 00:03:38.367525 env[1913]: 2025-08-13 00:03:37.976 [INFO][4941] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Aug 13 00:03:38.367525 env[1913]: 2025-08-13 00:03:38.280 [INFO][5023] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" HandleID="k8s-pod-network.4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Workload="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:03:38.367525 env[1913]: 2025-08-13 00:03:38.281 [INFO][5023] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:38.367525 env[1913]: 2025-08-13 00:03:38.281 [INFO][5023] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:38.367525 env[1913]: 2025-08-13 00:03:38.306 [WARNING][5023] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" HandleID="k8s-pod-network.4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Workload="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:03:38.367525 env[1913]: 2025-08-13 00:03:38.306 [INFO][5023] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" HandleID="k8s-pod-network.4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Workload="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:03:38.367525 env[1913]: 2025-08-13 00:03:38.314 [INFO][5023] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:38.367525 env[1913]: 2025-08-13 00:03:38.320 [INFO][4941] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Aug 13 00:03:38.367525 env[1913]: time="2025-08-13T00:03:38.357377116Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6ff96b959-jggcm,Uid:6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9\"" Aug 13 00:03:38.356000 systemd[1]: run-netns-cni\x2dd1d5b105\x2db1af\x2d29c0\x2d5352\x2d57c21cd80429.mount: Deactivated successfully. Aug 13 00:03:38.381650 env[1913]: time="2025-08-13T00:03:38.368805932Z" level=info msg="TearDown network for sandbox \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\" successfully" Aug 13 00:03:38.381650 env[1913]: time="2025-08-13T00:03:38.368892741Z" level=info msg="StopPodSandbox for \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\" returns successfully" Aug 13 00:03:38.381650 env[1913]: time="2025-08-13T00:03:38.370480711Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-n4g7q,Uid:21d84c26-2961-4407-a494-a46bbc34cc12,Namespace:calico-system,Attempt:1,}" Aug 13 00:03:38.366735 systemd[1]: run-netns-cni\x2d1eb12bdf\x2d98d2\x2d99eb\x2d62e9\x2d9ae7796f4b29.mount: Deactivated successfully. Aug 13 00:03:38.753346 systemd-networkd[1587]: califc96d725571: Link UP Aug 13 00:03:38.768355 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Aug 13 00:03:38.768531 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): califc96d725571: link becomes ready Aug 13 00:03:38.768765 systemd-networkd[1587]: califc96d725571: Gained carrier Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.324 [INFO][5049] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0 coredns-7c65d6cfc9- kube-system ba5590d9-af5c-484d-9d4d-2e76ac5327b7 1016 0 2025-08-13 00:02:33 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ip-172-31-27-151 coredns-7c65d6cfc9-v6d44 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] califc96d725571 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" Namespace="kube-system" Pod="coredns-7c65d6cfc9-v6d44" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-" Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.324 [INFO][5049] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" Namespace="kube-system" Pod="coredns-7c65d6cfc9-v6d44" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.593 [INFO][5080] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" HandleID="k8s-pod-network.086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.593 [INFO][5080] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" HandleID="k8s-pod-network.086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004d6f0), Attrs:map[string]string{"namespace":"kube-system", "node":"ip-172-31-27-151", "pod":"coredns-7c65d6cfc9-v6d44", "timestamp":"2025-08-13 00:03:38.592976395 +0000 UTC"}, Hostname:"ip-172-31-27-151", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.593 [INFO][5080] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.593 [INFO][5080] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.593 [INFO][5080] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-27-151' Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.618 [INFO][5080] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" host="ip-172-31-27-151" Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.650 [INFO][5080] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-27-151" Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.669 [INFO][5080] ipam/ipam.go 511: Trying affinity for 192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.678 [INFO][5080] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.689 [INFO][5080] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.689 [INFO][5080] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.64/26 handle="k8s-pod-network.086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" host="ip-172-31-27-151" Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.696 [INFO][5080] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.706 [INFO][5080] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.64/26 handle="k8s-pod-network.086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" host="ip-172-31-27-151" Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.721 [INFO][5080] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.71/26] block=192.168.1.64/26 handle="k8s-pod-network.086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" host="ip-172-31-27-151" Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.721 [INFO][5080] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.71/26] handle="k8s-pod-network.086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" host="ip-172-31-27-151" Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.721 [INFO][5080] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:38.940401 env[1913]: 2025-08-13 00:03:38.721 [INFO][5080] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.71/26] IPv6=[] ContainerID="086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" HandleID="k8s-pod-network.086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:03:38.942271 env[1913]: 2025-08-13 00:03:38.736 [INFO][5049] cni-plugin/k8s.go 418: Populated endpoint ContainerID="086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" Namespace="kube-system" Pod="coredns-7c65d6cfc9-v6d44" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"ba5590d9-af5c-484d-9d4d-2e76ac5327b7", ResourceVersion:"1016", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 2, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"", Pod:"coredns-7c65d6cfc9-v6d44", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.1.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"califc96d725571", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:38.942271 env[1913]: 2025-08-13 00:03:38.737 [INFO][5049] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.71/32] ContainerID="086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" Namespace="kube-system" Pod="coredns-7c65d6cfc9-v6d44" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:03:38.942271 env[1913]: 2025-08-13 00:03:38.737 [INFO][5049] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to califc96d725571 ContainerID="086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" Namespace="kube-system" Pod="coredns-7c65d6cfc9-v6d44" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:03:38.942271 env[1913]: 2025-08-13 00:03:38.821 [INFO][5049] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" Namespace="kube-system" Pod="coredns-7c65d6cfc9-v6d44" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:03:38.942271 env[1913]: 2025-08-13 00:03:38.822 [INFO][5049] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" Namespace="kube-system" Pod="coredns-7c65d6cfc9-v6d44" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"ba5590d9-af5c-484d-9d4d-2e76ac5327b7", ResourceVersion:"1016", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 2, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf", Pod:"coredns-7c65d6cfc9-v6d44", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.1.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"califc96d725571", MAC:"9a:08:de:80:ea:00", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:38.942271 env[1913]: 2025-08-13 00:03:38.882 [INFO][5049] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf" Namespace="kube-system" Pod="coredns-7c65d6cfc9-v6d44" WorkloadEndpoint="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:03:39.010972 systemd-networkd[1587]: cali28fc58a798b: Gained IPv6LL Aug 13 00:03:39.055629 env[1913]: time="2025-08-13T00:03:39.055504325Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:03:39.055931 env[1913]: time="2025-08-13T00:03:39.055876556Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:03:39.056154 env[1913]: time="2025-08-13T00:03:39.056066277Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:03:39.056678 env[1913]: time="2025-08-13T00:03:39.056585724Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf pid=5122 runtime=io.containerd.runc.v2 Aug 13 00:03:39.183015 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): calid0dd8b3f31c: link becomes ready Aug 13 00:03:39.184016 systemd-networkd[1587]: calid0dd8b3f31c: Link UP Aug 13 00:03:39.184479 systemd-networkd[1587]: calid0dd8b3f31c: Gained carrier Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:38.802 [INFO][5084] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0 csi-node-driver- calico-system 21d84c26-2961-4407-a494-a46bbc34cc12 1019 0 2025-08-13 00:03:10 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:57bd658777 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ip-172-31-27-151 csi-node-driver-n4g7q eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calid0dd8b3f31c [] [] }} ContainerID="e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" Namespace="calico-system" Pod="csi-node-driver-n4g7q" WorkloadEndpoint="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-" Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:38.807 [INFO][5084] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" Namespace="calico-system" Pod="csi-node-driver-n4g7q" WorkloadEndpoint="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:38.947 [INFO][5104] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" HandleID="k8s-pod-network.e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" Workload="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:38.947 [INFO][5104] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" HandleID="k8s-pod-network.e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" Workload="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002cbb30), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-27-151", "pod":"csi-node-driver-n4g7q", "timestamp":"2025-08-13 00:03:38.947066134 +0000 UTC"}, Hostname:"ip-172-31-27-151", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:38.948 [INFO][5104] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:38.949 [INFO][5104] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:38.949 [INFO][5104] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-27-151' Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:38.988 [INFO][5104] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" host="ip-172-31-27-151" Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:39.019 [INFO][5104] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-27-151" Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:39.031 [INFO][5104] ipam/ipam.go 511: Trying affinity for 192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:39.038 [INFO][5104] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:39.047 [INFO][5104] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.64/26 host="ip-172-31-27-151" Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:39.047 [INFO][5104] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.64/26 handle="k8s-pod-network.e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" host="ip-172-31-27-151" Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:39.062 [INFO][5104] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:39.089 [INFO][5104] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.64/26 handle="k8s-pod-network.e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" host="ip-172-31-27-151" Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:39.135 [INFO][5104] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.72/26] block=192.168.1.64/26 handle="k8s-pod-network.e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" host="ip-172-31-27-151" Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:39.135 [INFO][5104] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.72/26] handle="k8s-pod-network.e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" host="ip-172-31-27-151" Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:39.135 [INFO][5104] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:03:39.242943 env[1913]: 2025-08-13 00:03:39.135 [INFO][5104] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.72/26] IPv6=[] ContainerID="e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" HandleID="k8s-pod-network.e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" Workload="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:03:39.244433 env[1913]: 2025-08-13 00:03:39.147 [INFO][5084] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" Namespace="calico-system" Pod="csi-node-driver-n4g7q" WorkloadEndpoint="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"21d84c26-2961-4407-a494-a46bbc34cc12", ResourceVersion:"1019", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"", Pod:"csi-node-driver-n4g7q", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.1.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calid0dd8b3f31c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:39.244433 env[1913]: 2025-08-13 00:03:39.147 [INFO][5084] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.72/32] ContainerID="e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" Namespace="calico-system" Pod="csi-node-driver-n4g7q" WorkloadEndpoint="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:03:39.244433 env[1913]: 2025-08-13 00:03:39.147 [INFO][5084] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid0dd8b3f31c ContainerID="e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" Namespace="calico-system" Pod="csi-node-driver-n4g7q" WorkloadEndpoint="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:03:39.244433 env[1913]: 2025-08-13 00:03:39.197 [INFO][5084] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" Namespace="calico-system" Pod="csi-node-driver-n4g7q" WorkloadEndpoint="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:03:39.244433 env[1913]: 2025-08-13 00:03:39.201 [INFO][5084] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" Namespace="calico-system" Pod="csi-node-driver-n4g7q" WorkloadEndpoint="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"21d84c26-2961-4407-a494-a46bbc34cc12", ResourceVersion:"1019", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a", Pod:"csi-node-driver-n4g7q", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.1.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calid0dd8b3f31c", MAC:"76:18:8d:a5:d7:93", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:03:39.244433 env[1913]: 2025-08-13 00:03:39.234 [INFO][5084] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a" Namespace="calico-system" Pod="csi-node-driver-n4g7q" WorkloadEndpoint="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:03:39.334199 env[1913]: time="2025-08-13T00:03:39.332327040Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:03:39.334199 env[1913]: time="2025-08-13T00:03:39.332477197Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:03:39.334199 env[1913]: time="2025-08-13T00:03:39.332587141Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:03:39.334199 env[1913]: time="2025-08-13T00:03:39.332878575Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a pid=5164 runtime=io.containerd.runc.v2 Aug 13 00:03:39.403207 kernel: kauditd_printk_skb: 562 callbacks suppressed Aug 13 00:03:39.403356 kernel: audit: type=1325 audit(1755043419.391:407): table=filter:110 family=2 entries=52 op=nft_register_chain pid=5163 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Aug 13 00:03:39.391000 audit[5163]: NETFILTER_CFG table=filter:110 family=2 entries=52 op=nft_register_chain pid=5163 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Aug 13 00:03:39.432145 kernel: audit: type=1300 audit(1755043419.391:407): arch=c00000b7 syscall=211 success=yes exit=23908 a0=3 a1=fffffe416770 a2=0 a3=ffffb966efa8 items=0 ppid=4310 pid=5163 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:39.391000 audit[5163]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=23908 a0=3 a1=fffffe416770 a2=0 a3=ffffb966efa8 items=0 ppid=4310 pid=5163 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:39.447383 systemd[1]: run-containerd-runc-k8s.io-e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a-runc.HCP7dJ.mount: Deactivated successfully. Aug 13 00:03:39.461253 kernel: audit: type=1327 audit(1755043419.391:407): proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Aug 13 00:03:39.391000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Aug 13 00:03:39.495200 env[1913]: time="2025-08-13T00:03:39.495143437Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-v6d44,Uid:ba5590d9-af5c-484d-9d4d-2e76ac5327b7,Namespace:kube-system,Attempt:1,} returns sandbox id \"086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf\"" Aug 13 00:03:39.514193 env[1913]: time="2025-08-13T00:03:39.514138278Z" level=info msg="CreateContainer within sandbox \"086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 00:03:39.529000 audit[5193]: NETFILTER_CFG table=filter:111 family=2 entries=20 op=nft_register_rule pid=5193 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:39.554601 kernel: audit: type=1325 audit(1755043419.529:408): table=filter:111 family=2 entries=20 op=nft_register_rule pid=5193 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:39.554851 kernel: audit: type=1300 audit(1755043419.529:408): arch=c00000b7 syscall=211 success=yes exit=7480 a0=3 a1=fffff6cd3770 a2=0 a3=1 items=0 ppid=3113 pid=5193 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:39.529000 audit[5193]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=7480 a0=3 a1=fffff6cd3770 a2=0 a3=1 items=0 ppid=3113 pid=5193 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:39.529000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:39.563486 kernel: audit: type=1327 audit(1755043419.529:408): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:39.537000 audit[5193]: NETFILTER_CFG table=nat:112 family=2 entries=14 op=nft_register_rule pid=5193 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:39.537000 audit[5193]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=3468 a0=3 a1=fffff6cd3770 a2=0 a3=1 items=0 ppid=3113 pid=5193 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:39.617474 kernel: audit: type=1325 audit(1755043419.537:409): table=nat:112 family=2 entries=14 op=nft_register_rule pid=5193 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:39.617676 kernel: audit: type=1300 audit(1755043419.537:409): arch=c00000b7 syscall=211 success=yes exit=3468 a0=3 a1=fffff6cd3770 a2=0 a3=1 items=0 ppid=3113 pid=5193 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:39.622222 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1285429570.mount: Deactivated successfully. Aug 13 00:03:39.537000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:39.625474 kernel: audit: type=1327 audit(1755043419.537:409): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:39.590000 audit[5196]: NETFILTER_CFG table=filter:113 family=2 entries=60 op=nft_register_chain pid=5196 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Aug 13 00:03:39.641805 kernel: audit: type=1325 audit(1755043419.590:410): table=filter:113 family=2 entries=60 op=nft_register_chain pid=5196 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" Aug 13 00:03:39.647832 env[1913]: time="2025-08-13T00:03:39.647646199Z" level=info msg="CreateContainer within sandbox \"086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"bbe154d9c6a7e22fb7364cc8d53c3a1226c519d7f93f675879874cdecf627c06\"" Aug 13 00:03:39.651683 systemd-networkd[1587]: cali3abb69b4f11: Gained IPv6LL Aug 13 00:03:39.652926 env[1913]: time="2025-08-13T00:03:39.652872221Z" level=info msg="StartContainer for \"bbe154d9c6a7e22fb7364cc8d53c3a1226c519d7f93f675879874cdecf627c06\"" Aug 13 00:03:39.590000 audit[5196]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=26704 a0=3 a1=ffffe8f650b0 a2=0 a3=ffff8c5cdfa8 items=0 ppid=4310 pid=5196 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:39.590000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 Aug 13 00:03:39.821044 env[1913]: time="2025-08-13T00:03:39.820970778Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-n4g7q,Uid:21d84c26-2961-4407-a494-a46bbc34cc12,Namespace:calico-system,Attempt:1,} returns sandbox id \"e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a\"" Aug 13 00:03:39.927267 env[1913]: time="2025-08-13T00:03:39.927189907Z" level=info msg="StartContainer for \"bbe154d9c6a7e22fb7364cc8d53c3a1226c519d7f93f675879874cdecf627c06\" returns successfully" Aug 13 00:03:40.034441 systemd-networkd[1587]: califc96d725571: Gained IPv6LL Aug 13 00:03:40.242359 env[1913]: time="2025-08-13T00:03:40.242191841Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/whisker-backend:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:40.248335 env[1913]: time="2025-08-13T00:03:40.248245929Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:8763d908c0cd23d0e87bc61ce1ba8371b86449688baf955e5eeff7f7d7e101c4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:40.252090 env[1913]: time="2025-08-13T00:03:40.250648560Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/whisker-backend:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:40.254609 env[1913]: time="2025-08-13T00:03:40.254538025Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:40.256220 env[1913]: time="2025-08-13T00:03:40.256165356Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" returns image reference \"sha256:8763d908c0cd23d0e87bc61ce1ba8371b86449688baf955e5eeff7f7d7e101c4\"" Aug 13 00:03:40.263061 env[1913]: time="2025-08-13T00:03:40.259091695Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\"" Aug 13 00:03:40.266219 env[1913]: time="2025-08-13T00:03:40.266151268Z" level=info msg="CreateContainer within sandbox \"e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Aug 13 00:03:40.353813 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1080897784.mount: Deactivated successfully. Aug 13 00:03:40.392699 env[1913]: time="2025-08-13T00:03:40.392630033Z" level=info msg="CreateContainer within sandbox \"e8e050780a5348b0d0565e3ee2f03c5e48024717c780917e39aabf5c449ffd2e\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"f4a6b585fff3f2512ab585063c8216dda125965b39810c594fa4a9a5c22d19f9\"" Aug 13 00:03:40.394182 env[1913]: time="2025-08-13T00:03:40.394125866Z" level=info msg="StartContainer for \"f4a6b585fff3f2512ab585063c8216dda125965b39810c594fa4a9a5c22d19f9\"" Aug 13 00:03:40.414415 kubelet[2984]: I0813 00:03:40.412759 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-kqx4k" podStartSLOduration=67.412735203 podStartE2EDuration="1m7.412735203s" podCreationTimestamp="2025-08-13 00:02:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:03:39.288585684 +0000 UTC m=+72.035764628" watchObservedRunningTime="2025-08-13 00:03:40.412735203 +0000 UTC m=+73.159914123" Aug 13 00:03:40.414415 kubelet[2984]: I0813 00:03:40.413464 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-v6d44" podStartSLOduration=67.413448343 podStartE2EDuration="1m7.413448343s" podCreationTimestamp="2025-08-13 00:02:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:03:40.407863207 +0000 UTC m=+73.155042127" watchObservedRunningTime="2025-08-13 00:03:40.413448343 +0000 UTC m=+73.160627251" Aug 13 00:03:40.549825 systemd[1]: run-containerd-runc-k8s.io-f4a6b585fff3f2512ab585063c8216dda125965b39810c594fa4a9a5c22d19f9-runc.phfXyV.mount: Deactivated successfully. Aug 13 00:03:40.574000 audit[5268]: NETFILTER_CFG table=filter:114 family=2 entries=20 op=nft_register_rule pid=5268 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:40.574000 audit[5268]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=7480 a0=3 a1=ffffc49a6fc0 a2=0 a3=1 items=0 ppid=3113 pid=5268 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:40.574000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:40.582000 audit[5268]: NETFILTER_CFG table=nat:115 family=2 entries=14 op=nft_register_rule pid=5268 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:40.582000 audit[5268]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=3468 a0=3 a1=ffffc49a6fc0 a2=0 a3=1 items=0 ppid=3113 pid=5268 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:40.582000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:40.608000 audit[5272]: NETFILTER_CFG table=filter:116 family=2 entries=17 op=nft_register_rule pid=5272 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:40.608000 audit[5272]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=5248 a0=3 a1=ffffdc13dc30 a2=0 a3=1 items=0 ppid=3113 pid=5272 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:40.608000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:40.647000 audit[5272]: NETFILTER_CFG table=nat:117 family=2 entries=47 op=nft_register_chain pid=5272 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:40.647000 audit[5272]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=19860 a0=3 a1=ffffdc13dc30 a2=0 a3=1 items=0 ppid=3113 pid=5272 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:40.647000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:40.729173 env[1913]: time="2025-08-13T00:03:40.729071417Z" level=info msg="StartContainer for \"f4a6b585fff3f2512ab585063c8216dda125965b39810c594fa4a9a5c22d19f9\" returns successfully" Aug 13 00:03:40.778975 systemd[1]: Started sshd@7-172.31.27.151:22-139.178.89.65:58348.service. Aug 13 00:03:40.778000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-172.31.27.151:22-139.178.89.65:58348 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:03:40.866444 systemd-networkd[1587]: calid0dd8b3f31c: Gained IPv6LL Aug 13 00:03:41.021000 audit[5299]: USER_ACCT pid=5299 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:41.022985 sshd[5299]: Accepted publickey for core from 139.178.89.65 port 58348 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:03:41.023000 audit[5299]: CRED_ACQ pid=5299 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:41.024000 audit[5299]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffffa50efe0 a2=3 a3=1 items=0 ppid=1 pid=5299 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=8 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:41.024000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:03:41.026830 sshd[5299]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:03:41.045488 systemd[1]: Started session-8.scope. Aug 13 00:03:41.051867 systemd-logind[1901]: New session 8 of user core. Aug 13 00:03:41.120000 audit[5299]: USER_START pid=5299 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:41.123000 audit[5318]: CRED_ACQ pid=5318 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:41.348131 systemd[1]: run-containerd-runc-k8s.io-be0ed1e29b588d67e52fae942c6b3420ba4d8c77c6e0d8347717d2a5797a2a37-runc.x30txf.mount: Deactivated successfully. Aug 13 00:03:41.490490 sshd[5299]: pam_unix(sshd:session): session closed for user core Aug 13 00:03:41.493000 audit[5299]: USER_END pid=5299 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:41.494000 audit[5299]: CRED_DISP pid=5299 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:41.497000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-172.31.27.151:22-139.178.89.65:58348 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:03:41.498201 systemd[1]: sshd@7-172.31.27.151:22-139.178.89.65:58348.service: Deactivated successfully. Aug 13 00:03:41.501697 systemd[1]: session-8.scope: Deactivated successfully. Aug 13 00:03:41.508723 systemd-logind[1901]: Session 8 logged out. Waiting for processes to exit. Aug 13 00:03:41.512466 systemd-logind[1901]: Removed session 8. Aug 13 00:03:41.697000 audit[5330]: NETFILTER_CFG table=filter:118 family=2 entries=13 op=nft_register_rule pid=5330 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:41.697000 audit[5330]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=4504 a0=3 a1=fffffd5639e0 a2=0 a3=1 items=0 ppid=3113 pid=5330 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:41.697000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:41.705000 audit[5330]: NETFILTER_CFG table=nat:119 family=2 entries=27 op=nft_register_chain pid=5330 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:41.705000 audit[5330]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=9348 a0=3 a1=fffffd5639e0 a2=0 a3=1 items=0 ppid=3113 pid=5330 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:41.705000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:43.417085 env[1913]: time="2025-08-13T00:03:43.417026045Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/kube-controllers:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:43.421255 env[1913]: time="2025-08-13T00:03:43.421202742Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ba9e7793995ca67a9b78aa06adda4e89cbd435b1e88ab1032ca665140517fa7a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:43.425819 env[1913]: time="2025-08-13T00:03:43.425765266Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/kube-controllers:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:43.430984 env[1913]: time="2025-08-13T00:03:43.430905762Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:43.431532 env[1913]: time="2025-08-13T00:03:43.431467737Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" returns image reference \"sha256:ba9e7793995ca67a9b78aa06adda4e89cbd435b1e88ab1032ca665140517fa7a\"" Aug 13 00:03:43.440415 env[1913]: time="2025-08-13T00:03:43.440331964Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\"" Aug 13 00:03:43.479275 env[1913]: time="2025-08-13T00:03:43.472215152Z" level=info msg="CreateContainer within sandbox \"00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Aug 13 00:03:43.544317 env[1913]: time="2025-08-13T00:03:43.544255386Z" level=info msg="CreateContainer within sandbox \"00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"7ee0d4f88f3c0fde73248650354bfcaccd45af41c5d859d00eea0aeb5c899949\"" Aug 13 00:03:43.548909 env[1913]: time="2025-08-13T00:03:43.548843722Z" level=info msg="StartContainer for \"7ee0d4f88f3c0fde73248650354bfcaccd45af41c5d859d00eea0aeb5c899949\"" Aug 13 00:03:43.699238 env[1913]: time="2025-08-13T00:03:43.695083591Z" level=info msg="StartContainer for \"7ee0d4f88f3c0fde73248650354bfcaccd45af41c5d859d00eea0aeb5c899949\" returns successfully" Aug 13 00:03:44.425968 kubelet[2984]: I0813 00:03:44.425786 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-66f7b8977f-mqv9d" podStartSLOduration=5.040149718 podStartE2EDuration="12.425724354s" podCreationTimestamp="2025-08-13 00:03:32 +0000 UTC" firstStartedPulling="2025-08-13 00:03:32.872921675 +0000 UTC m=+65.620100583" lastFinishedPulling="2025-08-13 00:03:40.258496299 +0000 UTC m=+73.005675219" observedRunningTime="2025-08-13 00:03:41.415176305 +0000 UTC m=+74.162355249" watchObservedRunningTime="2025-08-13 00:03:44.425724354 +0000 UTC m=+77.172903262" Aug 13 00:03:44.428297 kubelet[2984]: I0813 00:03:44.428170 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-d75f5fd6d-27n9j" podStartSLOduration=27.091237197 podStartE2EDuration="34.428139596s" podCreationTimestamp="2025-08-13 00:03:10 +0000 UTC" firstStartedPulling="2025-08-13 00:03:36.096980257 +0000 UTC m=+68.844159153" lastFinishedPulling="2025-08-13 00:03:43.433882644 +0000 UTC m=+76.181061552" observedRunningTime="2025-08-13 00:03:44.427400248 +0000 UTC m=+77.174579156" watchObservedRunningTime="2025-08-13 00:03:44.428139596 +0000 UTC m=+77.175318516" Aug 13 00:03:44.498784 systemd[1]: run-containerd-runc-k8s.io-7ee0d4f88f3c0fde73248650354bfcaccd45af41c5d859d00eea0aeb5c899949-runc.LOMlxg.mount: Deactivated successfully. Aug 13 00:03:45.986903 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1519722679.mount: Deactivated successfully. Aug 13 00:03:46.516000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-172.31.27.151:22-139.178.89.65:58362 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:03:46.516865 systemd[1]: Started sshd@8-172.31.27.151:22-139.178.89.65:58362.service. Aug 13 00:03:46.519383 kernel: kauditd_printk_skb: 31 callbacks suppressed Aug 13 00:03:46.519489 kernel: audit: type=1130 audit(1755043426.516:426): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-172.31.27.151:22-139.178.89.65:58362 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:03:46.729000 audit[5407]: USER_ACCT pid=5407 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:46.741441 sshd[5407]: Accepted publickey for core from 139.178.89.65 port 58362 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:03:46.742262 kernel: audit: type=1101 audit(1755043426.729:427): pid=5407 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:46.742000 audit[5407]: CRED_ACQ pid=5407 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:46.755009 sshd[5407]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:03:46.756670 kernel: audit: type=1103 audit(1755043426.742:428): pid=5407 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:46.777538 kernel: audit: type=1006 audit(1755043426.742:429): pid=5407 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=9 res=1 Aug 13 00:03:46.777680 kernel: audit: type=1300 audit(1755043426.742:429): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffcb00e820 a2=3 a3=1 items=0 ppid=1 pid=5407 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=9 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:46.742000 audit[5407]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffcb00e820 a2=3 a3=1 items=0 ppid=1 pid=5407 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=9 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:46.742000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:03:46.783451 kernel: audit: type=1327 audit(1755043426.742:429): proctitle=737368643A20636F7265205B707269765D Aug 13 00:03:46.784790 systemd-logind[1901]: New session 9 of user core. Aug 13 00:03:46.786012 systemd[1]: Started session-9.scope. Aug 13 00:03:46.802000 audit[5407]: USER_START pid=5407 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:46.805000 audit[5410]: CRED_ACQ pid=5410 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:46.825427 kernel: audit: type=1105 audit(1755043426.802:430): pid=5407 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:46.825550 kernel: audit: type=1103 audit(1755043426.805:431): pid=5410 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:47.131482 sshd[5407]: pam_unix(sshd:session): session closed for user core Aug 13 00:03:47.133000 audit[5407]: USER_END pid=5407 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:47.137406 systemd[1]: sshd@8-172.31.27.151:22-139.178.89.65:58362.service: Deactivated successfully. Aug 13 00:03:47.138821 systemd[1]: session-9.scope: Deactivated successfully. Aug 13 00:03:47.141948 systemd-logind[1901]: Session 9 logged out. Waiting for processes to exit. Aug 13 00:03:47.147018 systemd-logind[1901]: Removed session 9. Aug 13 00:03:47.133000 audit[5407]: CRED_DISP pid=5407 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:47.157426 kernel: audit: type=1106 audit(1755043427.133:432): pid=5407 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:47.159360 kernel: audit: type=1104 audit(1755043427.133:433): pid=5407 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:47.136000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-172.31.27.151:22-139.178.89.65:58362 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:03:47.219296 env[1913]: time="2025-08-13T00:03:47.219238801Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/goldmane:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:47.224798 env[1913]: time="2025-08-13T00:03:47.224745513Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:1389d38feb576cfff09a57a2c028a53e51a72c658f295166960f770eaf07985f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:47.228791 env[1913]: time="2025-08-13T00:03:47.228704504Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/goldmane:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:47.232849 env[1913]: time="2025-08-13T00:03:47.232796887Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:47.234685 env[1913]: time="2025-08-13T00:03:47.234581489Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" returns image reference \"sha256:1389d38feb576cfff09a57a2c028a53e51a72c658f295166960f770eaf07985f\"" Aug 13 00:03:47.241824 env[1913]: time="2025-08-13T00:03:47.241738031Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Aug 13 00:03:47.244372 env[1913]: time="2025-08-13T00:03:47.244022844Z" level=info msg="CreateContainer within sandbox \"bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Aug 13 00:03:47.283978 env[1913]: time="2025-08-13T00:03:47.283387102Z" level=info msg="CreateContainer within sandbox \"bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"1b9102b7b9558849293d48e5ebe5883262c62e0e7362ffcc9c556a1dfae7d31d\"" Aug 13 00:03:47.286052 env[1913]: time="2025-08-13T00:03:47.284815519Z" level=info msg="StartContainer for \"1b9102b7b9558849293d48e5ebe5883262c62e0e7362ffcc9c556a1dfae7d31d\"" Aug 13 00:03:47.454957 env[1913]: time="2025-08-13T00:03:47.454866170Z" level=info msg="StartContainer for \"1b9102b7b9558849293d48e5ebe5883262c62e0e7362ffcc9c556a1dfae7d31d\" returns successfully" Aug 13 00:03:48.478658 kubelet[2984]: I0813 00:03:48.478573 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-58fd7646b9-6tzfz" podStartSLOduration=28.341561001 podStartE2EDuration="39.478483565s" podCreationTimestamp="2025-08-13 00:03:09 +0000 UTC" firstStartedPulling="2025-08-13 00:03:36.101520656 +0000 UTC m=+68.848699564" lastFinishedPulling="2025-08-13 00:03:47.238443232 +0000 UTC m=+79.985622128" observedRunningTime="2025-08-13 00:03:48.475998999 +0000 UTC m=+81.223177955" watchObservedRunningTime="2025-08-13 00:03:48.478483565 +0000 UTC m=+81.225662485" Aug 13 00:03:48.508000 audit[5463]: NETFILTER_CFG table=filter:120 family=2 entries=12 op=nft_register_rule pid=5463 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:48.508000 audit[5463]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=4504 a0=3 a1=fffffec7beb0 a2=0 a3=1 items=0 ppid=3113 pid=5463 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:48.508000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:48.523000 audit[5463]: NETFILTER_CFG table=nat:121 family=2 entries=22 op=nft_register_rule pid=5463 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:48.523000 audit[5463]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=6540 a0=3 a1=fffffec7beb0 a2=0 a3=1 items=0 ppid=3113 pid=5463 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:48.523000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:48.593477 systemd[1]: run-containerd-runc-k8s.io-1b9102b7b9558849293d48e5ebe5883262c62e0e7362ffcc9c556a1dfae7d31d-runc.QOlhWx.mount: Deactivated successfully. Aug 13 00:03:49.497370 systemd[1]: run-containerd-runc-k8s.io-1b9102b7b9558849293d48e5ebe5883262c62e0e7362ffcc9c556a1dfae7d31d-runc.A3VRlw.mount: Deactivated successfully. Aug 13 00:03:50.411653 env[1913]: time="2025-08-13T00:03:50.411588412Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/apiserver:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:50.414271 env[1913]: time="2025-08-13T00:03:50.414202422Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3371ea1b18040228ef58c964e49b96f4291def748753dfbc0aef87a55f906b8f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:50.416684 env[1913]: time="2025-08-13T00:03:50.416633995Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/apiserver:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:50.420546 env[1913]: time="2025-08-13T00:03:50.420472972Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:50.427397 env[1913]: time="2025-08-13T00:03:50.427310646Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:3371ea1b18040228ef58c964e49b96f4291def748753dfbc0aef87a55f906b8f\"" Aug 13 00:03:50.432250 env[1913]: time="2025-08-13T00:03:50.432195153Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Aug 13 00:03:50.437700 env[1913]: time="2025-08-13T00:03:50.437642319Z" level=info msg="CreateContainer within sandbox \"df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Aug 13 00:03:50.492991 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount888911583.mount: Deactivated successfully. Aug 13 00:03:50.499019 env[1913]: time="2025-08-13T00:03:50.498949672Z" level=info msg="CreateContainer within sandbox \"df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"f8fae5043f8a9a94184f948f132428be4cfcddbf04eb9eeb5371f64c8830bb4a\"" Aug 13 00:03:50.502144 env[1913]: time="2025-08-13T00:03:50.501750056Z" level=info msg="StartContainer for \"f8fae5043f8a9a94184f948f132428be4cfcddbf04eb9eeb5371f64c8830bb4a\"" Aug 13 00:03:50.742795 env[1913]: time="2025-08-13T00:03:50.742016986Z" level=info msg="StartContainer for \"f8fae5043f8a9a94184f948f132428be4cfcddbf04eb9eeb5371f64c8830bb4a\" returns successfully" Aug 13 00:03:50.746666 env[1913]: time="2025-08-13T00:03:50.745361632Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/apiserver:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:50.751689 env[1913]: time="2025-08-13T00:03:50.751632927Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:3371ea1b18040228ef58c964e49b96f4291def748753dfbc0aef87a55f906b8f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:50.757324 env[1913]: time="2025-08-13T00:03:50.757268482Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/apiserver:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:50.762387 env[1913]: time="2025-08-13T00:03:50.762326053Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:50.765320 env[1913]: time="2025-08-13T00:03:50.763927354Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:3371ea1b18040228ef58c964e49b96f4291def748753dfbc0aef87a55f906b8f\"" Aug 13 00:03:50.770792 env[1913]: time="2025-08-13T00:03:50.770732508Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\"" Aug 13 00:03:50.781455 env[1913]: time="2025-08-13T00:03:50.781366138Z" level=info msg="CreateContainer within sandbox \"c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Aug 13 00:03:50.813532 env[1913]: time="2025-08-13T00:03:50.813446899Z" level=info msg="CreateContainer within sandbox \"c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"5b95e84c30647dd733856e0fd59f53fb11885492e798dc259086d007ae4bf64a\"" Aug 13 00:03:50.816967 env[1913]: time="2025-08-13T00:03:50.816911126Z" level=info msg="StartContainer for \"5b95e84c30647dd733856e0fd59f53fb11885492e798dc259086d007ae4bf64a\"" Aug 13 00:03:50.984542 env[1913]: time="2025-08-13T00:03:50.984478680Z" level=info msg="StartContainer for \"5b95e84c30647dd733856e0fd59f53fb11885492e798dc259086d007ae4bf64a\" returns successfully" Aug 13 00:03:51.544070 kubelet[2984]: I0813 00:03:51.543017 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-6ff96b959-jggcm" podStartSLOduration=38.143957437 podStartE2EDuration="50.542995178s" podCreationTimestamp="2025-08-13 00:03:01 +0000 UTC" firstStartedPulling="2025-08-13 00:03:38.367737829 +0000 UTC m=+71.114916737" lastFinishedPulling="2025-08-13 00:03:50.766775558 +0000 UTC m=+83.513954478" observedRunningTime="2025-08-13 00:03:51.54262146 +0000 UTC m=+84.289800404" watchObservedRunningTime="2025-08-13 00:03:51.542995178 +0000 UTC m=+84.290174086" Aug 13 00:03:51.544070 kubelet[2984]: I0813 00:03:51.543822 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-6ff96b959-s8g8j" podStartSLOduration=35.985153167 podStartE2EDuration="49.543806562s" podCreationTimestamp="2025-08-13 00:03:02 +0000 UTC" firstStartedPulling="2025-08-13 00:03:36.872852654 +0000 UTC m=+69.620031562" lastFinishedPulling="2025-08-13 00:03:50.431506037 +0000 UTC m=+83.178684957" observedRunningTime="2025-08-13 00:03:51.509663965 +0000 UTC m=+84.256842909" watchObservedRunningTime="2025-08-13 00:03:51.543806562 +0000 UTC m=+84.290985458" Aug 13 00:03:51.573000 audit[5598]: NETFILTER_CFG table=filter:122 family=2 entries=12 op=nft_register_rule pid=5598 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:51.576731 kernel: kauditd_printk_skb: 7 callbacks suppressed Aug 13 00:03:51.576855 kernel: audit: type=1325 audit(1755043431.573:437): table=filter:122 family=2 entries=12 op=nft_register_rule pid=5598 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:51.573000 audit[5598]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=4504 a0=3 a1=ffffc29eda50 a2=0 a3=1 items=0 ppid=3113 pid=5598 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:51.605180 kernel: audit: type=1300 audit(1755043431.573:437): arch=c00000b7 syscall=211 success=yes exit=4504 a0=3 a1=ffffc29eda50 a2=0 a3=1 items=0 ppid=3113 pid=5598 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:51.573000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:51.613068 kernel: audit: type=1327 audit(1755043431.573:437): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:51.591000 audit[5598]: NETFILTER_CFG table=nat:123 family=2 entries=22 op=nft_register_rule pid=5598 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:51.620315 kernel: audit: type=1325 audit(1755043431.591:438): table=nat:123 family=2 entries=22 op=nft_register_rule pid=5598 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:51.591000 audit[5598]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=6540 a0=3 a1=ffffc29eda50 a2=0 a3=1 items=0 ppid=3113 pid=5598 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:51.634475 kernel: audit: type=1300 audit(1755043431.591:438): arch=c00000b7 syscall=211 success=yes exit=6540 a0=3 a1=ffffc29eda50 a2=0 a3=1 items=0 ppid=3113 pid=5598 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:51.591000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:51.643221 kernel: audit: type=1327 audit(1755043431.591:438): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:51.652813 systemd[1]: run-containerd-runc-k8s.io-1b9102b7b9558849293d48e5ebe5883262c62e0e7362ffcc9c556a1dfae7d31d-runc.NDTd9l.mount: Deactivated successfully. Aug 13 00:03:51.695000 audit[5615]: NETFILTER_CFG table=filter:124 family=2 entries=12 op=nft_register_rule pid=5615 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:51.729023 kernel: audit: type=1325 audit(1755043431.695:439): table=filter:124 family=2 entries=12 op=nft_register_rule pid=5615 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:51.729249 kernel: audit: type=1300 audit(1755043431.695:439): arch=c00000b7 syscall=211 success=yes exit=4504 a0=3 a1=fffffd5d5740 a2=0 a3=1 items=0 ppid=3113 pid=5615 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:51.695000 audit[5615]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=4504 a0=3 a1=fffffd5d5740 a2=0 a3=1 items=0 ppid=3113 pid=5615 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:51.778843 kernel: audit: type=1327 audit(1755043431.695:439): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:51.779001 kernel: audit: type=1325 audit(1755043431.712:440): table=nat:125 family=2 entries=22 op=nft_register_rule pid=5615 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:51.695000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:51.712000 audit[5615]: NETFILTER_CFG table=nat:125 family=2 entries=22 op=nft_register_rule pid=5615 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:51.776518 systemd[1]: run-containerd-runc-k8s.io-7ee0d4f88f3c0fde73248650354bfcaccd45af41c5d859d00eea0aeb5c899949-runc.iC6WMn.mount: Deactivated successfully. Aug 13 00:03:51.712000 audit[5615]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=6540 a0=3 a1=fffffd5d5740 a2=0 a3=1 items=0 ppid=3113 pid=5615 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:51.712000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:52.157000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-172.31.27.151:22-139.178.89.65:40810 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:03:52.158216 systemd[1]: Started sshd@9-172.31.27.151:22-139.178.89.65:40810.service. Aug 13 00:03:52.390000 audit[5641]: USER_ACCT pid=5641 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:52.391626 sshd[5641]: Accepted publickey for core from 139.178.89.65 port 40810 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:03:52.392000 audit[5641]: CRED_ACQ pid=5641 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:52.392000 audit[5641]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffff0285f90 a2=3 a3=1 items=0 ppid=1 pid=5641 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:52.392000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:03:52.395560 sshd[5641]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:03:52.407832 systemd[1]: Started session-10.scope. Aug 13 00:03:52.408754 systemd-logind[1901]: New session 10 of user core. Aug 13 00:03:52.428000 audit[5641]: USER_START pid=5641 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:52.431000 audit[5644]: CRED_ACQ pid=5644 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:52.504517 env[1913]: time="2025-08-13T00:03:52.504457368Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/csi:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:52.511247 env[1913]: time="2025-08-13T00:03:52.511189980Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:14ecfabbdbebd1f5a36708f8b11a95a43baddd6a935d7d78c89a9c333849fcd2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:52.517210 env[1913]: time="2025-08-13T00:03:52.517154192Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/csi:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:52.531060 env[1913]: time="2025-08-13T00:03:52.530999614Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:52.540096 env[1913]: time="2025-08-13T00:03:52.532558699Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\" returns image reference \"sha256:14ecfabbdbebd1f5a36708f8b11a95a43baddd6a935d7d78c89a9c333849fcd2\"" Aug 13 00:03:52.543925 env[1913]: time="2025-08-13T00:03:52.543862279Z" level=info msg="CreateContainer within sandbox \"e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Aug 13 00:03:52.589853 env[1913]: time="2025-08-13T00:03:52.589771380Z" level=info msg="CreateContainer within sandbox \"e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"5e1fd12fbac59d69d033ff7171ca54ebe6c900723fb031a3321595b1de387066\"" Aug 13 00:03:52.591097 env[1913]: time="2025-08-13T00:03:52.591026539Z" level=info msg="StartContainer for \"5e1fd12fbac59d69d033ff7171ca54ebe6c900723fb031a3321595b1de387066\"" Aug 13 00:03:52.608671 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount830491269.mount: Deactivated successfully. Aug 13 00:03:52.716647 systemd[1]: run-containerd-runc-k8s.io-5e1fd12fbac59d69d033ff7171ca54ebe6c900723fb031a3321595b1de387066-runc.EFY9Kd.mount: Deactivated successfully. Aug 13 00:03:52.959470 sshd[5641]: pam_unix(sshd:session): session closed for user core Aug 13 00:03:52.960000 audit[5641]: USER_END pid=5641 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:52.961000 audit[5641]: CRED_DISP pid=5641 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:52.968000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-172.31.27.151:22-139.178.89.65:40810 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:03:52.968864 systemd[1]: sshd@9-172.31.27.151:22-139.178.89.65:40810.service: Deactivated successfully. Aug 13 00:03:52.970580 systemd[1]: session-10.scope: Deactivated successfully. Aug 13 00:03:52.971981 systemd-logind[1901]: Session 10 logged out. Waiting for processes to exit. Aug 13 00:03:52.983658 systemd-logind[1901]: Removed session 10. Aug 13 00:03:52.988000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-172.31.27.151:22-139.178.89.65:40814 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:03:52.989409 systemd[1]: Started sshd@10-172.31.27.151:22-139.178.89.65:40814.service. Aug 13 00:03:53.040799 env[1913]: time="2025-08-13T00:03:53.040721032Z" level=info msg="StartContainer for \"5e1fd12fbac59d69d033ff7171ca54ebe6c900723fb031a3321595b1de387066\" returns successfully" Aug 13 00:03:53.042578 env[1913]: time="2025-08-13T00:03:53.042514417Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\"" Aug 13 00:03:53.185281 sshd[5678]: Accepted publickey for core from 139.178.89.65 port 40814 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:03:53.184000 audit[5678]: USER_ACCT pid=5678 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:53.187000 audit[5678]: CRED_ACQ pid=5678 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:53.187000 audit[5678]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffefea7a70 a2=3 a3=1 items=0 ppid=1 pid=5678 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=11 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:53.187000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:03:53.189638 sshd[5678]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:03:53.200202 systemd-logind[1901]: New session 11 of user core. Aug 13 00:03:53.201569 systemd[1]: Started session-11.scope. Aug 13 00:03:53.233000 audit[5678]: USER_START pid=5678 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:53.237000 audit[5690]: CRED_ACQ pid=5690 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:53.499348 kubelet[2984]: I0813 00:03:53.498606 2984 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:03:53.745054 sshd[5678]: pam_unix(sshd:session): session closed for user core Aug 13 00:03:53.749000 audit[5678]: USER_END pid=5678 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:53.749000 audit[5678]: CRED_DISP pid=5678 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:53.755685 systemd-logind[1901]: Session 11 logged out. Waiting for processes to exit. Aug 13 00:03:53.757694 systemd[1]: sshd@10-172.31.27.151:22-139.178.89.65:40814.service: Deactivated successfully. Aug 13 00:03:53.757000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-172.31.27.151:22-139.178.89.65:40814 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:03:53.759229 systemd[1]: session-11.scope: Deactivated successfully. Aug 13 00:03:53.764954 systemd-logind[1901]: Removed session 11. Aug 13 00:03:53.790620 systemd[1]: Started sshd@11-172.31.27.151:22-139.178.89.65:40820.service. Aug 13 00:03:53.789000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-172.31.27.151:22-139.178.89.65:40820 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:03:54.016000 audit[5698]: USER_ACCT pid=5698 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:54.019409 sshd[5698]: Accepted publickey for core from 139.178.89.65 port 40820 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:03:54.019000 audit[5698]: CRED_ACQ pid=5698 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:54.019000 audit[5698]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffe336a5a0 a2=3 a3=1 items=0 ppid=1 pid=5698 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=12 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:54.019000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:03:54.021857 sshd[5698]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:03:54.032518 systemd-logind[1901]: New session 12 of user core. Aug 13 00:03:54.034728 systemd[1]: Started session-12.scope. Aug 13 00:03:54.054000 audit[5698]: USER_START pid=5698 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:54.056000 audit[5701]: CRED_ACQ pid=5701 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:54.537505 sshd[5698]: pam_unix(sshd:session): session closed for user core Aug 13 00:03:54.538000 audit[5698]: USER_END pid=5698 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:54.538000 audit[5698]: CRED_DISP pid=5698 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:54.542215 systemd[1]: sshd@11-172.31.27.151:22-139.178.89.65:40820.service: Deactivated successfully. Aug 13 00:03:54.541000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-172.31.27.151:22-139.178.89.65:40820 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:03:54.543703 systemd[1]: session-12.scope: Deactivated successfully. Aug 13 00:03:54.545681 systemd-logind[1901]: Session 12 logged out. Waiting for processes to exit. Aug 13 00:03:54.549355 systemd-logind[1901]: Removed session 12. Aug 13 00:03:54.561000 audit[5711]: NETFILTER_CFG table=filter:126 family=2 entries=11 op=nft_register_rule pid=5711 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:54.561000 audit[5711]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=3760 a0=3 a1=ffffc56ea300 a2=0 a3=1 items=0 ppid=3113 pid=5711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:54.561000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:54.569000 audit[5711]: NETFILTER_CFG table=nat:127 family=2 entries=29 op=nft_register_chain pid=5711 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:54.569000 audit[5711]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=10116 a0=3 a1=ffffc56ea300 a2=0 a3=1 items=0 ppid=3113 pid=5711 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:54.569000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:54.922416 env[1913]: time="2025-08-13T00:03:54.922363318Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:54.927060 env[1913]: time="2025-08-13T00:03:54.926989858Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:664ed31fb4687b0de23d6e6e116bc87b236790d7355871d3237c54452e02e27c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:54.933341 env[1913]: time="2025-08-13T00:03:54.932269334Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:54.938674 env[1913]: time="2025-08-13T00:03:54.935725076Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Aug 13 00:03:54.938674 env[1913]: time="2025-08-13T00:03:54.936327779Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" returns image reference \"sha256:664ed31fb4687b0de23d6e6e116bc87b236790d7355871d3237c54452e02e27c\"" Aug 13 00:03:54.949132 env[1913]: time="2025-08-13T00:03:54.941881900Z" level=info msg="CreateContainer within sandbox \"e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Aug 13 00:03:54.975936 env[1913]: time="2025-08-13T00:03:54.975807972Z" level=info msg="CreateContainer within sandbox \"e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"855d6e2045364641ccbb029b3f94c5bc2c83cfc2f64688760de7d8814f8a1c3e\"" Aug 13 00:03:54.981363 env[1913]: time="2025-08-13T00:03:54.981307757Z" level=info msg="StartContainer for \"855d6e2045364641ccbb029b3f94c5bc2c83cfc2f64688760de7d8814f8a1c3e\"" Aug 13 00:03:55.093007 systemd[1]: run-containerd-runc-k8s.io-855d6e2045364641ccbb029b3f94c5bc2c83cfc2f64688760de7d8814f8a1c3e-runc.DRpX7u.mount: Deactivated successfully. Aug 13 00:03:55.250292 env[1913]: time="2025-08-13T00:03:55.250086588Z" level=info msg="StartContainer for \"855d6e2045364641ccbb029b3f94c5bc2c83cfc2f64688760de7d8814f8a1c3e\" returns successfully" Aug 13 00:03:55.758015 kubelet[2984]: I0813 00:03:55.757922 2984 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-n4g7q" podStartSLOduration=30.641922943 podStartE2EDuration="45.757899996s" podCreationTimestamp="2025-08-13 00:03:10 +0000 UTC" firstStartedPulling="2025-08-13 00:03:39.823690583 +0000 UTC m=+72.570869491" lastFinishedPulling="2025-08-13 00:03:54.939667636 +0000 UTC m=+87.686846544" observedRunningTime="2025-08-13 00:03:55.538543035 +0000 UTC m=+88.285721979" watchObservedRunningTime="2025-08-13 00:03:55.757899996 +0000 UTC m=+88.505078904" Aug 13 00:03:55.788000 audit[5760]: NETFILTER_CFG table=filter:128 family=2 entries=10 op=nft_register_rule pid=5760 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:55.788000 audit[5760]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=3760 a0=3 a1=ffffc5cf72f0 a2=0 a3=1 items=0 ppid=3113 pid=5760 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:55.788000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:55.797000 audit[5760]: NETFILTER_CFG table=nat:129 family=2 entries=36 op=nft_register_chain pid=5760 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:03:55.797000 audit[5760]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=12004 a0=3 a1=ffffc5cf72f0 a2=0 a3=1 items=0 ppid=3113 pid=5760 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:55.797000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:03:55.832727 kubelet[2984]: I0813 00:03:55.832693 2984 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Aug 13 00:03:55.832936 kubelet[2984]: I0813 00:03:55.832915 2984 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Aug 13 00:03:59.561973 systemd[1]: Started sshd@12-172.31.27.151:22-139.178.89.65:34642.service. Aug 13 00:03:59.568335 kernel: kauditd_printk_skb: 47 callbacks suppressed Aug 13 00:03:59.568426 kernel: audit: type=1130 audit(1755043439.561:472): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-172.31.27.151:22-139.178.89.65:34642 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:03:59.561000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-172.31.27.151:22-139.178.89.65:34642 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:03:59.763000 audit[5761]: USER_ACCT pid=5761 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:59.764684 sshd[5761]: Accepted publickey for core from 139.178.89.65 port 34642 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:03:59.768401 sshd[5761]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:03:59.765000 audit[5761]: CRED_ACQ pid=5761 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:59.786182 kernel: audit: type=1101 audit(1755043439.763:473): pid=5761 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:59.786346 kernel: audit: type=1103 audit(1755043439.765:474): pid=5761 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:59.793157 systemd-logind[1901]: New session 13 of user core. Aug 13 00:03:59.795544 systemd[1]: Started session-13.scope. Aug 13 00:03:59.803959 kernel: audit: type=1006 audit(1755043439.766:475): pid=5761 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=13 res=1 Aug 13 00:03:59.766000 audit[5761]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffe9d461e0 a2=3 a3=1 items=0 ppid=1 pid=5761 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=13 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:59.815429 kernel: audit: type=1300 audit(1755043439.766:475): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffe9d461e0 a2=3 a3=1 items=0 ppid=1 pid=5761 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=13 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:03:59.766000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:03:59.823318 kernel: audit: type=1327 audit(1755043439.766:475): proctitle=737368643A20636F7265205B707269765D Aug 13 00:03:59.825000 audit[5761]: USER_START pid=5761 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:59.839000 audit[5767]: CRED_ACQ pid=5767 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:59.851893 kernel: audit: type=1105 audit(1755043439.825:476): pid=5761 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:03:59.852036 kernel: audit: type=1103 audit(1755043439.839:477): pid=5767 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:00.110976 sshd[5761]: pam_unix(sshd:session): session closed for user core Aug 13 00:04:00.112000 audit[5761]: USER_END pid=5761 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:00.118040 systemd-logind[1901]: Session 13 logged out. Waiting for processes to exit. Aug 13 00:04:00.121041 systemd[1]: sshd@12-172.31.27.151:22-139.178.89.65:34642.service: Deactivated successfully. Aug 13 00:04:00.122482 systemd[1]: session-13.scope: Deactivated successfully. Aug 13 00:04:00.125225 systemd-logind[1901]: Removed session 13. Aug 13 00:04:00.112000 audit[5761]: CRED_DISP pid=5761 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:00.135970 kernel: audit: type=1106 audit(1755043440.112:478): pid=5761 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:00.136191 kernel: audit: type=1104 audit(1755043440.112:479): pid=5761 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:00.118000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-172.31.27.151:22-139.178.89.65:34642 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:05.137192 systemd[1]: Started sshd@13-172.31.27.151:22-139.178.89.65:34656.service. Aug 13 00:04:05.140702 kernel: kauditd_printk_skb: 1 callbacks suppressed Aug 13 00:04:05.140780 kernel: audit: type=1130 audit(1755043445.136:481): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-172.31.27.151:22-139.178.89.65:34656 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:05.136000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-172.31.27.151:22-139.178.89.65:34656 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:05.314000 audit[5783]: USER_ACCT pid=5783 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:05.316117 sshd[5783]: Accepted publickey for core from 139.178.89.65 port 34656 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:04:05.327208 kernel: audit: type=1101 audit(1755043445.314:482): pid=5783 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:05.327000 audit[5783]: CRED_ACQ pid=5783 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:05.332037 sshd[5783]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:04:05.345572 kernel: audit: type=1103 audit(1755043445.327:483): pid=5783 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:05.345710 kernel: audit: type=1006 audit(1755043445.327:484): pid=5783 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=14 res=1 Aug 13 00:04:05.327000 audit[5783]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffff85aecc0 a2=3 a3=1 items=0 ppid=1 pid=5783 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:05.357945 kernel: audit: type=1300 audit(1755043445.327:484): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffff85aecc0 a2=3 a3=1 items=0 ppid=1 pid=5783 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:05.354563 systemd[1]: Started session-14.scope. Aug 13 00:04:05.356193 systemd-logind[1901]: New session 14 of user core. Aug 13 00:04:05.327000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:05.375175 kernel: audit: type=1327 audit(1755043445.327:484): proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:05.376000 audit[5783]: USER_START pid=5783 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:05.379000 audit[5786]: CRED_ACQ pid=5786 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:05.400259 kernel: audit: type=1105 audit(1755043445.376:485): pid=5783 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:05.400407 kernel: audit: type=1103 audit(1755043445.379:486): pid=5786 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:05.670520 sshd[5783]: pam_unix(sshd:session): session closed for user core Aug 13 00:04:05.671000 audit[5783]: USER_END pid=5783 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:05.677410 systemd-logind[1901]: Session 14 logged out. Waiting for processes to exit. Aug 13 00:04:05.679945 systemd[1]: sshd@13-172.31.27.151:22-139.178.89.65:34656.service: Deactivated successfully. Aug 13 00:04:05.682514 systemd[1]: session-14.scope: Deactivated successfully. Aug 13 00:04:05.685790 systemd-logind[1901]: Removed session 14. Aug 13 00:04:05.671000 audit[5783]: CRED_DISP pid=5783 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:05.692133 kernel: audit: type=1106 audit(1755043445.671:487): pid=5783 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:05.679000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-172.31.27.151:22-139.178.89.65:34656 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:05.703183 kernel: audit: type=1104 audit(1755043445.671:488): pid=5783 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:10.715407 kernel: kauditd_printk_skb: 1 callbacks suppressed Aug 13 00:04:10.715784 kernel: audit: type=1130 audit(1755043450.696:490): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-172.31.27.151:22-139.178.89.65:60948 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:10.696000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-172.31.27.151:22-139.178.89.65:60948 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:10.697249 systemd[1]: Started sshd@14-172.31.27.151:22-139.178.89.65:60948.service. Aug 13 00:04:10.786249 systemd[1]: run-containerd-runc-k8s.io-be0ed1e29b588d67e52fae942c6b3420ba4d8c77c6e0d8347717d2a5797a2a37-runc.4sJsdQ.mount: Deactivated successfully. Aug 13 00:04:10.911000 audit[5796]: USER_ACCT pid=5796 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:10.916010 systemd[1]: run-containerd-runc-k8s.io-7ee0d4f88f3c0fde73248650354bfcaccd45af41c5d859d00eea0aeb5c899949-runc.inxfau.mount: Deactivated successfully. Aug 13 00:04:10.915651 sshd[5796]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:04:10.924714 sshd[5796]: Accepted publickey for core from 139.178.89.65 port 60948 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:04:10.911000 audit[5796]: CRED_ACQ pid=5796 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:10.939861 kernel: audit: type=1101 audit(1755043450.911:491): pid=5796 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:10.940053 kernel: audit: type=1103 audit(1755043450.911:492): pid=5796 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:10.946658 kernel: audit: type=1006 audit(1755043450.911:493): pid=5796 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=15 res=1 Aug 13 00:04:10.911000 audit[5796]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffc3ba3d40 a2=3 a3=1 items=0 ppid=1 pid=5796 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:10.946897 systemd[1]: Started session-15.scope. Aug 13 00:04:10.949023 systemd-logind[1901]: New session 15 of user core. Aug 13 00:04:10.984005 kernel: audit: type=1300 audit(1755043450.911:493): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffc3ba3d40 a2=3 a3=1 items=0 ppid=1 pid=5796 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:10.984174 kernel: audit: type=1327 audit(1755043450.911:493): proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:10.911000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:11.003063 kernel: audit: type=1105 audit(1755043450.979:494): pid=5796 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:10.979000 audit[5796]: USER_START pid=5796 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:10.986000 audit[5822]: CRED_ACQ pid=5822 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:11.050169 kernel: audit: type=1103 audit(1755043450.986:495): pid=5822 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:11.328625 sshd[5796]: pam_unix(sshd:session): session closed for user core Aug 13 00:04:11.329000 audit[5796]: USER_END pid=5796 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:11.334398 systemd-logind[1901]: Session 15 logged out. Waiting for processes to exit. Aug 13 00:04:11.335680 systemd[1]: sshd@14-172.31.27.151:22-139.178.89.65:60948.service: Deactivated successfully. Aug 13 00:04:11.337098 systemd[1]: session-15.scope: Deactivated successfully. Aug 13 00:04:11.339331 systemd-logind[1901]: Removed session 15. Aug 13 00:04:11.329000 audit[5796]: CRED_DISP pid=5796 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:11.357744 kernel: audit: type=1106 audit(1755043451.329:496): pid=5796 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:11.357913 kernel: audit: type=1104 audit(1755043451.329:497): pid=5796 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:11.335000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-172.31.27.151:22-139.178.89.65:60948 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:16.352000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-172.31.27.151:22-139.178.89.65:60954 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:16.353664 systemd[1]: Started sshd@15-172.31.27.151:22-139.178.89.65:60954.service. Aug 13 00:04:16.356137 kernel: kauditd_printk_skb: 1 callbacks suppressed Aug 13 00:04:16.357269 kernel: audit: type=1130 audit(1755043456.352:499): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-172.31.27.151:22-139.178.89.65:60954 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:16.588000 audit[5853]: USER_ACCT pid=5853 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:16.600609 sshd[5853]: Accepted publickey for core from 139.178.89.65 port 60954 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:04:16.603831 sshd[5853]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:04:16.601000 audit[5853]: CRED_ACQ pid=5853 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:16.620321 kernel: audit: type=1101 audit(1755043456.588:500): pid=5853 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:16.620488 kernel: audit: type=1103 audit(1755043456.601:501): pid=5853 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:16.614401 systemd[1]: Started session-16.scope. Aug 13 00:04:16.616620 systemd-logind[1901]: New session 16 of user core. Aug 13 00:04:16.650348 kernel: audit: type=1006 audit(1755043456.601:502): pid=5853 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=16 res=1 Aug 13 00:04:16.601000 audit[5853]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffc8c6d390 a2=3 a3=1 items=0 ppid=1 pid=5853 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:16.601000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:16.670855 kernel: audit: type=1300 audit(1755043456.601:502): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffc8c6d390 a2=3 a3=1 items=0 ppid=1 pid=5853 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:16.670955 kernel: audit: type=1327 audit(1755043456.601:502): proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:16.636000 audit[5853]: USER_START pid=5853 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:16.683014 kernel: audit: type=1105 audit(1755043456.636:503): pid=5853 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:16.639000 audit[5856]: CRED_ACQ pid=5856 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:16.692913 kernel: audit: type=1103 audit(1755043456.639:504): pid=5856 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:16.950321 sshd[5853]: pam_unix(sshd:session): session closed for user core Aug 13 00:04:16.952000 audit[5853]: USER_END pid=5853 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:16.956057 systemd-logind[1901]: Session 16 logged out. Waiting for processes to exit. Aug 13 00:04:16.968575 systemd[1]: sshd@15-172.31.27.151:22-139.178.89.65:60954.service: Deactivated successfully. Aug 13 00:04:16.970195 systemd[1]: session-16.scope: Deactivated successfully. Aug 13 00:04:16.952000 audit[5853]: CRED_DISP pid=5853 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:16.975140 kernel: audit: type=1106 audit(1755043456.952:505): pid=5853 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:16.976778 systemd[1]: Started sshd@16-172.31.27.151:22-139.178.89.65:60966.service. Aug 13 00:04:16.967000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-172.31.27.151:22-139.178.89.65:60954 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:16.984000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-172.31.27.151:22-139.178.89.65:60966 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:17.000142 kernel: audit: type=1104 audit(1755043456.952:506): pid=5853 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:17.024229 systemd-logind[1901]: Removed session 16. Aug 13 00:04:17.209307 sshd[5866]: Accepted publickey for core from 139.178.89.65 port 60966 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:04:17.207000 audit[5866]: USER_ACCT pid=5866 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:17.211000 audit[5866]: CRED_ACQ pid=5866 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:17.211000 audit[5866]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffdc40eda0 a2=3 a3=1 items=0 ppid=1 pid=5866 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=17 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:17.211000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:17.214186 sshd[5866]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:04:17.226207 systemd-logind[1901]: New session 17 of user core. Aug 13 00:04:17.229407 systemd[1]: Started session-17.scope. Aug 13 00:04:17.240000 audit[5866]: USER_START pid=5866 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:17.243000 audit[5869]: CRED_ACQ pid=5869 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:17.652334 systemd[1]: run-containerd-runc-k8s.io-1b9102b7b9558849293d48e5ebe5883262c62e0e7362ffcc9c556a1dfae7d31d-runc.kRUt9k.mount: Deactivated successfully. Aug 13 00:04:18.026226 sshd[5866]: pam_unix(sshd:session): session closed for user core Aug 13 00:04:18.027000 audit[5866]: USER_END pid=5866 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:18.027000 audit[5866]: CRED_DISP pid=5866 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:18.032672 systemd[1]: sshd@16-172.31.27.151:22-139.178.89.65:60966.service: Deactivated successfully. Aug 13 00:04:18.032000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-172.31.27.151:22-139.178.89.65:60966 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:18.034199 systemd-logind[1901]: Session 17 logged out. Waiting for processes to exit. Aug 13 00:04:18.035210 systemd[1]: session-17.scope: Deactivated successfully. Aug 13 00:04:18.038350 systemd-logind[1901]: Removed session 17. Aug 13 00:04:18.054764 systemd[1]: Started sshd@17-172.31.27.151:22-139.178.89.65:60970.service. Aug 13 00:04:18.054000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-172.31.27.151:22-139.178.89.65:60970 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:18.263686 sshd[5895]: Accepted publickey for core from 139.178.89.65 port 60970 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:04:18.262000 audit[5895]: USER_ACCT pid=5895 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:18.264000 audit[5895]: CRED_ACQ pid=5895 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:18.264000 audit[5895]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffed20f9c0 a2=3 a3=1 items=0 ppid=1 pid=5895 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=18 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:18.264000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:18.266885 sshd[5895]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:04:18.277022 systemd[1]: Started session-18.scope. Aug 13 00:04:18.284673 systemd-logind[1901]: New session 18 of user core. Aug 13 00:04:18.301000 audit[5895]: USER_START pid=5895 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:18.304000 audit[5898]: CRED_ACQ pid=5898 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:21.568647 systemd[1]: run-containerd-runc-k8s.io-1b9102b7b9558849293d48e5ebe5883262c62e0e7362ffcc9c556a1dfae7d31d-runc.XNGSWT.mount: Deactivated successfully. Aug 13 00:04:22.343938 kernel: kauditd_printk_skb: 20 callbacks suppressed Aug 13 00:04:22.344145 kernel: audit: type=1325 audit(1755043462.334:523): table=filter:130 family=2 entries=9 op=nft_register_rule pid=5948 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:04:22.334000 audit[5948]: NETFILTER_CFG table=filter:130 family=2 entries=9 op=nft_register_rule pid=5948 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:04:22.334000 audit[5948]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=3016 a0=3 a1=ffffe38412f0 a2=0 a3=1 items=0 ppid=3113 pid=5948 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:22.334000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:04:22.386875 kernel: audit: type=1300 audit(1755043462.334:523): arch=c00000b7 syscall=211 success=yes exit=3016 a0=3 a1=ffffe38412f0 a2=0 a3=1 items=0 ppid=3113 pid=5948 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:22.387041 kernel: audit: type=1327 audit(1755043462.334:523): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:04:22.499000 audit[5948]: NETFILTER_CFG table=nat:131 family=2 entries=31 op=nft_register_chain pid=5948 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:04:22.499000 audit[5948]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=10884 a0=3 a1=ffffe38412f0 a2=0 a3=1 items=0 ppid=3113 pid=5948 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:22.527236 kernel: audit: type=1325 audit(1755043462.499:524): table=nat:131 family=2 entries=31 op=nft_register_chain pid=5948 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:04:22.527434 kernel: audit: type=1300 audit(1755043462.499:524): arch=c00000b7 syscall=211 success=yes exit=10884 a0=3 a1=ffffe38412f0 a2=0 a3=1 items=0 ppid=3113 pid=5948 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:22.499000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:04:22.539900 kernel: audit: type=1327 audit(1755043462.499:524): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:04:22.930959 sshd[5895]: pam_unix(sshd:session): session closed for user core Aug 13 00:04:22.933000 audit[5895]: USER_END pid=5895 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:22.951513 systemd[1]: sshd@17-172.31.27.151:22-139.178.89.65:60970.service: Deactivated successfully. Aug 13 00:04:22.961890 systemd[1]: Started sshd@18-172.31.27.151:22-139.178.89.65:47478.service. Aug 13 00:04:22.962600 systemd-logind[1901]: Session 18 logged out. Waiting for processes to exit. Aug 13 00:04:22.964484 systemd[1]: session-18.scope: Deactivated successfully. Aug 13 00:04:22.933000 audit[5895]: CRED_DISP pid=5895 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:22.990214 kernel: audit: type=1106 audit(1755043462.933:525): pid=5895 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:22.990322 kernel: audit: type=1104 audit(1755043462.933:526): pid=5895 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:22.990277 systemd-logind[1901]: Removed session 18. Aug 13 00:04:22.950000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-172.31.27.151:22-139.178.89.65:60970 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:23.011629 kernel: audit: type=1131 audit(1755043462.950:527): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-172.31.27.151:22-139.178.89.65:60970 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:22.962000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-172.31.27.151:22-139.178.89.65:47478 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:23.035574 kernel: audit: type=1130 audit(1755043462.962:528): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-172.31.27.151:22-139.178.89.65:47478 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:23.060000 audit[5954]: NETFILTER_CFG table=filter:132 family=2 entries=8 op=nft_register_rule pid=5954 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:04:23.060000 audit[5954]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=3016 a0=3 a1=fffff0e0b7d0 a2=0 a3=1 items=0 ppid=3113 pid=5954 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:23.060000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:04:23.066000 audit[5954]: NETFILTER_CFG table=nat:133 family=2 entries=26 op=nft_register_rule pid=5954 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:04:23.066000 audit[5954]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=8076 a0=3 a1=fffff0e0b7d0 a2=0 a3=1 items=0 ppid=3113 pid=5954 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:23.066000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:04:23.222000 audit[5952]: USER_ACCT pid=5952 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:23.225960 sshd[5952]: Accepted publickey for core from 139.178.89.65 port 47478 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:04:23.225000 audit[5952]: CRED_ACQ pid=5952 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:23.226000 audit[5952]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffff0a137a0 a2=3 a3=1 items=0 ppid=1 pid=5952 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=19 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:23.226000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:23.227926 sshd[5952]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:04:23.242350 systemd-logind[1901]: New session 19 of user core. Aug 13 00:04:23.244548 systemd[1]: Started session-19.scope. Aug 13 00:04:23.265000 audit[5952]: USER_START pid=5952 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:23.268000 audit[5956]: CRED_ACQ pid=5956 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:23.904334 sshd[5952]: pam_unix(sshd:session): session closed for user core Aug 13 00:04:23.905000 audit[5952]: USER_END pid=5952 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:23.905000 audit[5952]: CRED_DISP pid=5952 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:23.910202 systemd-logind[1901]: Session 19 logged out. Waiting for processes to exit. Aug 13 00:04:23.911856 systemd[1]: sshd@18-172.31.27.151:22-139.178.89.65:47478.service: Deactivated successfully. Aug 13 00:04:23.911000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-172.31.27.151:22-139.178.89.65:47478 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:23.913622 systemd[1]: session-19.scope: Deactivated successfully. Aug 13 00:04:23.915591 systemd-logind[1901]: Removed session 19. Aug 13 00:04:23.929000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-172.31.27.151:22-139.178.89.65:47490 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:23.930653 systemd[1]: Started sshd@19-172.31.27.151:22-139.178.89.65:47490.service. Aug 13 00:04:24.104000 audit[5967]: NETFILTER_CFG table=filter:134 family=2 entries=20 op=nft_register_rule pid=5967 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:04:24.104000 audit[5967]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=11944 a0=3 a1=ffffe7cfb3b0 a2=0 a3=1 items=0 ppid=3113 pid=5967 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:24.104000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:04:24.109000 audit[5967]: NETFILTER_CFG table=nat:135 family=2 entries=26 op=nft_register_rule pid=5967 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:04:24.109000 audit[5967]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=8076 a0=3 a1=ffffe7cfb3b0 a2=0 a3=1 items=0 ppid=3113 pid=5967 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:24.109000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:04:24.118000 audit[5964]: USER_ACCT pid=5964 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:24.119787 sshd[5964]: Accepted publickey for core from 139.178.89.65 port 47490 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:04:24.121000 audit[5964]: CRED_ACQ pid=5964 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:24.121000 audit[5964]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffd1f83e60 a2=3 a3=1 items=0 ppid=1 pid=5964 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=20 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:24.121000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:24.123646 sshd[5964]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:04:24.136231 systemd-logind[1901]: New session 20 of user core. Aug 13 00:04:24.139510 systemd[1]: Started session-20.scope. Aug 13 00:04:24.151000 audit[5964]: USER_START pid=5964 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:24.154000 audit[5969]: CRED_ACQ pid=5969 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:24.534261 sshd[5964]: pam_unix(sshd:session): session closed for user core Aug 13 00:04:24.534000 audit[5964]: USER_END pid=5964 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:24.535000 audit[5964]: CRED_DISP pid=5964 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:24.539815 systemd-logind[1901]: Session 20 logged out. Waiting for processes to exit. Aug 13 00:04:24.541000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-172.31.27.151:22-139.178.89.65:47490 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:24.541733 systemd[1]: sshd@19-172.31.27.151:22-139.178.89.65:47490.service: Deactivated successfully. Aug 13 00:04:24.543244 systemd[1]: session-20.scope: Deactivated successfully. Aug 13 00:04:24.547633 systemd-logind[1901]: Removed session 20. Aug 13 00:04:27.608998 env[1913]: time="2025-08-13T00:04:27.608915442Z" level=info msg="StopPodSandbox for \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\"" Aug 13 00:04:27.868869 env[1913]: 2025-08-13 00:04:27.743 [WARNING][5988] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0", GenerateName:"calico-apiserver-6ff96b959-", Namespace:"calico-apiserver", SelfLink:"", UID:"71cf2b5c-a9b2-457e-a496-c93a3a2d8126", ResourceVersion:"1212", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6ff96b959", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19", Pod:"calico-apiserver-6ff96b959-s8g8j", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.1.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali9af93f385cf", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:04:27.868869 env[1913]: 2025-08-13 00:04:27.744 [INFO][5988] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Aug 13 00:04:27.868869 env[1913]: 2025-08-13 00:04:27.744 [INFO][5988] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" iface="eth0" netns="" Aug 13 00:04:27.868869 env[1913]: 2025-08-13 00:04:27.744 [INFO][5988] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Aug 13 00:04:27.868869 env[1913]: 2025-08-13 00:04:27.744 [INFO][5988] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Aug 13 00:04:27.868869 env[1913]: 2025-08-13 00:04:27.838 [INFO][5995] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" HandleID="k8s-pod-network.df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:04:27.868869 env[1913]: 2025-08-13 00:04:27.838 [INFO][5995] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:27.868869 env[1913]: 2025-08-13 00:04:27.838 [INFO][5995] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:27.868869 env[1913]: 2025-08-13 00:04:27.853 [WARNING][5995] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" HandleID="k8s-pod-network.df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:04:27.868869 env[1913]: 2025-08-13 00:04:27.853 [INFO][5995] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" HandleID="k8s-pod-network.df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:04:27.868869 env[1913]: 2025-08-13 00:04:27.860 [INFO][5995] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:27.868869 env[1913]: 2025-08-13 00:04:27.864 [INFO][5988] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Aug 13 00:04:27.868869 env[1913]: time="2025-08-13T00:04:27.868505116Z" level=info msg="TearDown network for sandbox \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\" successfully" Aug 13 00:04:27.868869 env[1913]: time="2025-08-13T00:04:27.868553189Z" level=info msg="StopPodSandbox for \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\" returns successfully" Aug 13 00:04:27.872173 env[1913]: time="2025-08-13T00:04:27.871040318Z" level=info msg="RemovePodSandbox for \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\"" Aug 13 00:04:27.872173 env[1913]: time="2025-08-13T00:04:27.871141034Z" level=info msg="Forcibly stopping sandbox \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\"" Aug 13 00:04:28.150380 env[1913]: 2025-08-13 00:04:27.983 [WARNING][6009] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0", GenerateName:"calico-apiserver-6ff96b959-", Namespace:"calico-apiserver", SelfLink:"", UID:"71cf2b5c-a9b2-457e-a496-c93a3a2d8126", ResourceVersion:"1212", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6ff96b959", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"df7aa80981f6ddec445a35cf1ece12895f6d0dc189352178705bb41e379e1c19", Pod:"calico-apiserver-6ff96b959-s8g8j", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.1.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali9af93f385cf", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:04:28.150380 env[1913]: 2025-08-13 00:04:27.984 [INFO][6009] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Aug 13 00:04:28.150380 env[1913]: 2025-08-13 00:04:27.984 [INFO][6009] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" iface="eth0" netns="" Aug 13 00:04:28.150380 env[1913]: 2025-08-13 00:04:27.984 [INFO][6009] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Aug 13 00:04:28.150380 env[1913]: 2025-08-13 00:04:27.984 [INFO][6009] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Aug 13 00:04:28.150380 env[1913]: 2025-08-13 00:04:28.096 [INFO][6016] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" HandleID="k8s-pod-network.df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:04:28.150380 env[1913]: 2025-08-13 00:04:28.107 [INFO][6016] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:28.150380 env[1913]: 2025-08-13 00:04:28.107 [INFO][6016] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:28.150380 env[1913]: 2025-08-13 00:04:28.138 [WARNING][6016] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" HandleID="k8s-pod-network.df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:04:28.150380 env[1913]: 2025-08-13 00:04:28.138 [INFO][6016] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" HandleID="k8s-pod-network.df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--s8g8j-eth0" Aug 13 00:04:28.150380 env[1913]: 2025-08-13 00:04:28.141 [INFO][6016] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:28.150380 env[1913]: 2025-08-13 00:04:28.143 [INFO][6009] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c" Aug 13 00:04:28.150380 env[1913]: time="2025-08-13T00:04:28.150344562Z" level=info msg="TearDown network for sandbox \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\" successfully" Aug 13 00:04:28.159683 env[1913]: time="2025-08-13T00:04:28.159569489Z" level=info msg="RemovePodSandbox \"df8e6b65fde9fa45520048922f4d4c98edc75b3c181f63ef454f307575cf0a4c\" returns successfully" Aug 13 00:04:28.160574 env[1913]: time="2025-08-13T00:04:28.160506476Z" level=info msg="StopPodSandbox for \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\"" Aug 13 00:04:28.363602 env[1913]: 2025-08-13 00:04:28.281 [WARNING][6030] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0", GenerateName:"calico-apiserver-6ff96b959-", Namespace:"calico-apiserver", SelfLink:"", UID:"6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e", ResourceVersion:"1227", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6ff96b959", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9", Pod:"calico-apiserver-6ff96b959-jggcm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.1.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3abb69b4f11", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:04:28.363602 env[1913]: 2025-08-13 00:04:28.282 [INFO][6030] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Aug 13 00:04:28.363602 env[1913]: 2025-08-13 00:04:28.282 [INFO][6030] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" iface="eth0" netns="" Aug 13 00:04:28.363602 env[1913]: 2025-08-13 00:04:28.282 [INFO][6030] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Aug 13 00:04:28.363602 env[1913]: 2025-08-13 00:04:28.282 [INFO][6030] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Aug 13 00:04:28.363602 env[1913]: 2025-08-13 00:04:28.341 [INFO][6036] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" HandleID="k8s-pod-network.359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:04:28.363602 env[1913]: 2025-08-13 00:04:28.342 [INFO][6036] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:28.363602 env[1913]: 2025-08-13 00:04:28.342 [INFO][6036] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:28.363602 env[1913]: 2025-08-13 00:04:28.356 [WARNING][6036] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" HandleID="k8s-pod-network.359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:04:28.363602 env[1913]: 2025-08-13 00:04:28.356 [INFO][6036] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" HandleID="k8s-pod-network.359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:04:28.363602 env[1913]: 2025-08-13 00:04:28.358 [INFO][6036] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:28.363602 env[1913]: 2025-08-13 00:04:28.361 [INFO][6030] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Aug 13 00:04:28.364491 env[1913]: time="2025-08-13T00:04:28.363644286Z" level=info msg="TearDown network for sandbox \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\" successfully" Aug 13 00:04:28.364491 env[1913]: time="2025-08-13T00:04:28.363690631Z" level=info msg="StopPodSandbox for \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\" returns successfully" Aug 13 00:04:28.364610 env[1913]: time="2025-08-13T00:04:28.364483174Z" level=info msg="RemovePodSandbox for \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\"" Aug 13 00:04:28.364610 env[1913]: time="2025-08-13T00:04:28.364531858Z" level=info msg="Forcibly stopping sandbox \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\"" Aug 13 00:04:28.560634 env[1913]: 2025-08-13 00:04:28.479 [WARNING][6050] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0", GenerateName:"calico-apiserver-6ff96b959-", Namespace:"calico-apiserver", SelfLink:"", UID:"6e027b9e-faa1-4ba5-bf5e-ffbac2270a8e", ResourceVersion:"1227", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 1, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6ff96b959", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"c6f3c3aca4991ff81e510d0e5b474666d6114a13daa593a733d452cf8cbc74f9", Pod:"calico-apiserver-6ff96b959-jggcm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.1.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3abb69b4f11", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:04:28.560634 env[1913]: 2025-08-13 00:04:28.480 [INFO][6050] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Aug 13 00:04:28.560634 env[1913]: 2025-08-13 00:04:28.480 [INFO][6050] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" iface="eth0" netns="" Aug 13 00:04:28.560634 env[1913]: 2025-08-13 00:04:28.480 [INFO][6050] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Aug 13 00:04:28.560634 env[1913]: 2025-08-13 00:04:28.480 [INFO][6050] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Aug 13 00:04:28.560634 env[1913]: 2025-08-13 00:04:28.527 [INFO][6057] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" HandleID="k8s-pod-network.359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:04:28.560634 env[1913]: 2025-08-13 00:04:28.527 [INFO][6057] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:28.560634 env[1913]: 2025-08-13 00:04:28.528 [INFO][6057] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:28.560634 env[1913]: 2025-08-13 00:04:28.552 [WARNING][6057] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" HandleID="k8s-pod-network.359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:04:28.560634 env[1913]: 2025-08-13 00:04:28.552 [INFO][6057] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" HandleID="k8s-pod-network.359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Workload="ip--172--31--27--151-k8s-calico--apiserver--6ff96b959--jggcm-eth0" Aug 13 00:04:28.560634 env[1913]: 2025-08-13 00:04:28.554 [INFO][6057] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:28.560634 env[1913]: 2025-08-13 00:04:28.556 [INFO][6050] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014" Aug 13 00:04:28.561823 env[1913]: time="2025-08-13T00:04:28.561770278Z" level=info msg="TearDown network for sandbox \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\" successfully" Aug 13 00:04:28.569277 env[1913]: time="2025-08-13T00:04:28.569189654Z" level=info msg="RemovePodSandbox \"359c743935844633814177d6cc64312d79d14c463f789b4abadde48ce05c3014\" returns successfully" Aug 13 00:04:28.570245 env[1913]: time="2025-08-13T00:04:28.570187913Z" level=info msg="StopPodSandbox for \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\"" Aug 13 00:04:28.769060 env[1913]: 2025-08-13 00:04:28.658 [WARNING][6072] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" WorkloadEndpoint="ip--172--31--27--151-k8s-whisker--755dcdc665--9mnng-eth0" Aug 13 00:04:28.769060 env[1913]: 2025-08-13 00:04:28.659 [INFO][6072] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Aug 13 00:04:28.769060 env[1913]: 2025-08-13 00:04:28.659 [INFO][6072] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" iface="eth0" netns="" Aug 13 00:04:28.769060 env[1913]: 2025-08-13 00:04:28.659 [INFO][6072] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Aug 13 00:04:28.769060 env[1913]: 2025-08-13 00:04:28.659 [INFO][6072] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Aug 13 00:04:28.769060 env[1913]: 2025-08-13 00:04:28.743 [INFO][6079] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" HandleID="k8s-pod-network.12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Workload="ip--172--31--27--151-k8s-whisker--755dcdc665--9mnng-eth0" Aug 13 00:04:28.769060 env[1913]: 2025-08-13 00:04:28.743 [INFO][6079] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:28.769060 env[1913]: 2025-08-13 00:04:28.743 [INFO][6079] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:28.769060 env[1913]: 2025-08-13 00:04:28.757 [WARNING][6079] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" HandleID="k8s-pod-network.12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Workload="ip--172--31--27--151-k8s-whisker--755dcdc665--9mnng-eth0" Aug 13 00:04:28.769060 env[1913]: 2025-08-13 00:04:28.757 [INFO][6079] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" HandleID="k8s-pod-network.12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Workload="ip--172--31--27--151-k8s-whisker--755dcdc665--9mnng-eth0" Aug 13 00:04:28.769060 env[1913]: 2025-08-13 00:04:28.761 [INFO][6079] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:28.769060 env[1913]: 2025-08-13 00:04:28.766 [INFO][6072] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Aug 13 00:04:28.770531 env[1913]: time="2025-08-13T00:04:28.770473745Z" level=info msg="TearDown network for sandbox \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\" successfully" Aug 13 00:04:28.770681 env[1913]: time="2025-08-13T00:04:28.770645310Z" level=info msg="StopPodSandbox for \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\" returns successfully" Aug 13 00:04:28.773277 env[1913]: time="2025-08-13T00:04:28.771630262Z" level=info msg="RemovePodSandbox for \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\"" Aug 13 00:04:28.773277 env[1913]: time="2025-08-13T00:04:28.771715714Z" level=info msg="Forcibly stopping sandbox \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\"" Aug 13 00:04:29.071656 env[1913]: 2025-08-13 00:04:28.849 [WARNING][6094] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" WorkloadEndpoint="ip--172--31--27--151-k8s-whisker--755dcdc665--9mnng-eth0" Aug 13 00:04:29.071656 env[1913]: 2025-08-13 00:04:28.849 [INFO][6094] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Aug 13 00:04:29.071656 env[1913]: 2025-08-13 00:04:28.849 [INFO][6094] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" iface="eth0" netns="" Aug 13 00:04:29.071656 env[1913]: 2025-08-13 00:04:28.849 [INFO][6094] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Aug 13 00:04:29.071656 env[1913]: 2025-08-13 00:04:28.850 [INFO][6094] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Aug 13 00:04:29.071656 env[1913]: 2025-08-13 00:04:28.965 [INFO][6101] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" HandleID="k8s-pod-network.12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Workload="ip--172--31--27--151-k8s-whisker--755dcdc665--9mnng-eth0" Aug 13 00:04:29.071656 env[1913]: 2025-08-13 00:04:28.966 [INFO][6101] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:29.071656 env[1913]: 2025-08-13 00:04:28.966 [INFO][6101] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:29.071656 env[1913]: 2025-08-13 00:04:29.034 [WARNING][6101] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" HandleID="k8s-pod-network.12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Workload="ip--172--31--27--151-k8s-whisker--755dcdc665--9mnng-eth0" Aug 13 00:04:29.071656 env[1913]: 2025-08-13 00:04:29.034 [INFO][6101] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" HandleID="k8s-pod-network.12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Workload="ip--172--31--27--151-k8s-whisker--755dcdc665--9mnng-eth0" Aug 13 00:04:29.071656 env[1913]: 2025-08-13 00:04:29.042 [INFO][6101] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:29.071656 env[1913]: 2025-08-13 00:04:29.061 [INFO][6094] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309" Aug 13 00:04:29.072681 env[1913]: time="2025-08-13T00:04:29.071685986Z" level=info msg="TearDown network for sandbox \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\" successfully" Aug 13 00:04:29.080445 env[1913]: time="2025-08-13T00:04:29.078939245Z" level=info msg="RemovePodSandbox \"12ca975b0e0ff55fb1b6651c6c084a2b73acc293c56df49b961d1628eaa63309\" returns successfully" Aug 13 00:04:29.081053 env[1913]: time="2025-08-13T00:04:29.080983416Z" level=info msg="StopPodSandbox for \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\"" Aug 13 00:04:29.284691 env[1913]: 2025-08-13 00:04:29.173 [WARNING][6116] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0", GenerateName:"calico-kube-controllers-d75f5fd6d-", Namespace:"calico-system", SelfLink:"", UID:"ce695f2d-ac34-4846-848c-9adafe443946", ResourceVersion:"1126", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"d75f5fd6d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0", Pod:"calico-kube-controllers-d75f5fd6d-27n9j", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.1.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali03a672766d6", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:04:29.284691 env[1913]: 2025-08-13 00:04:29.174 [INFO][6116] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Aug 13 00:04:29.284691 env[1913]: 2025-08-13 00:04:29.174 [INFO][6116] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" iface="eth0" netns="" Aug 13 00:04:29.284691 env[1913]: 2025-08-13 00:04:29.174 [INFO][6116] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Aug 13 00:04:29.284691 env[1913]: 2025-08-13 00:04:29.174 [INFO][6116] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Aug 13 00:04:29.284691 env[1913]: 2025-08-13 00:04:29.240 [INFO][6123] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" HandleID="k8s-pod-network.3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Workload="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:04:29.284691 env[1913]: 2025-08-13 00:04:29.241 [INFO][6123] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:29.284691 env[1913]: 2025-08-13 00:04:29.241 [INFO][6123] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:29.284691 env[1913]: 2025-08-13 00:04:29.263 [WARNING][6123] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" HandleID="k8s-pod-network.3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Workload="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:04:29.284691 env[1913]: 2025-08-13 00:04:29.263 [INFO][6123] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" HandleID="k8s-pod-network.3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Workload="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:04:29.284691 env[1913]: 2025-08-13 00:04:29.266 [INFO][6123] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:29.284691 env[1913]: 2025-08-13 00:04:29.271 [INFO][6116] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Aug 13 00:04:29.285901 env[1913]: time="2025-08-13T00:04:29.285848484Z" level=info msg="TearDown network for sandbox \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\" successfully" Aug 13 00:04:29.286038 env[1913]: time="2025-08-13T00:04:29.286005013Z" level=info msg="StopPodSandbox for \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\" returns successfully" Aug 13 00:04:29.286918 env[1913]: time="2025-08-13T00:04:29.286869340Z" level=info msg="RemovePodSandbox for \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\"" Aug 13 00:04:29.287272 env[1913]: time="2025-08-13T00:04:29.287086049Z" level=info msg="Forcibly stopping sandbox \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\"" Aug 13 00:04:29.462464 env[1913]: 2025-08-13 00:04:29.390 [WARNING][6138] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0", GenerateName:"calico-kube-controllers-d75f5fd6d-", Namespace:"calico-system", SelfLink:"", UID:"ce695f2d-ac34-4846-848c-9adafe443946", ResourceVersion:"1126", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"d75f5fd6d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"00440da2100543b4e78c13b0166757dde4bdf656f4de78d8f993f48ee7d4e1c0", Pod:"calico-kube-controllers-d75f5fd6d-27n9j", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.1.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali03a672766d6", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:04:29.462464 env[1913]: 2025-08-13 00:04:29.390 [INFO][6138] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Aug 13 00:04:29.462464 env[1913]: 2025-08-13 00:04:29.390 [INFO][6138] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" iface="eth0" netns="" Aug 13 00:04:29.462464 env[1913]: 2025-08-13 00:04:29.390 [INFO][6138] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Aug 13 00:04:29.462464 env[1913]: 2025-08-13 00:04:29.390 [INFO][6138] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Aug 13 00:04:29.462464 env[1913]: 2025-08-13 00:04:29.438 [INFO][6146] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" HandleID="k8s-pod-network.3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Workload="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:04:29.462464 env[1913]: 2025-08-13 00:04:29.438 [INFO][6146] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:29.462464 env[1913]: 2025-08-13 00:04:29.438 [INFO][6146] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:29.462464 env[1913]: 2025-08-13 00:04:29.454 [WARNING][6146] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" HandleID="k8s-pod-network.3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Workload="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:04:29.462464 env[1913]: 2025-08-13 00:04:29.454 [INFO][6146] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" HandleID="k8s-pod-network.3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Workload="ip--172--31--27--151-k8s-calico--kube--controllers--d75f5fd6d--27n9j-eth0" Aug 13 00:04:29.462464 env[1913]: 2025-08-13 00:04:29.457 [INFO][6146] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:29.462464 env[1913]: 2025-08-13 00:04:29.459 [INFO][6138] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403" Aug 13 00:04:29.463350 env[1913]: time="2025-08-13T00:04:29.462504856Z" level=info msg="TearDown network for sandbox \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\" successfully" Aug 13 00:04:29.469538 env[1913]: time="2025-08-13T00:04:29.469434846Z" level=info msg="RemovePodSandbox \"3d50bb1173156c66eb20411ebc3be46a606dbe9bd7ad833564cbfbf572578403\" returns successfully" Aug 13 00:04:29.470255 env[1913]: time="2025-08-13T00:04:29.470192661Z" level=info msg="StopPodSandbox for \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\"" Aug 13 00:04:29.571365 kernel: kauditd_printk_skb: 33 callbacks suppressed Aug 13 00:04:29.571500 kernel: audit: type=1130 audit(1755043469.557:550): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-172.31.27.151:22-139.178.89.65:38090 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:29.557000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-172.31.27.151:22-139.178.89.65:38090 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:29.557936 systemd[1]: Started sshd@20-172.31.27.151:22-139.178.89.65:38090.service. Aug 13 00:04:29.756828 env[1913]: 2025-08-13 00:04:29.672 [WARNING][6160] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"ba5590d9-af5c-484d-9d4d-2e76ac5327b7", ResourceVersion:"1084", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 2, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf", Pod:"coredns-7c65d6cfc9-v6d44", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.1.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"califc96d725571", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:04:29.756828 env[1913]: 2025-08-13 00:04:29.673 [INFO][6160] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Aug 13 00:04:29.756828 env[1913]: 2025-08-13 00:04:29.673 [INFO][6160] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" iface="eth0" netns="" Aug 13 00:04:29.756828 env[1913]: 2025-08-13 00:04:29.673 [INFO][6160] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Aug 13 00:04:29.756828 env[1913]: 2025-08-13 00:04:29.673 [INFO][6160] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Aug 13 00:04:29.756828 env[1913]: 2025-08-13 00:04:29.728 [INFO][6169] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" HandleID="k8s-pod-network.6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:04:29.756828 env[1913]: 2025-08-13 00:04:29.728 [INFO][6169] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:29.756828 env[1913]: 2025-08-13 00:04:29.728 [INFO][6169] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:29.756828 env[1913]: 2025-08-13 00:04:29.743 [WARNING][6169] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" HandleID="k8s-pod-network.6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:04:29.756828 env[1913]: 2025-08-13 00:04:29.743 [INFO][6169] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" HandleID="k8s-pod-network.6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:04:29.756828 env[1913]: 2025-08-13 00:04:29.748 [INFO][6169] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:29.756828 env[1913]: 2025-08-13 00:04:29.751 [INFO][6160] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Aug 13 00:04:29.756828 env[1913]: time="2025-08-13T00:04:29.755397881Z" level=info msg="TearDown network for sandbox \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\" successfully" Aug 13 00:04:29.756828 env[1913]: time="2025-08-13T00:04:29.755443277Z" level=info msg="StopPodSandbox for \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\" returns successfully" Aug 13 00:04:29.756828 env[1913]: time="2025-08-13T00:04:29.756158600Z" level=info msg="RemovePodSandbox for \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\"" Aug 13 00:04:29.756828 env[1913]: time="2025-08-13T00:04:29.756213788Z" level=info msg="Forcibly stopping sandbox \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\"" Aug 13 00:04:29.786728 sshd[6164]: Accepted publickey for core from 139.178.89.65 port 38090 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:04:29.785000 audit[6164]: USER_ACCT pid=6164 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:29.799394 sshd[6164]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:04:29.820601 kernel: audit: type=1101 audit(1755043469.785:551): pid=6164 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:29.820792 kernel: audit: type=1103 audit(1755043469.797:552): pid=6164 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:29.797000 audit[6164]: CRED_ACQ pid=6164 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:29.797000 audit[6164]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffdef16040 a2=3 a3=1 items=0 ppid=1 pid=6164 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=21 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:29.848794 kernel: audit: type=1006 audit(1755043469.797:553): pid=6164 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=21 res=1 Aug 13 00:04:29.848864 kernel: audit: type=1300 audit(1755043469.797:553): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffdef16040 a2=3 a3=1 items=0 ppid=1 pid=6164 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=21 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:29.840509 systemd[1]: Started session-21.scope. Aug 13 00:04:29.850362 systemd-logind[1901]: New session 21 of user core. Aug 13 00:04:29.797000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:29.872865 kernel: audit: type=1327 audit(1755043469.797:553): proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:29.888000 audit[6164]: USER_START pid=6164 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:29.902000 audit[6192]: CRED_ACQ pid=6192 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:29.917850 kernel: audit: type=1105 audit(1755043469.888:554): pid=6164 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:29.918020 kernel: audit: type=1103 audit(1755043469.902:555): pid=6192 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:29.992156 env[1913]: 2025-08-13 00:04:29.904 [WARNING][6185] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"ba5590d9-af5c-484d-9d4d-2e76ac5327b7", ResourceVersion:"1084", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 2, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"086b91db53901203d1b1b43ff7f6950633d805c16c7ce91fcc8338276c6fefbf", Pod:"coredns-7c65d6cfc9-v6d44", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.1.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"califc96d725571", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:04:29.992156 env[1913]: 2025-08-13 00:04:29.904 [INFO][6185] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Aug 13 00:04:29.992156 env[1913]: 2025-08-13 00:04:29.904 [INFO][6185] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" iface="eth0" netns="" Aug 13 00:04:29.992156 env[1913]: 2025-08-13 00:04:29.904 [INFO][6185] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Aug 13 00:04:29.992156 env[1913]: 2025-08-13 00:04:29.904 [INFO][6185] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Aug 13 00:04:29.992156 env[1913]: 2025-08-13 00:04:29.958 [INFO][6194] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" HandleID="k8s-pod-network.6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:04:29.992156 env[1913]: 2025-08-13 00:04:29.959 [INFO][6194] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:29.992156 env[1913]: 2025-08-13 00:04:29.960 [INFO][6194] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:29.992156 env[1913]: 2025-08-13 00:04:29.977 [WARNING][6194] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" HandleID="k8s-pod-network.6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:04:29.992156 env[1913]: 2025-08-13 00:04:29.978 [INFO][6194] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" HandleID="k8s-pod-network.6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--v6d44-eth0" Aug 13 00:04:29.992156 env[1913]: 2025-08-13 00:04:29.980 [INFO][6194] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:29.992156 env[1913]: 2025-08-13 00:04:29.983 [INFO][6185] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a" Aug 13 00:04:29.992156 env[1913]: time="2025-08-13T00:04:29.990294776Z" level=info msg="TearDown network for sandbox \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\" successfully" Aug 13 00:04:30.015319 env[1913]: time="2025-08-13T00:04:30.015078059Z" level=info msg="RemovePodSandbox \"6a7b0c5aeb813116576eb442048c96265d800389d5ae640b883d6ca13b7e526a\" returns successfully" Aug 13 00:04:30.019132 env[1913]: time="2025-08-13T00:04:30.018130690Z" level=info msg="StopPodSandbox for \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\"" Aug 13 00:04:30.259300 env[1913]: 2025-08-13 00:04:30.141 [WARNING][6215] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"48c5f0fe-d3a5-4b24-a4f6-448275db6b4d", ResourceVersion:"1076", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 2, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b", Pod:"coredns-7c65d6cfc9-kqx4k", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.1.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali28fc58a798b", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:04:30.259300 env[1913]: 2025-08-13 00:04:30.142 [INFO][6215] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Aug 13 00:04:30.259300 env[1913]: 2025-08-13 00:04:30.142 [INFO][6215] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" iface="eth0" netns="" Aug 13 00:04:30.259300 env[1913]: 2025-08-13 00:04:30.142 [INFO][6215] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Aug 13 00:04:30.259300 env[1913]: 2025-08-13 00:04:30.142 [INFO][6215] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Aug 13 00:04:30.259300 env[1913]: 2025-08-13 00:04:30.217 [INFO][6222] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" HandleID="k8s-pod-network.a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:04:30.259300 env[1913]: 2025-08-13 00:04:30.218 [INFO][6222] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:30.259300 env[1913]: 2025-08-13 00:04:30.218 [INFO][6222] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:30.259300 env[1913]: 2025-08-13 00:04:30.245 [WARNING][6222] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" HandleID="k8s-pod-network.a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:04:30.259300 env[1913]: 2025-08-13 00:04:30.245 [INFO][6222] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" HandleID="k8s-pod-network.a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:04:30.259300 env[1913]: 2025-08-13 00:04:30.253 [INFO][6222] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:30.259300 env[1913]: 2025-08-13 00:04:30.255 [INFO][6215] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Aug 13 00:04:30.260229 env[1913]: time="2025-08-13T00:04:30.259343622Z" level=info msg="TearDown network for sandbox \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\" successfully" Aug 13 00:04:30.260229 env[1913]: time="2025-08-13T00:04:30.259390759Z" level=info msg="StopPodSandbox for \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\" returns successfully" Aug 13 00:04:30.261647 env[1913]: time="2025-08-13T00:04:30.261579603Z" level=info msg="RemovePodSandbox for \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\"" Aug 13 00:04:30.261802 env[1913]: time="2025-08-13T00:04:30.261652587Z" level=info msg="Forcibly stopping sandbox \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\"" Aug 13 00:04:30.289855 sshd[6164]: pam_unix(sshd:session): session closed for user core Aug 13 00:04:30.292000 audit[6164]: USER_END pid=6164 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:30.293000 audit[6164]: CRED_DISP pid=6164 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:30.306348 systemd[1]: sshd@20-172.31.27.151:22-139.178.89.65:38090.service: Deactivated successfully. Aug 13 00:04:30.307822 systemd[1]: session-21.scope: Deactivated successfully. Aug 13 00:04:30.321562 kernel: audit: type=1106 audit(1755043470.292:556): pid=6164 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:30.321719 kernel: audit: type=1104 audit(1755043470.293:557): pid=6164 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:30.305000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-172.31.27.151:22-139.178.89.65:38090 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:30.329132 systemd-logind[1901]: Session 21 logged out. Waiting for processes to exit. Aug 13 00:04:30.331911 systemd-logind[1901]: Removed session 21. Aug 13 00:04:30.470013 env[1913]: 2025-08-13 00:04:30.383 [WARNING][6236] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"48c5f0fe-d3a5-4b24-a4f6-448275db6b4d", ResourceVersion:"1076", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 2, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"dc95269cc9bafb9f18e5fcbed52374714f90e09916276e12823b635dd08b0b1b", Pod:"coredns-7c65d6cfc9-kqx4k", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.1.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali28fc58a798b", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:04:30.470013 env[1913]: 2025-08-13 00:04:30.383 [INFO][6236] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Aug 13 00:04:30.470013 env[1913]: 2025-08-13 00:04:30.384 [INFO][6236] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" iface="eth0" netns="" Aug 13 00:04:30.470013 env[1913]: 2025-08-13 00:04:30.384 [INFO][6236] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Aug 13 00:04:30.470013 env[1913]: 2025-08-13 00:04:30.384 [INFO][6236] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Aug 13 00:04:30.470013 env[1913]: 2025-08-13 00:04:30.448 [INFO][6245] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" HandleID="k8s-pod-network.a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:04:30.470013 env[1913]: 2025-08-13 00:04:30.448 [INFO][6245] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:30.470013 env[1913]: 2025-08-13 00:04:30.448 [INFO][6245] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:30.470013 env[1913]: 2025-08-13 00:04:30.460 [WARNING][6245] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" HandleID="k8s-pod-network.a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:04:30.470013 env[1913]: 2025-08-13 00:04:30.460 [INFO][6245] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" HandleID="k8s-pod-network.a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Workload="ip--172--31--27--151-k8s-coredns--7c65d6cfc9--kqx4k-eth0" Aug 13 00:04:30.470013 env[1913]: 2025-08-13 00:04:30.463 [INFO][6245] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:30.470013 env[1913]: 2025-08-13 00:04:30.465 [INFO][6236] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf" Aug 13 00:04:30.471143 env[1913]: time="2025-08-13T00:04:30.471043719Z" level=info msg="TearDown network for sandbox \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\" successfully" Aug 13 00:04:30.478661 env[1913]: time="2025-08-13T00:04:30.478586742Z" level=info msg="RemovePodSandbox \"a83ea8348679e6fb09dcb5c75f1463a95870c046efc22e01d3bb0e4fa6939cdf\" returns successfully" Aug 13 00:04:30.479327 env[1913]: time="2025-08-13T00:04:30.479277405Z" level=info msg="StopPodSandbox for \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\"" Aug 13 00:04:30.652067 env[1913]: 2025-08-13 00:04:30.572 [WARNING][6261] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"cf609856-35b0-4517-9fba-45b28ef2be55", ResourceVersion:"1353", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb", Pod:"goldmane-58fd7646b9-6tzfz", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.1.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali4658d01368a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:04:30.652067 env[1913]: 2025-08-13 00:04:30.572 [INFO][6261] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Aug 13 00:04:30.652067 env[1913]: 2025-08-13 00:04:30.572 [INFO][6261] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" iface="eth0" netns="" Aug 13 00:04:30.652067 env[1913]: 2025-08-13 00:04:30.572 [INFO][6261] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Aug 13 00:04:30.652067 env[1913]: 2025-08-13 00:04:30.573 [INFO][6261] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Aug 13 00:04:30.652067 env[1913]: 2025-08-13 00:04:30.627 [INFO][6268] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" HandleID="k8s-pod-network.b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Workload="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:04:30.652067 env[1913]: 2025-08-13 00:04:30.628 [INFO][6268] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:30.652067 env[1913]: 2025-08-13 00:04:30.628 [INFO][6268] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:30.652067 env[1913]: 2025-08-13 00:04:30.642 [WARNING][6268] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" HandleID="k8s-pod-network.b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Workload="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:04:30.652067 env[1913]: 2025-08-13 00:04:30.642 [INFO][6268] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" HandleID="k8s-pod-network.b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Workload="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:04:30.652067 env[1913]: 2025-08-13 00:04:30.645 [INFO][6268] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:30.652067 env[1913]: 2025-08-13 00:04:30.648 [INFO][6261] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Aug 13 00:04:30.652067 env[1913]: time="2025-08-13T00:04:30.650902238Z" level=info msg="TearDown network for sandbox \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\" successfully" Aug 13 00:04:30.652067 env[1913]: time="2025-08-13T00:04:30.650948666Z" level=info msg="StopPodSandbox for \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\" returns successfully" Aug 13 00:04:30.652067 env[1913]: time="2025-08-13T00:04:30.651688193Z" level=info msg="RemovePodSandbox for \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\"" Aug 13 00:04:30.652067 env[1913]: time="2025-08-13T00:04:30.651735605Z" level=info msg="Forcibly stopping sandbox \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\"" Aug 13 00:04:30.834734 env[1913]: 2025-08-13 00:04:30.749 [WARNING][6282] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"cf609856-35b0-4517-9fba-45b28ef2be55", ResourceVersion:"1353", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"bd235c5b763088c9007e45f2ba465a9e90ebf88dd26ca2319c7b312a550970bb", Pod:"goldmane-58fd7646b9-6tzfz", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.1.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali4658d01368a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:04:30.834734 env[1913]: 2025-08-13 00:04:30.750 [INFO][6282] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Aug 13 00:04:30.834734 env[1913]: 2025-08-13 00:04:30.750 [INFO][6282] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" iface="eth0" netns="" Aug 13 00:04:30.834734 env[1913]: 2025-08-13 00:04:30.750 [INFO][6282] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Aug 13 00:04:30.834734 env[1913]: 2025-08-13 00:04:30.750 [INFO][6282] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Aug 13 00:04:30.834734 env[1913]: 2025-08-13 00:04:30.804 [INFO][6289] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" HandleID="k8s-pod-network.b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Workload="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:04:30.834734 env[1913]: 2025-08-13 00:04:30.807 [INFO][6289] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:30.834734 env[1913]: 2025-08-13 00:04:30.808 [INFO][6289] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:30.834734 env[1913]: 2025-08-13 00:04:30.826 [WARNING][6289] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" HandleID="k8s-pod-network.b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Workload="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:04:30.834734 env[1913]: 2025-08-13 00:04:30.826 [INFO][6289] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" HandleID="k8s-pod-network.b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Workload="ip--172--31--27--151-k8s-goldmane--58fd7646b9--6tzfz-eth0" Aug 13 00:04:30.834734 env[1913]: 2025-08-13 00:04:30.829 [INFO][6289] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:30.834734 env[1913]: 2025-08-13 00:04:30.832 [INFO][6282] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802" Aug 13 00:04:30.835701 env[1913]: time="2025-08-13T00:04:30.834783700Z" level=info msg="TearDown network for sandbox \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\" successfully" Aug 13 00:04:30.841559 env[1913]: time="2025-08-13T00:04:30.841488245Z" level=info msg="RemovePodSandbox \"b8fea0adc2701f3ef944b43bf9bed14193822b84f65132c04d59aa28fdd81802\" returns successfully" Aug 13 00:04:30.842413 env[1913]: time="2025-08-13T00:04:30.842325584Z" level=info msg="StopPodSandbox for \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\"" Aug 13 00:04:31.038862 env[1913]: 2025-08-13 00:04:30.959 [WARNING][6305] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"21d84c26-2961-4407-a494-a46bbc34cc12", ResourceVersion:"1224", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a", Pod:"csi-node-driver-n4g7q", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.1.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calid0dd8b3f31c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:04:31.038862 env[1913]: 2025-08-13 00:04:30.959 [INFO][6305] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Aug 13 00:04:31.038862 env[1913]: 2025-08-13 00:04:30.959 [INFO][6305] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" iface="eth0" netns="" Aug 13 00:04:31.038862 env[1913]: 2025-08-13 00:04:30.959 [INFO][6305] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Aug 13 00:04:31.038862 env[1913]: 2025-08-13 00:04:30.959 [INFO][6305] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Aug 13 00:04:31.038862 env[1913]: 2025-08-13 00:04:31.016 [INFO][6314] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" HandleID="k8s-pod-network.4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Workload="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:04:31.038862 env[1913]: 2025-08-13 00:04:31.017 [INFO][6314] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:31.038862 env[1913]: 2025-08-13 00:04:31.017 [INFO][6314] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:31.038862 env[1913]: 2025-08-13 00:04:31.030 [WARNING][6314] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" HandleID="k8s-pod-network.4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Workload="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:04:31.038862 env[1913]: 2025-08-13 00:04:31.030 [INFO][6314] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" HandleID="k8s-pod-network.4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Workload="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:04:31.038862 env[1913]: 2025-08-13 00:04:31.032 [INFO][6314] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:31.038862 env[1913]: 2025-08-13 00:04:31.035 [INFO][6305] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Aug 13 00:04:31.040292 env[1913]: time="2025-08-13T00:04:31.040239165Z" level=info msg="TearDown network for sandbox \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\" successfully" Aug 13 00:04:31.040442 env[1913]: time="2025-08-13T00:04:31.040409841Z" level=info msg="StopPodSandbox for \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\" returns successfully" Aug 13 00:04:31.041990 env[1913]: time="2025-08-13T00:04:31.041941803Z" level=info msg="RemovePodSandbox for \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\"" Aug 13 00:04:31.042288 env[1913]: time="2025-08-13T00:04:31.042212620Z" level=info msg="Forcibly stopping sandbox \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\"" Aug 13 00:04:31.219261 env[1913]: 2025-08-13 00:04:31.147 [WARNING][6329] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"21d84c26-2961-4407-a494-a46bbc34cc12", ResourceVersion:"1224", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 3, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-27-151", ContainerID:"e777b7d37ac1d6ebd3d9fccd53c106cc25bdcca1bdad105a82ebeee0363eff9a", Pod:"csi-node-driver-n4g7q", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.1.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calid0dd8b3f31c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:04:31.219261 env[1913]: 2025-08-13 00:04:31.148 [INFO][6329] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Aug 13 00:04:31.219261 env[1913]: 2025-08-13 00:04:31.148 [INFO][6329] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" iface="eth0" netns="" Aug 13 00:04:31.219261 env[1913]: 2025-08-13 00:04:31.148 [INFO][6329] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Aug 13 00:04:31.219261 env[1913]: 2025-08-13 00:04:31.148 [INFO][6329] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Aug 13 00:04:31.219261 env[1913]: 2025-08-13 00:04:31.199 [INFO][6336] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" HandleID="k8s-pod-network.4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Workload="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:04:31.219261 env[1913]: 2025-08-13 00:04:31.199 [INFO][6336] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:04:31.219261 env[1913]: 2025-08-13 00:04:31.199 [INFO][6336] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:04:31.219261 env[1913]: 2025-08-13 00:04:31.211 [WARNING][6336] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" HandleID="k8s-pod-network.4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Workload="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:04:31.219261 env[1913]: 2025-08-13 00:04:31.211 [INFO][6336] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" HandleID="k8s-pod-network.4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Workload="ip--172--31--27--151-k8s-csi--node--driver--n4g7q-eth0" Aug 13 00:04:31.219261 env[1913]: 2025-08-13 00:04:31.214 [INFO][6336] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:04:31.219261 env[1913]: 2025-08-13 00:04:31.216 [INFO][6329] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678" Aug 13 00:04:31.220231 env[1913]: time="2025-08-13T00:04:31.219306049Z" level=info msg="TearDown network for sandbox \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\" successfully" Aug 13 00:04:31.226506 env[1913]: time="2025-08-13T00:04:31.226419507Z" level=info msg="RemovePodSandbox \"4164a7e1fa04d976de0ea87c20daaf92177cf2cd7f76cd15a3fb340ed67f6678\" returns successfully" Aug 13 00:04:31.712000 audit[6343]: NETFILTER_CFG table=filter:136 family=2 entries=20 op=nft_register_rule pid=6343 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:04:31.712000 audit[6343]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=3016 a0=3 a1=ffffd7c690c0 a2=0 a3=1 items=0 ppid=3113 pid=6343 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:31.712000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:04:31.718000 audit[6343]: NETFILTER_CFG table=nat:137 family=2 entries=110 op=nft_register_chain pid=6343 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" Aug 13 00:04:31.718000 audit[6343]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=50988 a0=3 a1=ffffd7c690c0 a2=0 a3=1 items=0 ppid=3113 pid=6343 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:31.718000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 Aug 13 00:04:35.330874 kernel: kauditd_printk_skb: 7 callbacks suppressed Aug 13 00:04:35.331027 kernel: audit: type=1130 audit(1755043475.318:561): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-172.31.27.151:22-139.178.89.65:38100 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:35.318000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-172.31.27.151:22-139.178.89.65:38100 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:35.319273 systemd[1]: Started sshd@21-172.31.27.151:22-139.178.89.65:38100.service. Aug 13 00:04:35.520000 audit[6347]: USER_ACCT pid=6347 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:35.531997 sshd[6347]: Accepted publickey for core from 139.178.89.65 port 38100 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:04:35.535357 sshd[6347]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:04:35.533000 audit[6347]: CRED_ACQ pid=6347 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:35.547998 kernel: audit: type=1101 audit(1755043475.520:562): pid=6347 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:35.548152 kernel: audit: type=1103 audit(1755043475.533:563): pid=6347 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:35.556929 kernel: audit: type=1006 audit(1755043475.533:564): pid=6347 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=22 res=1 Aug 13 00:04:35.533000 audit[6347]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffd9fa5e50 a2=3 a3=1 items=0 ppid=1 pid=6347 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=22 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:35.570723 kernel: audit: type=1300 audit(1755043475.533:564): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffd9fa5e50 a2=3 a3=1 items=0 ppid=1 pid=6347 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=22 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:35.572345 systemd-logind[1901]: New session 22 of user core. Aug 13 00:04:35.573005 systemd[1]: Started session-22.scope. Aug 13 00:04:35.533000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:35.577507 kernel: audit: type=1327 audit(1755043475.533:564): proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:35.594000 audit[6347]: USER_START pid=6347 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:35.611000 audit[6350]: CRED_ACQ pid=6350 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:35.622612 kernel: audit: type=1105 audit(1755043475.594:565): pid=6347 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:35.622747 kernel: audit: type=1103 audit(1755043475.611:566): pid=6350 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:35.898448 sshd[6347]: pam_unix(sshd:session): session closed for user core Aug 13 00:04:35.899000 audit[6347]: USER_END pid=6347 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:35.916047 systemd-logind[1901]: Session 22 logged out. Waiting for processes to exit. Aug 13 00:04:35.919291 systemd[1]: sshd@21-172.31.27.151:22-139.178.89.65:38100.service: Deactivated successfully. Aug 13 00:04:35.921279 systemd[1]: session-22.scope: Deactivated successfully. Aug 13 00:04:35.926165 systemd-logind[1901]: Removed session 22. Aug 13 00:04:35.911000 audit[6347]: CRED_DISP pid=6347 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:35.943592 kernel: audit: type=1106 audit(1755043475.899:567): pid=6347 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:35.943710 kernel: audit: type=1104 audit(1755043475.911:568): pid=6347 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:35.918000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-172.31.27.151:22-139.178.89.65:38100 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:40.766892 systemd[1]: run-containerd-runc-k8s.io-be0ed1e29b588d67e52fae942c6b3420ba4d8c77c6e0d8347717d2a5797a2a37-runc.Kx1joh.mount: Deactivated successfully. Aug 13 00:04:40.924717 systemd[1]: Started sshd@22-172.31.27.151:22-139.178.89.65:57956.service. Aug 13 00:04:40.928000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-172.31.27.151:22-139.178.89.65:57956 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:40.931392 kernel: kauditd_printk_skb: 1 callbacks suppressed Aug 13 00:04:40.931489 kernel: audit: type=1130 audit(1755043480.928:570): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-172.31.27.151:22-139.178.89.65:57956 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:41.132000 audit[6381]: USER_ACCT pid=6381 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:41.134231 sshd[6381]: Accepted publickey for core from 139.178.89.65 port 57956 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:04:41.146189 sshd[6381]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:04:41.144000 audit[6381]: CRED_ACQ pid=6381 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:41.164061 kernel: audit: type=1101 audit(1755043481.132:571): pid=6381 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:41.164249 kernel: audit: type=1103 audit(1755043481.144:572): pid=6381 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:41.166888 systemd[1]: Started session-23.scope. Aug 13 00:04:41.168332 systemd-logind[1901]: New session 23 of user core. Aug 13 00:04:41.196678 kernel: audit: type=1006 audit(1755043481.144:573): pid=6381 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=23 res=1 Aug 13 00:04:41.196865 kernel: audit: type=1300 audit(1755043481.144:573): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffec1ffa20 a2=3 a3=1 items=0 ppid=1 pid=6381 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=23 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:41.144000 audit[6381]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffec1ffa20 a2=3 a3=1 items=0 ppid=1 pid=6381 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=23 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:41.144000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:41.211492 kernel: audit: type=1327 audit(1755043481.144:573): proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:41.211607 kernel: audit: type=1105 audit(1755043481.184:574): pid=6381 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:41.184000 audit[6381]: USER_START pid=6381 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:41.187000 audit[6384]: CRED_ACQ pid=6384 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:41.239778 kernel: audit: type=1103 audit(1755043481.187:575): pid=6384 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:41.527386 sshd[6381]: pam_unix(sshd:session): session closed for user core Aug 13 00:04:41.528000 audit[6381]: USER_END pid=6381 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:41.542070 systemd[1]: sshd@22-172.31.27.151:22-139.178.89.65:57956.service: Deactivated successfully. Aug 13 00:04:41.544323 systemd[1]: session-23.scope: Deactivated successfully. Aug 13 00:04:41.544734 systemd-logind[1901]: Session 23 logged out. Waiting for processes to exit. Aug 13 00:04:41.546800 systemd-logind[1901]: Removed session 23. Aug 13 00:04:41.534000 audit[6381]: CRED_DISP pid=6381 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:41.557952 kernel: audit: type=1106 audit(1755043481.528:576): pid=6381 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:41.558056 kernel: audit: type=1104 audit(1755043481.534:577): pid=6381 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:41.541000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-172.31.27.151:22-139.178.89.65:57956 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:46.552841 systemd[1]: Started sshd@23-172.31.27.151:22-139.178.89.65:57960.service. Aug 13 00:04:46.552000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-172.31.27.151:22-139.178.89.65:57960 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:46.555367 kernel: kauditd_printk_skb: 1 callbacks suppressed Aug 13 00:04:46.555485 kernel: audit: type=1130 audit(1755043486.552:579): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-172.31.27.151:22-139.178.89.65:57960 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:46.745585 sshd[6394]: Accepted publickey for core from 139.178.89.65 port 57960 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:04:46.744000 audit[6394]: USER_ACCT pid=6394 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:46.759161 kernel: audit: type=1101 audit(1755043486.744:580): pid=6394 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:46.759293 kernel: audit: type=1103 audit(1755043486.756:581): pid=6394 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:46.756000 audit[6394]: CRED_ACQ pid=6394 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:46.758931 sshd[6394]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:04:46.775508 kernel: audit: type=1006 audit(1755043486.757:582): pid=6394 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=24 res=1 Aug 13 00:04:46.757000 audit[6394]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffe24a0610 a2=3 a3=1 items=0 ppid=1 pid=6394 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=24 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:46.786994 kernel: audit: type=1300 audit(1755043486.757:582): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffe24a0610 a2=3 a3=1 items=0 ppid=1 pid=6394 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=24 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:46.757000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:46.792870 kernel: audit: type=1327 audit(1755043486.757:582): proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:46.796189 systemd-logind[1901]: New session 24 of user core. Aug 13 00:04:46.798586 systemd[1]: Started session-24.scope. Aug 13 00:04:46.814000 audit[6394]: USER_START pid=6394 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:46.829000 audit[6397]: CRED_ACQ pid=6397 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:46.841119 kernel: audit: type=1105 audit(1755043486.814:583): pid=6394 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:46.841261 kernel: audit: type=1103 audit(1755043486.829:584): pid=6397 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:47.240880 sshd[6394]: pam_unix(sshd:session): session closed for user core Aug 13 00:04:47.242000 audit[6394]: USER_END pid=6394 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:47.242000 audit[6394]: CRED_DISP pid=6394 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:47.256169 systemd[1]: sshd@23-172.31.27.151:22-139.178.89.65:57960.service: Deactivated successfully. Aug 13 00:04:47.257552 systemd[1]: session-24.scope: Deactivated successfully. Aug 13 00:04:47.267764 kernel: audit: type=1106 audit(1755043487.242:585): pid=6394 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:47.267921 kernel: audit: type=1104 audit(1755043487.242:586): pid=6394 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:47.255000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-172.31.27.151:22-139.178.89.65:57960 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:47.268540 systemd-logind[1901]: Session 24 logged out. Waiting for processes to exit. Aug 13 00:04:47.270290 systemd-logind[1901]: Removed session 24. Aug 13 00:04:51.605347 systemd[1]: run-containerd-runc-k8s.io-7ee0d4f88f3c0fde73248650354bfcaccd45af41c5d859d00eea0aeb5c899949-runc.etOcWR.mount: Deactivated successfully. Aug 13 00:04:52.265000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@24-172.31.27.151:22-139.178.89.65:42460 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:52.266070 systemd[1]: Started sshd@24-172.31.27.151:22-139.178.89.65:42460.service. Aug 13 00:04:52.268637 kernel: kauditd_printk_skb: 1 callbacks suppressed Aug 13 00:04:52.268724 kernel: audit: type=1130 audit(1755043492.265:588): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@24-172.31.27.151:22-139.178.89.65:42460 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:52.470000 audit[6445]: USER_ACCT pid=6445 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:52.471539 sshd[6445]: Accepted publickey for core from 139.178.89.65 port 42460 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:04:52.482000 audit[6445]: CRED_ACQ pid=6445 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:52.484982 sshd[6445]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:04:52.493962 kernel: audit: type=1101 audit(1755043492.470:589): pid=6445 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:52.494174 kernel: audit: type=1103 audit(1755043492.482:590): pid=6445 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:52.504687 systemd[1]: Started session-25.scope. Aug 13 00:04:52.507896 systemd-logind[1901]: New session 25 of user core. Aug 13 00:04:52.515451 kernel: audit: type=1006 audit(1755043492.483:591): pid=6445 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=25 res=1 Aug 13 00:04:52.483000 audit[6445]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffe5a0bfb0 a2=3 a3=1 items=0 ppid=1 pid=6445 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=25 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:52.529599 kernel: audit: type=1300 audit(1755043492.483:591): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffe5a0bfb0 a2=3 a3=1 items=0 ppid=1 pid=6445 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=25 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:52.483000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:52.536721 kernel: audit: type=1327 audit(1755043492.483:591): proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:52.526000 audit[6445]: USER_START pid=6445 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:52.549918 kernel: audit: type=1105 audit(1755043492.526:592): pid=6445 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:52.536000 audit[6448]: CRED_ACQ pid=6448 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:52.560764 kernel: audit: type=1103 audit(1755043492.536:593): pid=6448 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:52.861200 sshd[6445]: pam_unix(sshd:session): session closed for user core Aug 13 00:04:52.862000 audit[6445]: USER_END pid=6445 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:52.879806 systemd-logind[1901]: Session 25 logged out. Waiting for processes to exit. Aug 13 00:04:52.882672 systemd[1]: sshd@24-172.31.27.151:22-139.178.89.65:42460.service: Deactivated successfully. Aug 13 00:04:52.884236 systemd[1]: session-25.scope: Deactivated successfully. Aug 13 00:04:52.887147 kernel: audit: type=1106 audit(1755043492.862:594): pid=6445 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:52.887932 systemd-logind[1901]: Removed session 25. Aug 13 00:04:52.875000 audit[6445]: CRED_DISP pid=6445 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:52.882000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@24-172.31.27.151:22-139.178.89.65:42460 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:52.910191 kernel: audit: type=1104 audit(1755043492.875:595): pid=6445 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:57.889495 systemd[1]: Started sshd@25-172.31.27.151:22-139.178.89.65:42462.service. Aug 13 00:04:57.889000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-172.31.27.151:22-139.178.89.65:42462 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:57.897209 kernel: kauditd_printk_skb: 1 callbacks suppressed Aug 13 00:04:57.897350 kernel: audit: type=1130 audit(1755043497.889:597): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-172.31.27.151:22-139.178.89.65:42462 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:58.071493 sshd[6464]: Accepted publickey for core from 139.178.89.65 port 42462 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:04:58.070000 audit[6464]: USER_ACCT pid=6464 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:58.081631 sshd[6464]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:04:58.098835 systemd[1]: Started session-26.scope. Aug 13 00:04:58.100239 systemd-logind[1901]: New session 26 of user core. Aug 13 00:04:58.079000 audit[6464]: CRED_ACQ pid=6464 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:58.112121 kernel: audit: type=1101 audit(1755043498.070:598): pid=6464 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:58.112266 kernel: audit: type=1103 audit(1755043498.079:599): pid=6464 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:58.140422 kernel: audit: type=1006 audit(1755043498.079:600): pid=6464 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=26 res=1 Aug 13 00:04:58.079000 audit[6464]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffff5540c00 a2=3 a3=1 items=0 ppid=1 pid=6464 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=26 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:58.162801 kernel: audit: type=1300 audit(1755043498.079:600): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=fffff5540c00 a2=3 a3=1 items=0 ppid=1 pid=6464 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=26 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:04:58.079000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:58.184730 kernel: audit: type=1327 audit(1755043498.079:600): proctitle=737368643A20636F7265205B707269765D Aug 13 00:04:58.145000 audit[6464]: USER_START pid=6464 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:58.196485 kernel: audit: type=1105 audit(1755043498.145:601): pid=6464 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:58.148000 audit[6467]: CRED_ACQ pid=6467 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:58.218587 kernel: audit: type=1103 audit(1755043498.148:602): pid=6467 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:58.472376 sshd[6464]: pam_unix(sshd:session): session closed for user core Aug 13 00:04:58.473000 audit[6464]: USER_END pid=6464 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:58.483533 systemd[1]: sshd@25-172.31.27.151:22-139.178.89.65:42462.service: Deactivated successfully. Aug 13 00:04:58.484986 systemd[1]: session-26.scope: Deactivated successfully. Aug 13 00:04:58.479000 audit[6464]: CRED_DISP pid=6464 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:58.496602 kernel: audit: type=1106 audit(1755043498.473:603): pid=6464 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:58.496766 kernel: audit: type=1104 audit(1755043498.479:604): pid=6464 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:04:58.482000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-172.31.27.151:22-139.178.89.65:42462 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:04:58.497601 systemd-logind[1901]: Session 26 logged out. Waiting for processes to exit. Aug 13 00:04:58.500412 systemd-logind[1901]: Removed session 26. Aug 13 00:05:03.497629 systemd[1]: Started sshd@26-172.31.27.151:22-139.178.89.65:38880.service. Aug 13 00:05:03.509640 kernel: kauditd_printk_skb: 1 callbacks suppressed Aug 13 00:05:03.509744 kernel: audit: type=1130 audit(1755043503.497:606): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-172.31.27.151:22-139.178.89.65:38880 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:05:03.497000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-172.31.27.151:22-139.178.89.65:38880 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Aug 13 00:05:03.679192 sshd[6480]: Accepted publickey for core from 139.178.89.65 port 38880 ssh2: RSA SHA256:VHsMFr69bquTAmMxwTFVl2uJICDxuaiLF9bCAtZNNGI Aug 13 00:05:03.677000 audit[6480]: USER_ACCT pid=6480 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:05:03.691841 sshd[6480]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:05:03.689000 audit[6480]: CRED_ACQ pid=6480 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:05:03.706225 kernel: audit: type=1101 audit(1755043503.677:607): pid=6480 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:05:03.706356 kernel: audit: type=1103 audit(1755043503.689:608): pid=6480 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:05:03.713854 kernel: audit: type=1006 audit(1755043503.690:609): pid=6480 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=27 res=1 Aug 13 00:05:03.690000 audit[6480]: SYSCALL arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffedef6190 a2=3 a3=1 items=0 ppid=1 pid=6480 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=27 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:05:03.727825 kernel: audit: type=1300 audit(1755043503.690:609): arch=c00000b7 syscall=64 success=yes exit=3 a0=5 a1=ffffedef6190 a2=3 a3=1 items=0 ppid=1 pid=6480 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=27 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) Aug 13 00:05:03.729230 systemd-logind[1901]: New session 27 of user core. Aug 13 00:05:03.732572 systemd[1]: Started session-27.scope. Aug 13 00:05:03.690000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D Aug 13 00:05:03.737664 kernel: audit: type=1327 audit(1755043503.690:609): proctitle=737368643A20636F7265205B707269765D Aug 13 00:05:03.752000 audit[6480]: USER_START pid=6480 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:05:03.765000 audit[6483]: CRED_ACQ pid=6483 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:05:03.768379 kernel: audit: type=1105 audit(1755043503.752:610): pid=6480 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:05:03.779289 kernel: audit: type=1103 audit(1755043503.765:611): pid=6483 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:05:04.116131 sshd[6480]: pam_unix(sshd:session): session closed for user core Aug 13 00:05:04.118000 audit[6480]: USER_END pid=6480 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:05:04.135310 systemd-logind[1901]: Session 27 logged out. Waiting for processes to exit. Aug 13 00:05:04.135888 systemd[1]: sshd@26-172.31.27.151:22-139.178.89.65:38880.service: Deactivated successfully. Aug 13 00:05:04.137965 systemd[1]: session-27.scope: Deactivated successfully. Aug 13 00:05:04.138964 systemd-logind[1901]: Removed session 27. Aug 13 00:05:04.118000 audit[6480]: CRED_DISP pid=6480 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:05:04.156823 kernel: audit: type=1106 audit(1755043504.118:612): pid=6480 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:05:04.156980 kernel: audit: type=1104 audit(1755043504.118:613): pid=6480 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=139.178.89.65 addr=139.178.89.65 terminal=ssh res=success' Aug 13 00:05:04.135000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-172.31.27.151:22-139.178.89.65:38880 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'