May 8 00:45:11.080713 kernel: Linux version 5.15.180-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP Wed May 7 23:10:51 -00 2025 May 8 00:45:11.080740 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a54efb5fced97d6fa50818abcad373184ba88ccc0f58664d2cd82270befba488 May 8 00:45:11.080754 kernel: BIOS-provided physical RAM map: May 8 00:45:11.080761 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable May 8 00:45:11.080768 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable May 8 00:45:11.080776 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS May 8 00:45:11.080785 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable May 8 00:45:11.080792 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS May 8 00:45:11.080800 kernel: BIOS-e820: [mem 0x000000000080c000-0x000000000080ffff] usable May 8 00:45:11.080809 kernel: BIOS-e820: [mem 0x0000000000810000-0x00000000008fffff] ACPI NVS May 8 00:45:11.080817 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009c8eefff] usable May 8 00:45:11.080824 kernel: BIOS-e820: [mem 0x000000009c8ef000-0x000000009cb6efff] reserved May 8 00:45:11.080832 kernel: BIOS-e820: [mem 0x000000009cb6f000-0x000000009cb7efff] ACPI data May 8 00:45:11.080840 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS May 8 00:45:11.080849 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009cf3ffff] usable May 8 00:45:11.080859 kernel: BIOS-e820: [mem 0x000000009cf40000-0x000000009cf5ffff] reserved May 8 00:45:11.080867 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS May 8 00:45:11.080875 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved May 8 00:45:11.080896 kernel: NX (Execute Disable) protection: active May 8 00:45:11.080905 kernel: e820: update [mem 0x9b475018-0x9b47ec57] usable ==> usable May 8 00:45:11.080913 kernel: e820: update [mem 0x9b475018-0x9b47ec57] usable ==> usable May 8 00:45:11.080921 kernel: e820: update [mem 0x9b438018-0x9b474e57] usable ==> usable May 8 00:45:11.080928 kernel: e820: update [mem 0x9b438018-0x9b474e57] usable ==> usable May 8 00:45:11.080936 kernel: extended physical RAM map: May 8 00:45:11.080944 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable May 8 00:45:11.080954 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable May 8 00:45:11.080962 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS May 8 00:45:11.080970 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable May 8 00:45:11.080978 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS May 8 00:45:11.080986 kernel: reserve setup_data: [mem 0x000000000080c000-0x000000000080ffff] usable May 8 00:45:11.080998 kernel: reserve setup_data: [mem 0x0000000000810000-0x00000000008fffff] ACPI NVS May 8 00:45:11.081006 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b438017] usable May 8 00:45:11.081013 kernel: reserve setup_data: [mem 0x000000009b438018-0x000000009b474e57] usable May 8 00:45:11.081021 kernel: reserve setup_data: [mem 0x000000009b474e58-0x000000009b475017] usable May 8 00:45:11.081029 kernel: reserve setup_data: [mem 0x000000009b475018-0x000000009b47ec57] usable May 8 00:45:11.081037 kernel: reserve setup_data: [mem 0x000000009b47ec58-0x000000009c8eefff] usable May 8 00:45:11.081047 kernel: reserve setup_data: [mem 0x000000009c8ef000-0x000000009cb6efff] reserved May 8 00:45:11.081055 kernel: reserve setup_data: [mem 0x000000009cb6f000-0x000000009cb7efff] ACPI data May 8 00:45:11.081063 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS May 8 00:45:11.081071 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009cf3ffff] usable May 8 00:45:11.081083 kernel: reserve setup_data: [mem 0x000000009cf40000-0x000000009cf5ffff] reserved May 8 00:45:11.081091 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS May 8 00:45:11.081100 kernel: reserve setup_data: [mem 0x00000000b0000000-0x00000000bfffffff] reserved May 8 00:45:11.081110 kernel: efi: EFI v2.70 by EDK II May 8 00:45:11.081119 kernel: efi: SMBIOS=0x9c9ab000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b673018 RNG=0x9cb73018 May 8 00:45:11.081127 kernel: random: crng init done May 8 00:45:11.081136 kernel: SMBIOS 2.8 present. May 8 00:45:11.081144 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 0.0.0 02/06/2015 May 8 00:45:11.081153 kernel: Hypervisor detected: KVM May 8 00:45:11.081161 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 May 8 00:45:11.081170 kernel: kvm-clock: cpu 0, msr 4d198001, primary cpu clock May 8 00:45:11.081178 kernel: kvm-clock: using sched offset of 6284127616 cycles May 8 00:45:11.081193 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns May 8 00:45:11.081202 kernel: tsc: Detected 2794.748 MHz processor May 8 00:45:11.081211 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved May 8 00:45:11.081219 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable May 8 00:45:11.081228 kernel: last_pfn = 0x9cf40 max_arch_pfn = 0x400000000 May 8 00:45:11.081237 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT May 8 00:45:11.081246 kernel: Using GB pages for direct mapping May 8 00:45:11.081254 kernel: Secure boot disabled May 8 00:45:11.081263 kernel: ACPI: Early table checksum verification disabled May 8 00:45:11.081278 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) May 8 00:45:11.081286 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) May 8 00:45:11.081295 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) May 8 00:45:11.081304 kernel: ACPI: DSDT 0x000000009CB7A000 0021A8 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 8 00:45:11.081313 kernel: ACPI: FACS 0x000000009CBDD000 000040 May 8 00:45:11.081322 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 8 00:45:11.081330 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 8 00:45:11.081339 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) May 8 00:45:11.081348 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 8 00:45:11.081359 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) May 8 00:45:11.081367 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] May 8 00:45:11.081376 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1a7] May 8 00:45:11.081395 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] May 8 00:45:11.081404 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] May 8 00:45:11.081412 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] May 8 00:45:11.081427 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] May 8 00:45:11.081438 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] May 8 00:45:11.081449 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] May 8 00:45:11.081463 kernel: No NUMA configuration found May 8 00:45:11.081474 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cf3ffff] May 8 00:45:11.081486 kernel: NODE_DATA(0) allocated [mem 0x9cea6000-0x9ceabfff] May 8 00:45:11.081497 kernel: Zone ranges: May 8 00:45:11.081508 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] May 8 00:45:11.081519 kernel: DMA32 [mem 0x0000000001000000-0x000000009cf3ffff] May 8 00:45:11.081529 kernel: Normal empty May 8 00:45:11.081540 kernel: Movable zone start for each node May 8 00:45:11.081551 kernel: Early memory node ranges May 8 00:45:11.081564 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] May 8 00:45:11.081575 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] May 8 00:45:11.081586 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] May 8 00:45:11.081597 kernel: node 0: [mem 0x000000000080c000-0x000000000080ffff] May 8 00:45:11.081608 kernel: node 0: [mem 0x0000000000900000-0x000000009c8eefff] May 8 00:45:11.081619 kernel: node 0: [mem 0x000000009cbff000-0x000000009cf3ffff] May 8 00:45:11.081630 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cf3ffff] May 8 00:45:11.081641 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 8 00:45:11.081650 kernel: On node 0, zone DMA: 96 pages in unavailable ranges May 8 00:45:11.081659 kernel: On node 0, zone DMA: 8 pages in unavailable ranges May 8 00:45:11.081669 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 8 00:45:11.081678 kernel: On node 0, zone DMA: 240 pages in unavailable ranges May 8 00:45:11.081687 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges May 8 00:45:11.081696 kernel: On node 0, zone DMA32: 12480 pages in unavailable ranges May 8 00:45:11.081704 kernel: ACPI: PM-Timer IO Port: 0x608 May 8 00:45:11.081713 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) May 8 00:45:11.081722 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 May 8 00:45:11.081731 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) May 8 00:45:11.081739 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) May 8 00:45:11.081750 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) May 8 00:45:11.081759 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) May 8 00:45:11.081768 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) May 8 00:45:11.081780 kernel: ACPI: Using ACPI (MADT) for SMP configuration information May 8 00:45:11.081789 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 May 8 00:45:11.081798 kernel: TSC deadline timer available May 8 00:45:11.081807 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs May 8 00:45:11.081815 kernel: kvm-guest: KVM setup pv remote TLB flush May 8 00:45:11.081824 kernel: kvm-guest: setup PV sched yield May 8 00:45:11.081835 kernel: [mem 0xc0000000-0xffffffff] available for PCI devices May 8 00:45:11.081844 kernel: Booting paravirtualized kernel on KVM May 8 00:45:11.081860 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns May 8 00:45:11.081872 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:512 nr_cpu_ids:4 nr_node_ids:1 May 8 00:45:11.081881 kernel: percpu: Embedded 56 pages/cpu s188696 r8192 d32488 u524288 May 8 00:45:11.081902 kernel: pcpu-alloc: s188696 r8192 d32488 u524288 alloc=1*2097152 May 8 00:45:11.081912 kernel: pcpu-alloc: [0] 0 1 2 3 May 8 00:45:11.081924 kernel: kvm-guest: setup async PF for cpu 0 May 8 00:45:11.081934 kernel: kvm-guest: stealtime: cpu 0, msr 9b21c0c0 May 8 00:45:11.081943 kernel: kvm-guest: PV spinlocks enabled May 8 00:45:11.081952 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) May 8 00:45:11.081961 kernel: Built 1 zonelists, mobility grouping on. Total pages: 629759 May 8 00:45:11.081973 kernel: Policy zone: DMA32 May 8 00:45:11.081983 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a54efb5fced97d6fa50818abcad373184ba88ccc0f58664d2cd82270befba488 May 8 00:45:11.081993 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 8 00:45:11.082002 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) May 8 00:45:11.082013 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 8 00:45:11.082023 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 8 00:45:11.082033 kernel: Memory: 2397432K/2567000K available (12294K kernel code, 2279K rwdata, 13724K rodata, 47464K init, 4116K bss, 169308K reserved, 0K cma-reserved) May 8 00:45:11.082042 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 May 8 00:45:11.082051 kernel: ftrace: allocating 34584 entries in 136 pages May 8 00:45:11.082060 kernel: ftrace: allocated 136 pages with 2 groups May 8 00:45:11.082069 kernel: rcu: Hierarchical RCU implementation. May 8 00:45:11.082079 kernel: rcu: RCU event tracing is enabled. May 8 00:45:11.082088 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. May 8 00:45:11.082100 kernel: Rude variant of Tasks RCU enabled. May 8 00:45:11.082109 kernel: Tracing variant of Tasks RCU enabled. May 8 00:45:11.082118 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 8 00:45:11.082128 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 May 8 00:45:11.082137 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 May 8 00:45:11.082146 kernel: Console: colour dummy device 80x25 May 8 00:45:11.082155 kernel: printk: console [ttyS0] enabled May 8 00:45:11.082164 kernel: ACPI: Core revision 20210730 May 8 00:45:11.082173 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns May 8 00:45:11.082189 kernel: APIC: Switch to symmetric I/O mode setup May 8 00:45:11.082198 kernel: x2apic enabled May 8 00:45:11.082207 kernel: Switched APIC routing to physical x2apic. May 8 00:45:11.082216 kernel: kvm-guest: setup PV IPIs May 8 00:45:11.082225 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 May 8 00:45:11.082234 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized May 8 00:45:11.082244 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) May 8 00:45:11.082253 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated May 8 00:45:11.082262 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 May 8 00:45:11.082274 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 May 8 00:45:11.082283 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization May 8 00:45:11.082292 kernel: Spectre V2 : Mitigation: Retpolines May 8 00:45:11.082302 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch May 8 00:45:11.082311 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT May 8 00:45:11.082320 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls May 8 00:45:11.082329 kernel: RETBleed: Mitigation: untrained return thunk May 8 00:45:11.082342 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier May 8 00:45:11.082352 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl and seccomp May 8 00:45:11.082363 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' May 8 00:45:11.082388 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' May 8 00:45:11.082401 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' May 8 00:45:11.082412 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 May 8 00:45:11.082424 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. May 8 00:45:11.082435 kernel: Freeing SMP alternatives memory: 32K May 8 00:45:11.082447 kernel: pid_max: default: 32768 minimum: 301 May 8 00:45:11.082458 kernel: LSM: Security Framework initializing May 8 00:45:11.082469 kernel: SELinux: Initializing. May 8 00:45:11.082489 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 8 00:45:11.082499 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 8 00:45:11.082508 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) May 8 00:45:11.082518 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. May 8 00:45:11.082527 kernel: ... version: 0 May 8 00:45:11.082536 kernel: ... bit width: 48 May 8 00:45:11.082545 kernel: ... generic registers: 6 May 8 00:45:11.082554 kernel: ... value mask: 0000ffffffffffff May 8 00:45:11.082563 kernel: ... max period: 00007fffffffffff May 8 00:45:11.082574 kernel: ... fixed-purpose events: 0 May 8 00:45:11.082583 kernel: ... event mask: 000000000000003f May 8 00:45:11.082592 kernel: signal: max sigframe size: 1776 May 8 00:45:11.082602 kernel: rcu: Hierarchical SRCU implementation. May 8 00:45:11.082610 kernel: smp: Bringing up secondary CPUs ... May 8 00:45:11.082620 kernel: x86: Booting SMP configuration: May 8 00:45:11.082629 kernel: .... node #0, CPUs: #1 May 8 00:45:11.082638 kernel: kvm-clock: cpu 1, msr 4d198041, secondary cpu clock May 8 00:45:11.082647 kernel: kvm-guest: setup async PF for cpu 1 May 8 00:45:11.082659 kernel: kvm-guest: stealtime: cpu 1, msr 9b29c0c0 May 8 00:45:11.082667 kernel: #2 May 8 00:45:11.082677 kernel: kvm-clock: cpu 2, msr 4d198081, secondary cpu clock May 8 00:45:11.082686 kernel: kvm-guest: setup async PF for cpu 2 May 8 00:45:11.082695 kernel: kvm-guest: stealtime: cpu 2, msr 9b31c0c0 May 8 00:45:11.082704 kernel: #3 May 8 00:45:11.082713 kernel: kvm-clock: cpu 3, msr 4d1980c1, secondary cpu clock May 8 00:45:11.082722 kernel: kvm-guest: setup async PF for cpu 3 May 8 00:45:11.082731 kernel: kvm-guest: stealtime: cpu 3, msr 9b39c0c0 May 8 00:45:11.082742 kernel: smp: Brought up 1 node, 4 CPUs May 8 00:45:11.082752 kernel: smpboot: Max logical packages: 1 May 8 00:45:11.082761 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) May 8 00:45:11.082770 kernel: devtmpfs: initialized May 8 00:45:11.082779 kernel: x86/mm: Memory block size: 128MB May 8 00:45:11.082788 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) May 8 00:45:11.082798 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) May 8 00:45:11.082807 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00810000-0x008fffff] (983040 bytes) May 8 00:45:11.082817 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) May 8 00:45:11.082829 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) May 8 00:45:11.082838 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 8 00:45:11.082847 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) May 8 00:45:11.082856 kernel: pinctrl core: initialized pinctrl subsystem May 8 00:45:11.082866 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 8 00:45:11.082875 kernel: audit: initializing netlink subsys (disabled) May 8 00:45:11.082896 kernel: audit: type=2000 audit(1746665109.468:1): state=initialized audit_enabled=0 res=1 May 8 00:45:11.082905 kernel: thermal_sys: Registered thermal governor 'step_wise' May 8 00:45:11.082915 kernel: thermal_sys: Registered thermal governor 'user_space' May 8 00:45:11.082926 kernel: cpuidle: using governor menu May 8 00:45:11.082935 kernel: ACPI: bus type PCI registered May 8 00:45:11.082944 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 8 00:45:11.082953 kernel: dca service started, version 1.12.1 May 8 00:45:11.082963 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) May 8 00:45:11.082972 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved in E820 May 8 00:45:11.082981 kernel: PCI: Using configuration type 1 for base access May 8 00:45:11.082994 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. May 8 00:45:11.083003 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages May 8 00:45:11.083018 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages May 8 00:45:11.083027 kernel: ACPI: Added _OSI(Module Device) May 8 00:45:11.083036 kernel: ACPI: Added _OSI(Processor Device) May 8 00:45:11.083045 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 8 00:45:11.083054 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 8 00:45:11.083063 kernel: ACPI: Added _OSI(Linux-Dell-Video) May 8 00:45:11.083072 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) May 8 00:45:11.083081 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) May 8 00:45:11.083090 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded May 8 00:45:11.083102 kernel: ACPI: Interpreter enabled May 8 00:45:11.083111 kernel: ACPI: PM: (supports S0 S3 S5) May 8 00:45:11.083120 kernel: ACPI: Using IOAPIC for interrupt routing May 8 00:45:11.083129 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug May 8 00:45:11.083138 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F May 8 00:45:11.083147 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) May 8 00:45:11.083327 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] May 8 00:45:11.083453 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] May 8 00:45:11.083561 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] May 8 00:45:11.083575 kernel: PCI host bridge to bus 0000:00 May 8 00:45:11.083696 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] May 8 00:45:11.083790 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] May 8 00:45:11.083896 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] May 8 00:45:11.083992 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] May 8 00:45:11.084085 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] May 8 00:45:11.084210 kernel: pci_bus 0000:00: root bus resource [mem 0x800000000-0xfffffffff window] May 8 00:45:11.084307 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] May 8 00:45:11.084514 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 May 8 00:45:11.084682 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 May 8 00:45:11.084805 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xc0000000-0xc0ffffff pref] May 8 00:45:11.084931 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xc1044000-0xc1044fff] May 8 00:45:11.085066 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] May 8 00:45:11.085177 kernel: pci 0000:00:01.0: BAR 0: assigned to efifb May 8 00:45:11.085282 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] May 8 00:45:11.085436 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 May 8 00:45:11.085559 kernel: pci 0000:00:02.0: reg 0x10: [io 0x6100-0x611f] May 8 00:45:11.085664 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xc1043000-0xc1043fff] May 8 00:45:11.085768 kernel: pci 0000:00:02.0: reg 0x20: [mem 0x800000000-0x800003fff 64bit pref] May 8 00:45:11.085934 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 May 8 00:45:11.086041 kernel: pci 0000:00:03.0: reg 0x10: [io 0x6000-0x607f] May 8 00:45:11.086144 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xc1042000-0xc1042fff] May 8 00:45:11.086246 kernel: pci 0000:00:03.0: reg 0x20: [mem 0x800004000-0x800007fff 64bit pref] May 8 00:45:11.086374 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 May 8 00:45:11.086512 kernel: pci 0000:00:04.0: reg 0x10: [io 0x60e0-0x60ff] May 8 00:45:11.086640 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xc1041000-0xc1041fff] May 8 00:45:11.086752 kernel: pci 0000:00:04.0: reg 0x20: [mem 0x800008000-0x80000bfff 64bit pref] May 8 00:45:11.086855 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfffc0000-0xffffffff pref] May 8 00:45:11.087008 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 May 8 00:45:11.087114 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO May 8 00:45:11.087236 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 May 8 00:45:11.087342 kernel: pci 0000:00:1f.2: reg 0x20: [io 0x60c0-0x60df] May 8 00:45:11.087512 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xc1040000-0xc1040fff] May 8 00:45:11.087661 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 May 8 00:45:11.087797 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x6080-0x60bf] May 8 00:45:11.087814 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 May 8 00:45:11.087824 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 May 8 00:45:11.087833 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 May 8 00:45:11.087842 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 May 8 00:45:11.087851 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 May 8 00:45:11.087865 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 May 8 00:45:11.087877 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 May 8 00:45:11.087900 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 May 8 00:45:11.087910 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 May 8 00:45:11.087920 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 May 8 00:45:11.087929 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 May 8 00:45:11.087938 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 May 8 00:45:11.087948 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 May 8 00:45:11.087957 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 May 8 00:45:11.087970 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 May 8 00:45:11.087980 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 May 8 00:45:11.087989 kernel: iommu: Default domain type: Translated May 8 00:45:11.087999 kernel: iommu: DMA domain TLB invalidation policy: lazy mode May 8 00:45:11.088110 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device May 8 00:45:11.088209 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none May 8 00:45:11.088311 kernel: pci 0000:00:01.0: vgaarb: bridge control possible May 8 00:45:11.088326 kernel: vgaarb: loaded May 8 00:45:11.088335 kernel: pps_core: LinuxPPS API ver. 1 registered May 8 00:45:11.088349 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti May 8 00:45:11.088358 kernel: PTP clock support registered May 8 00:45:11.088368 kernel: Registered efivars operations May 8 00:45:11.088389 kernel: PCI: Using ACPI for IRQ routing May 8 00:45:11.088399 kernel: PCI: pci_cache_line_size set to 64 bytes May 8 00:45:11.088409 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] May 8 00:45:11.088418 kernel: e820: reserve RAM buffer [mem 0x00810000-0x008fffff] May 8 00:45:11.088427 kernel: e820: reserve RAM buffer [mem 0x9b438018-0x9bffffff] May 8 00:45:11.088439 kernel: e820: reserve RAM buffer [mem 0x9b475018-0x9bffffff] May 8 00:45:11.088460 kernel: e820: reserve RAM buffer [mem 0x9c8ef000-0x9fffffff] May 8 00:45:11.088472 kernel: e820: reserve RAM buffer [mem 0x9cf40000-0x9fffffff] May 8 00:45:11.088483 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 May 8 00:45:11.088495 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter May 8 00:45:11.088507 kernel: clocksource: Switched to clocksource kvm-clock May 8 00:45:11.088520 kernel: VFS: Disk quotas dquot_6.6.0 May 8 00:45:11.088533 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 8 00:45:11.088544 kernel: pnp: PnP ACPI init May 8 00:45:11.088743 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved May 8 00:45:11.088766 kernel: pnp: PnP ACPI: found 6 devices May 8 00:45:11.088776 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns May 8 00:45:11.088785 kernel: NET: Registered PF_INET protocol family May 8 00:45:11.088795 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) May 8 00:45:11.088804 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) May 8 00:45:11.088814 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 8 00:45:11.088823 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) May 8 00:45:11.088832 kernel: TCP bind hash table entries: 32768 (order: 7, 524288 bytes, linear) May 8 00:45:11.088849 kernel: TCP: Hash tables configured (established 32768 bind 32768) May 8 00:45:11.088858 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) May 8 00:45:11.088868 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) May 8 00:45:11.088877 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 8 00:45:11.088913 kernel: NET: Registered PF_XDP protocol family May 8 00:45:11.089033 kernel: pci 0000:00:04.0: can't claim BAR 6 [mem 0xfffc0000-0xffffffff pref]: no compatible bridge window May 8 00:45:11.089139 kernel: pci 0000:00:04.0: BAR 6: assigned [mem 0x9d000000-0x9d03ffff pref] May 8 00:45:11.089233 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] May 8 00:45:11.089330 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] May 8 00:45:11.089455 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] May 8 00:45:11.089557 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] May 8 00:45:11.089647 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] May 8 00:45:11.089738 kernel: pci_bus 0000:00: resource 9 [mem 0x800000000-0xfffffffff window] May 8 00:45:11.089752 kernel: PCI: CLS 0 bytes, default 64 May 8 00:45:11.089762 kernel: Initialise system trusted keyrings May 8 00:45:11.089771 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 May 8 00:45:11.089789 kernel: Key type asymmetric registered May 8 00:45:11.089799 kernel: Asymmetric key parser 'x509' registered May 8 00:45:11.089808 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) May 8 00:45:11.089830 kernel: io scheduler mq-deadline registered May 8 00:45:11.089842 kernel: io scheduler kyber registered May 8 00:45:11.089851 kernel: io scheduler bfq registered May 8 00:45:11.089861 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 May 8 00:45:11.089871 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 May 8 00:45:11.089881 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 May 8 00:45:11.089905 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 May 8 00:45:11.089916 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 8 00:45:11.089926 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A May 8 00:45:11.089936 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 May 8 00:45:11.089945 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 May 8 00:45:11.089955 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 May 8 00:45:11.089968 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 May 8 00:45:11.090093 kernel: rtc_cmos 00:04: RTC can wake from S4 May 8 00:45:11.090192 kernel: rtc_cmos 00:04: registered as rtc0 May 8 00:45:11.090292 kernel: rtc_cmos 00:04: setting system clock to 2025-05-08T00:45:10 UTC (1746665110) May 8 00:45:11.090397 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs May 8 00:45:11.090412 kernel: efifb: probing for efifb May 8 00:45:11.090422 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k May 8 00:45:11.090432 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 May 8 00:45:11.090441 kernel: efifb: scrolling: redraw May 8 00:45:11.090451 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 May 8 00:45:11.090461 kernel: Console: switching to colour frame buffer device 160x50 May 8 00:45:11.090474 kernel: fb0: EFI VGA frame buffer device May 8 00:45:11.090484 kernel: pstore: Registered efi as persistent store backend May 8 00:45:11.090494 kernel: NET: Registered PF_INET6 protocol family May 8 00:45:11.090504 kernel: Segment Routing with IPv6 May 8 00:45:11.090520 kernel: In-situ OAM (IOAM) with IPv6 May 8 00:45:11.090530 kernel: NET: Registered PF_PACKET protocol family May 8 00:45:11.090544 kernel: Key type dns_resolver registered May 8 00:45:11.090554 kernel: IPI shorthand broadcast: enabled May 8 00:45:11.090564 kernel: sched_clock: Marking stable (630205188, 143131197)->(794800599, -21464214) May 8 00:45:11.090573 kernel: registered taskstats version 1 May 8 00:45:11.090583 kernel: Loading compiled-in X.509 certificates May 8 00:45:11.090593 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.180-flatcar: c9ff13353458e6fa2786638fdd3dcad841d1075c' May 8 00:45:11.090603 kernel: Key type .fscrypt registered May 8 00:45:11.090612 kernel: Key type fscrypt-provisioning registered May 8 00:45:11.090622 kernel: pstore: Using crash dump compression: deflate May 8 00:45:11.090637 kernel: ima: No TPM chip found, activating TPM-bypass! May 8 00:45:11.090647 kernel: ima: Allocated hash algorithm: sha1 May 8 00:45:11.090657 kernel: ima: No architecture policies found May 8 00:45:11.090666 kernel: clk: Disabling unused clocks May 8 00:45:11.090676 kernel: Freeing unused kernel image (initmem) memory: 47464K May 8 00:45:11.090686 kernel: Write protecting the kernel read-only data: 28672k May 8 00:45:11.090696 kernel: Freeing unused kernel image (text/rodata gap) memory: 2040K May 8 00:45:11.090705 kernel: Freeing unused kernel image (rodata/data gap) memory: 612K May 8 00:45:11.090715 kernel: Run /init as init process May 8 00:45:11.090731 kernel: with arguments: May 8 00:45:11.090741 kernel: /init May 8 00:45:11.090750 kernel: with environment: May 8 00:45:11.090759 kernel: HOME=/ May 8 00:45:11.090769 kernel: TERM=linux May 8 00:45:11.090778 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 8 00:45:11.090791 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) May 8 00:45:11.090803 systemd[1]: Detected virtualization kvm. May 8 00:45:11.090819 systemd[1]: Detected architecture x86-64. May 8 00:45:11.090829 systemd[1]: Running in initrd. May 8 00:45:11.090839 systemd[1]: No hostname configured, using default hostname. May 8 00:45:11.090849 systemd[1]: Hostname set to . May 8 00:45:11.090860 systemd[1]: Initializing machine ID from VM UUID. May 8 00:45:11.090870 systemd[1]: Queued start job for default target initrd.target. May 8 00:45:11.090880 systemd[1]: Started systemd-ask-password-console.path. May 8 00:45:11.090904 systemd[1]: Reached target cryptsetup.target. May 8 00:45:11.090921 systemd[1]: Reached target paths.target. May 8 00:45:11.090932 systemd[1]: Reached target slices.target. May 8 00:45:11.090942 systemd[1]: Reached target swap.target. May 8 00:45:11.090952 systemd[1]: Reached target timers.target. May 8 00:45:11.090962 systemd[1]: Listening on iscsid.socket. May 8 00:45:11.090973 systemd[1]: Listening on iscsiuio.socket. May 8 00:45:11.090983 systemd[1]: Listening on systemd-journald-audit.socket. May 8 00:45:11.090993 systemd[1]: Listening on systemd-journald-dev-log.socket. May 8 00:45:11.091010 systemd[1]: Listening on systemd-journald.socket. May 8 00:45:11.091020 systemd[1]: Listening on systemd-networkd.socket. May 8 00:45:11.091030 systemd[1]: Listening on systemd-udevd-control.socket. May 8 00:45:11.091041 systemd[1]: Listening on systemd-udevd-kernel.socket. May 8 00:45:11.091052 systemd[1]: Reached target sockets.target. May 8 00:45:11.091062 systemd[1]: Starting kmod-static-nodes.service... May 8 00:45:11.091072 systemd[1]: Finished network-cleanup.service. May 8 00:45:11.091083 systemd[1]: Starting systemd-fsck-usr.service... May 8 00:45:11.091096 systemd[1]: Starting systemd-journald.service... May 8 00:45:11.091112 systemd[1]: Starting systemd-modules-load.service... May 8 00:45:11.091136 systemd[1]: Starting systemd-resolved.service... May 8 00:45:11.091160 systemd[1]: Starting systemd-vconsole-setup.service... May 8 00:45:11.091172 systemd[1]: Finished kmod-static-nodes.service. May 8 00:45:11.091182 systemd[1]: Finished systemd-fsck-usr.service. May 8 00:45:11.091193 kernel: audit: type=1130 audit(1746665111.080:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.091204 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... May 8 00:45:11.091215 systemd[1]: Finished systemd-vconsole-setup.service. May 8 00:45:11.091234 kernel: audit: type=1130 audit(1746665111.091:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.091248 systemd-journald[197]: Journal started May 8 00:45:11.091306 systemd-journald[197]: Runtime Journal (/run/log/journal/31e67fdc4ddf4ee889a1c11818c6b221) is 6.0M, max 48.4M, 42.4M free. May 8 00:45:11.080000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.091000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.084128 systemd-modules-load[198]: Inserted module 'overlay' May 8 00:45:11.097189 systemd[1]: Started systemd-journald.service. May 8 00:45:11.101693 kernel: audit: type=1130 audit(1746665111.097:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.097000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.097488 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. May 8 00:45:11.107079 kernel: audit: type=1130 audit(1746665111.101:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.101000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.105956 systemd[1]: Starting dracut-cmdline-ask.service... May 8 00:45:11.116850 systemd-resolved[199]: Positive Trust Anchors: May 8 00:45:11.116874 systemd-resolved[199]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 8 00:45:11.116912 systemd-resolved[199]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test May 8 00:45:11.119318 systemd-resolved[199]: Defaulting to hostname 'linux'. May 8 00:45:11.116000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.120130 systemd[1]: Started systemd-resolved.service. May 8 00:45:11.124118 kernel: audit: type=1130 audit(1746665111.116:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.120479 systemd[1]: Reached target nss-lookup.target. May 8 00:45:11.135492 systemd[1]: Finished dracut-cmdline-ask.service. May 8 00:45:11.140145 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 8 00:45:11.140165 kernel: audit: type=1130 audit(1746665111.134:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.134000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.140234 systemd[1]: Starting dracut-cmdline.service... May 8 00:45:11.142264 kernel: Bridge firewalling registered May 8 00:45:11.142263 systemd-modules-load[198]: Inserted module 'br_netfilter' May 8 00:45:11.152072 dracut-cmdline[216]: dracut-dracut-053 May 8 00:45:11.154341 dracut-cmdline[216]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=a54efb5fced97d6fa50818abcad373184ba88ccc0f58664d2cd82270befba488 May 8 00:45:11.160910 kernel: SCSI subsystem initialized May 8 00:45:11.172456 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 8 00:45:11.172487 kernel: device-mapper: uevent: version 1.0.3 May 8 00:45:11.173837 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com May 8 00:45:11.177561 systemd-modules-load[198]: Inserted module 'dm_multipath' May 8 00:45:11.178159 systemd[1]: Finished systemd-modules-load.service. May 8 00:45:11.183498 kernel: audit: type=1130 audit(1746665111.179:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.179000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.183331 systemd[1]: Starting systemd-sysctl.service... May 8 00:45:11.191483 systemd[1]: Finished systemd-sysctl.service. May 8 00:45:11.196241 kernel: audit: type=1130 audit(1746665111.190:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.190000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.227933 kernel: Loading iSCSI transport class v2.0-870. May 8 00:45:11.245913 kernel: iscsi: registered transport (tcp) May 8 00:45:11.266914 kernel: iscsi: registered transport (qla4xxx) May 8 00:45:11.266941 kernel: QLogic iSCSI HBA Driver May 8 00:45:11.295351 systemd[1]: Finished dracut-cmdline.service. May 8 00:45:11.300745 kernel: audit: type=1130 audit(1746665111.294:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.294000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.296579 systemd[1]: Starting dracut-pre-udev.service... May 8 00:45:11.341942 kernel: raid6: avx2x4 gen() 28765 MB/s May 8 00:45:11.358952 kernel: raid6: avx2x4 xor() 7860 MB/s May 8 00:45:11.375917 kernel: raid6: avx2x2 gen() 31013 MB/s May 8 00:45:11.392912 kernel: raid6: avx2x2 xor() 18513 MB/s May 8 00:45:11.409921 kernel: raid6: avx2x1 gen() 25057 MB/s May 8 00:45:11.426913 kernel: raid6: avx2x1 xor() 14772 MB/s May 8 00:45:11.443915 kernel: raid6: sse2x4 gen() 13929 MB/s May 8 00:45:11.460925 kernel: raid6: sse2x4 xor() 7220 MB/s May 8 00:45:11.477919 kernel: raid6: sse2x2 gen() 15162 MB/s May 8 00:45:11.494915 kernel: raid6: sse2x2 xor() 9265 MB/s May 8 00:45:11.511917 kernel: raid6: sse2x1 gen() 11455 MB/s May 8 00:45:11.529362 kernel: raid6: sse2x1 xor() 7447 MB/s May 8 00:45:11.529390 kernel: raid6: using algorithm avx2x2 gen() 31013 MB/s May 8 00:45:11.529400 kernel: raid6: .... xor() 18513 MB/s, rmw enabled May 8 00:45:11.530103 kernel: raid6: using avx2x2 recovery algorithm May 8 00:45:11.542926 kernel: xor: automatically using best checksumming function avx May 8 00:45:11.641940 kernel: Btrfs loaded, crc32c=crc32c-intel, zoned=no, fsverity=no May 8 00:45:11.651525 systemd[1]: Finished dracut-pre-udev.service. May 8 00:45:11.650000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.653000 audit: BPF prog-id=7 op=LOAD May 8 00:45:11.653000 audit: BPF prog-id=8 op=LOAD May 8 00:45:11.654117 systemd[1]: Starting systemd-udevd.service... May 8 00:45:11.667297 systemd-udevd[402]: Using default interface naming scheme 'v252'. May 8 00:45:11.671663 systemd[1]: Started systemd-udevd.service. May 8 00:45:11.672526 systemd[1]: Starting dracut-pre-trigger.service... May 8 00:45:11.670000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.683796 dracut-pre-trigger[405]: rd.md=0: removing MD RAID activation May 8 00:45:11.712700 systemd[1]: Finished dracut-pre-trigger.service. May 8 00:45:11.711000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.713643 systemd[1]: Starting systemd-udev-trigger.service... May 8 00:45:11.760113 systemd[1]: Finished systemd-udev-trigger.service. May 8 00:45:11.759000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:11.791918 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) May 8 00:45:11.797607 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. May 8 00:45:11.797630 kernel: GPT:9289727 != 19775487 May 8 00:45:11.797639 kernel: GPT:Alternate GPT header not at the end of the disk. May 8 00:45:11.797648 kernel: GPT:9289727 != 19775487 May 8 00:45:11.797656 kernel: GPT: Use GNU Parted to correct GPT errors. May 8 00:45:11.797664 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 8 00:45:11.801912 kernel: cryptd: max_cpu_qlen set to 1000 May 8 00:45:11.810904 kernel: libata version 3.00 loaded. May 8 00:45:11.812999 kernel: AVX2 version of gcm_enc/dec engaged. May 8 00:45:11.813027 kernel: AES CTR mode by8 optimization enabled May 8 00:45:11.821602 kernel: ahci 0000:00:1f.2: version 3.0 May 8 00:45:11.853474 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 May 8 00:45:11.853498 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode May 8 00:45:11.853646 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only May 8 00:45:11.853781 kernel: scsi host0: ahci May 8 00:45:11.853975 kernel: scsi host1: ahci May 8 00:45:11.854122 kernel: scsi host2: ahci May 8 00:45:11.854253 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (447) May 8 00:45:11.854269 kernel: scsi host3: ahci May 8 00:45:11.854412 kernel: scsi host4: ahci May 8 00:45:11.854536 kernel: scsi host5: ahci May 8 00:45:11.854690 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 May 8 00:45:11.854705 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 May 8 00:45:11.854722 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 May 8 00:45:11.854739 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 May 8 00:45:11.854751 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 May 8 00:45:11.854764 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 May 8 00:45:11.837955 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. May 8 00:45:11.854375 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. May 8 00:45:11.859643 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. May 8 00:45:11.860710 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. May 8 00:45:11.867955 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. May 8 00:45:11.870540 systemd[1]: Starting disk-uuid.service... May 8 00:45:11.877879 disk-uuid[525]: Primary Header is updated. May 8 00:45:11.877879 disk-uuid[525]: Secondary Entries is updated. May 8 00:45:11.877879 disk-uuid[525]: Secondary Header is updated. May 8 00:45:11.881695 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 8 00:45:11.884916 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 8 00:45:12.158927 kernel: ata2: SATA link down (SStatus 0 SControl 300) May 8 00:45:12.159035 kernel: ata4: SATA link down (SStatus 0 SControl 300) May 8 00:45:12.159913 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) May 8 00:45:12.162935 kernel: ata5: SATA link down (SStatus 0 SControl 300) May 8 00:45:12.163030 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 May 8 00:45:12.163045 kernel: ata3.00: applying bridge limits May 8 00:45:12.164451 kernel: ata3.00: configured for UDMA/100 May 8 00:45:12.168930 kernel: ata6: SATA link down (SStatus 0 SControl 300) May 8 00:45:12.168955 kernel: ata1: SATA link down (SStatus 0 SControl 300) May 8 00:45:12.169923 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 May 8 00:45:12.205427 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray May 8 00:45:12.222672 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 May 8 00:45:12.222685 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 May 8 00:45:12.888949 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 8 00:45:12.890202 disk-uuid[526]: The operation has completed successfully. May 8 00:45:12.971695 systemd[1]: disk-uuid.service: Deactivated successfully. May 8 00:45:12.971799 systemd[1]: Finished disk-uuid.service. May 8 00:45:12.973000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:12.973000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:12.981349 systemd[1]: Starting verity-setup.service... May 8 00:45:12.995914 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" May 8 00:45:13.020628 systemd[1]: Found device dev-mapper-usr.device. May 8 00:45:13.022475 systemd[1]: Mounting sysusr-usr.mount... May 8 00:45:13.024939 systemd[1]: Finished verity-setup.service. May 8 00:45:13.026000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:13.097911 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. May 8 00:45:13.098337 systemd[1]: Mounted sysusr-usr.mount. May 8 00:45:13.098605 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. May 8 00:45:13.099523 systemd[1]: Starting ignition-setup.service... May 8 00:45:13.101592 systemd[1]: Starting parse-ip-for-networkd.service... May 8 00:45:13.110143 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 8 00:45:13.110172 kernel: BTRFS info (device vda6): using free space tree May 8 00:45:13.110182 kernel: BTRFS info (device vda6): has skinny extents May 8 00:45:13.119822 systemd[1]: mnt-oem.mount: Deactivated successfully. May 8 00:45:13.129078 systemd[1]: Finished ignition-setup.service. May 8 00:45:13.130000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:13.132071 systemd[1]: Starting ignition-fetch-offline.service... May 8 00:45:13.207021 systemd[1]: Finished parse-ip-for-networkd.service. May 8 00:45:13.208000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:13.208000 audit: BPF prog-id=9 op=LOAD May 8 00:45:13.210394 systemd[1]: Starting systemd-networkd.service... May 8 00:45:13.227846 ignition[641]: Ignition 2.14.0 May 8 00:45:13.227856 ignition[641]: Stage: fetch-offline May 8 00:45:13.227960 ignition[641]: no configs at "/usr/lib/ignition/base.d" May 8 00:45:13.227972 ignition[641]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 8 00:45:13.228091 ignition[641]: parsed url from cmdline: "" May 8 00:45:13.228094 ignition[641]: no config URL provided May 8 00:45:13.228099 ignition[641]: reading system config file "/usr/lib/ignition/user.ign" May 8 00:45:13.228105 ignition[641]: no config at "/usr/lib/ignition/user.ign" May 8 00:45:13.228122 ignition[641]: op(1): [started] loading QEMU firmware config module May 8 00:45:13.228126 ignition[641]: op(1): executing: "modprobe" "qemu_fw_cfg" May 8 00:45:13.238085 ignition[641]: op(1): [finished] loading QEMU firmware config module May 8 00:45:13.239051 systemd-networkd[714]: lo: Link UP May 8 00:45:13.239056 systemd-networkd[714]: lo: Gained carrier May 8 00:45:13.240000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:13.239692 systemd-networkd[714]: Enumeration completed May 8 00:45:13.239796 systemd[1]: Started systemd-networkd.service. May 8 00:45:13.240779 systemd-networkd[714]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 8 00:45:13.241290 systemd[1]: Reached target network.target. May 8 00:45:13.242821 systemd-networkd[714]: eth0: Link UP May 8 00:45:13.242825 systemd-networkd[714]: eth0: Gained carrier May 8 00:45:13.244019 systemd[1]: Starting iscsiuio.service... May 8 00:45:13.307042 systemd[1]: Started iscsiuio.service. May 8 00:45:13.307000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:13.309276 systemd[1]: Starting iscsid.service... May 8 00:45:13.314667 iscsid[721]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi May 8 00:45:13.314667 iscsid[721]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. May 8 00:45:13.314667 iscsid[721]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. May 8 00:45:13.314667 iscsid[721]: If using hardware iscsi like qla4xxx this message can be ignored. May 8 00:45:13.314667 iscsid[721]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi May 8 00:45:13.323000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:13.326773 ignition[641]: parsing config with SHA512: bd2aeb1a384c4c2931a0ac0998a8e0b37b9e9b632ead020e669503830034a3bfe9c20d936c8f4c7b0719bdcca22f132fd202f9f28c66dea250f9c3d3ac0b4b68 May 8 00:45:13.329698 iscsid[721]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf May 8 00:45:13.317240 systemd[1]: Started iscsid.service. May 8 00:45:13.325346 systemd[1]: Starting dracut-initqueue.service... May 8 00:45:13.335197 systemd-networkd[714]: eth0: DHCPv4 address 10.0.0.82/16, gateway 10.0.0.1 acquired from 10.0.0.1 May 8 00:45:13.342583 systemd[1]: Finished dracut-initqueue.service. May 8 00:45:13.343000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:13.344782 systemd[1]: Reached target remote-fs-pre.target. May 8 00:45:13.346808 systemd[1]: Reached target remote-cryptsetup.target. May 8 00:45:13.348893 systemd[1]: Reached target remote-fs.target. May 8 00:45:13.351625 systemd[1]: Starting dracut-pre-mount.service... May 8 00:45:13.352989 unknown[641]: fetched base config from "system" May 8 00:45:13.353584 ignition[641]: fetch-offline: fetch-offline passed May 8 00:45:13.352999 unknown[641]: fetched user config from "qemu" May 8 00:45:13.353695 ignition[641]: Ignition finished successfully May 8 00:45:13.357763 systemd[1]: Finished ignition-fetch-offline.service. May 8 00:45:13.357000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:13.358859 systemd[1]: ignition-fetch.service was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). May 8 00:45:13.359578 systemd[1]: Starting ignition-kargs.service... May 8 00:45:13.367548 systemd[1]: Finished dracut-pre-mount.service. May 8 00:45:13.368000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:13.459543 ignition[732]: Ignition 2.14.0 May 8 00:45:13.459557 ignition[732]: Stage: kargs May 8 00:45:13.459692 ignition[732]: no configs at "/usr/lib/ignition/base.d" May 8 00:45:13.459702 ignition[732]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 8 00:45:13.461242 ignition[732]: kargs: kargs passed May 8 00:45:13.464000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:13.462806 systemd[1]: Finished ignition-kargs.service. May 8 00:45:13.461292 ignition[732]: Ignition finished successfully May 8 00:45:13.465613 systemd[1]: Starting ignition-disks.service... May 8 00:45:13.481774 ignition[741]: Ignition 2.14.0 May 8 00:45:13.481786 ignition[741]: Stage: disks May 8 00:45:13.481903 ignition[741]: no configs at "/usr/lib/ignition/base.d" May 8 00:45:13.481913 ignition[741]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 8 00:45:13.484000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:13.484329 systemd[1]: Finished ignition-disks.service. May 8 00:45:13.483300 ignition[741]: disks: disks passed May 8 00:45:13.485283 systemd[1]: Reached target initrd-root-device.target. May 8 00:45:13.483351 ignition[741]: Ignition finished successfully May 8 00:45:13.487145 systemd[1]: Reached target local-fs-pre.target. May 8 00:45:13.488046 systemd[1]: Reached target local-fs.target. May 8 00:45:13.489658 systemd[1]: Reached target sysinit.target. May 8 00:45:13.491159 systemd[1]: Reached target basic.target. May 8 00:45:13.493944 systemd[1]: Starting systemd-fsck-root.service... May 8 00:45:13.509325 systemd-fsck[749]: ROOT: clean, 623/553520 files, 56023/553472 blocks May 8 00:45:13.521425 systemd[1]: Finished systemd-fsck-root.service. May 8 00:45:13.522000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:13.523702 systemd[1]: Mounting sysroot.mount... May 8 00:45:13.544916 kernel: EXT4-fs (vda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. May 8 00:45:13.545616 systemd[1]: Mounted sysroot.mount. May 8 00:45:13.547128 systemd[1]: Reached target initrd-root-fs.target. May 8 00:45:13.549147 systemd[1]: Mounting sysroot-usr.mount... May 8 00:45:13.550110 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. May 8 00:45:13.550141 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 8 00:45:13.550161 systemd[1]: Reached target ignition-diskful.target. May 8 00:45:13.552278 systemd[1]: Mounted sysroot-usr.mount. May 8 00:45:13.554427 systemd[1]: Starting initrd-setup-root.service... May 8 00:45:13.559645 initrd-setup-root[759]: cut: /sysroot/etc/passwd: No such file or directory May 8 00:45:13.563481 initrd-setup-root[767]: cut: /sysroot/etc/group: No such file or directory May 8 00:45:13.566570 initrd-setup-root[775]: cut: /sysroot/etc/shadow: No such file or directory May 8 00:45:13.569727 initrd-setup-root[783]: cut: /sysroot/etc/gshadow: No such file or directory May 8 00:45:13.598476 systemd[1]: Finished initrd-setup-root.service. May 8 00:45:13.599000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:13.600477 systemd[1]: Starting ignition-mount.service... May 8 00:45:13.602587 systemd[1]: Starting sysroot-boot.service... May 8 00:45:13.607838 bash[800]: umount: /sysroot/usr/share/oem: not mounted. May 8 00:45:13.621803 ignition[801]: INFO : Ignition 2.14.0 May 8 00:45:13.621803 ignition[801]: INFO : Stage: mount May 8 00:45:13.623841 ignition[801]: INFO : no configs at "/usr/lib/ignition/base.d" May 8 00:45:13.623841 ignition[801]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 8 00:45:13.623841 ignition[801]: INFO : mount: mount passed May 8 00:45:13.623841 ignition[801]: INFO : Ignition finished successfully May 8 00:45:13.623000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:13.623731 systemd[1]: Finished ignition-mount.service. May 8 00:45:13.643782 systemd[1]: Finished sysroot-boot.service. May 8 00:45:13.644000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:14.033643 systemd[1]: Mounting sysroot-usr-share-oem.mount... May 8 00:45:14.042579 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (810) May 8 00:45:14.045723 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 8 00:45:14.045744 kernel: BTRFS info (device vda6): using free space tree May 8 00:45:14.045766 kernel: BTRFS info (device vda6): has skinny extents May 8 00:45:14.049246 systemd[1]: Mounted sysroot-usr-share-oem.mount. May 8 00:45:14.050178 systemd[1]: Starting ignition-files.service... May 8 00:45:14.084124 ignition[830]: INFO : Ignition 2.14.0 May 8 00:45:14.084124 ignition[830]: INFO : Stage: files May 8 00:45:14.086098 ignition[830]: INFO : no configs at "/usr/lib/ignition/base.d" May 8 00:45:14.086098 ignition[830]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 8 00:45:14.088525 ignition[830]: DEBUG : files: compiled without relabeling support, skipping May 8 00:45:14.090153 ignition[830]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 8 00:45:14.090153 ignition[830]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 8 00:45:14.093071 ignition[830]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 8 00:45:14.094710 ignition[830]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 8 00:45:14.096237 ignition[830]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 8 00:45:14.096237 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" May 8 00:45:14.096237 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" May 8 00:45:14.096237 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" May 8 00:45:14.096237 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 May 8 00:45:14.095379 unknown[830]: wrote ssh authorized keys file for user: core May 8 00:45:14.151760 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK May 8 00:45:14.388411 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" May 8 00:45:14.388411 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" May 8 00:45:14.392671 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" May 8 00:45:14.392671 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" May 8 00:45:14.392671 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" May 8 00:45:14.392671 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 8 00:45:14.392671 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 8 00:45:14.392671 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 8 00:45:14.392671 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 8 00:45:14.392671 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" May 8 00:45:14.392671 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 8 00:45:14.392671 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 8 00:45:14.392671 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 8 00:45:14.392671 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 8 00:45:14.392671 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-x86-64.raw: attempt #1 May 8 00:45:14.840857 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK May 8 00:45:15.000099 systemd-networkd[714]: eth0: Gained IPv6LL May 8 00:45:16.245307 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 8 00:45:16.245307 ignition[830]: INFO : files: op(c): [started] processing unit "containerd.service" May 8 00:45:16.249863 ignition[830]: INFO : files: op(c): op(d): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" May 8 00:45:16.249863 ignition[830]: INFO : files: op(c): op(d): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" May 8 00:45:16.249863 ignition[830]: INFO : files: op(c): [finished] processing unit "containerd.service" May 8 00:45:16.249863 ignition[830]: INFO : files: op(e): [started] processing unit "prepare-helm.service" May 8 00:45:16.249863 ignition[830]: INFO : files: op(e): op(f): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 8 00:45:16.249863 ignition[830]: INFO : files: op(e): op(f): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 8 00:45:16.249863 ignition[830]: INFO : files: op(e): [finished] processing unit "prepare-helm.service" May 8 00:45:16.249863 ignition[830]: INFO : files: op(10): [started] processing unit "coreos-metadata.service" May 8 00:45:16.249863 ignition[830]: INFO : files: op(10): op(11): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" May 8 00:45:16.249863 ignition[830]: INFO : files: op(10): op(11): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" May 8 00:45:16.249863 ignition[830]: INFO : files: op(10): [finished] processing unit "coreos-metadata.service" May 8 00:45:16.249863 ignition[830]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" May 8 00:45:16.249863 ignition[830]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" May 8 00:45:16.249863 ignition[830]: INFO : files: op(13): [started] setting preset to disabled for "coreos-metadata.service" May 8 00:45:16.249863 ignition[830]: INFO : files: op(13): op(14): [started] removing enablement symlink(s) for "coreos-metadata.service" May 8 00:45:16.302967 ignition[830]: INFO : files: op(13): op(14): [finished] removing enablement symlink(s) for "coreos-metadata.service" May 8 00:45:16.302967 ignition[830]: INFO : files: op(13): [finished] setting preset to disabled for "coreos-metadata.service" May 8 00:45:16.306205 ignition[830]: INFO : files: createResultFile: createFiles: op(15): [started] writing file "/sysroot/etc/.ignition-result.json" May 8 00:45:16.306205 ignition[830]: INFO : files: createResultFile: createFiles: op(15): [finished] writing file "/sysroot/etc/.ignition-result.json" May 8 00:45:16.306205 ignition[830]: INFO : files: files passed May 8 00:45:16.306205 ignition[830]: INFO : Ignition finished successfully May 8 00:45:16.313059 systemd[1]: Finished ignition-files.service. May 8 00:45:16.318810 kernel: kauditd_printk_skb: 24 callbacks suppressed May 8 00:45:16.318843 kernel: audit: type=1130 audit(1746665116.314:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.314000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.318966 systemd[1]: Starting initrd-setup-root-after-ignition.service... May 8 00:45:16.319998 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). May 8 00:45:16.320914 systemd[1]: Starting ignition-quench.service... May 8 00:45:16.331077 kernel: audit: type=1130 audit(1746665116.324:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.331095 kernel: audit: type=1131 audit(1746665116.324:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.324000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.324000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.323683 systemd[1]: ignition-quench.service: Deactivated successfully. May 8 00:45:16.323756 systemd[1]: Finished ignition-quench.service. May 8 00:45:16.336309 initrd-setup-root-after-ignition[856]: grep: /sysroot/usr/share/oem/oem-release: No such file or directory May 8 00:45:16.339466 initrd-setup-root-after-ignition[858]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 8 00:45:16.341526 systemd[1]: Finished initrd-setup-root-after-ignition.service. May 8 00:45:16.343000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.343754 systemd[1]: Reached target ignition-complete.target. May 8 00:45:16.348833 kernel: audit: type=1130 audit(1746665116.343:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.348794 systemd[1]: Starting initrd-parse-etc.service... May 8 00:45:16.362303 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 8 00:45:16.362396 systemd[1]: Finished initrd-parse-etc.service. May 8 00:45:16.370923 kernel: audit: type=1130 audit(1746665116.364:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.370947 kernel: audit: type=1131 audit(1746665116.364:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.364000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.364000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.364295 systemd[1]: Reached target initrd-fs.target. May 8 00:45:16.371811 systemd[1]: Reached target initrd.target. May 8 00:45:16.373359 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. May 8 00:45:16.374199 systemd[1]: Starting dracut-pre-pivot.service... May 8 00:45:16.386779 systemd[1]: Finished dracut-pre-pivot.service. May 8 00:45:16.392356 kernel: audit: type=1130 audit(1746665116.387:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.387000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.388492 systemd[1]: Starting initrd-cleanup.service... May 8 00:45:16.399995 systemd[1]: Stopped target nss-lookup.target. May 8 00:45:16.400982 systemd[1]: Stopped target remote-cryptsetup.target. May 8 00:45:16.402739 systemd[1]: Stopped target timers.target. May 8 00:45:16.404521 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 8 00:45:16.410125 kernel: audit: type=1131 audit(1746665116.405:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.405000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.404658 systemd[1]: Stopped dracut-pre-pivot.service. May 8 00:45:16.406334 systemd[1]: Stopped target initrd.target. May 8 00:45:16.411071 systemd[1]: Stopped target basic.target. May 8 00:45:16.411961 systemd[1]: Stopped target ignition-complete.target. May 8 00:45:16.414323 systemd[1]: Stopped target ignition-diskful.target. May 8 00:45:16.415419 systemd[1]: Stopped target initrd-root-device.target. May 8 00:45:16.417216 systemd[1]: Stopped target remote-fs.target. May 8 00:45:16.419014 systemd[1]: Stopped target remote-fs-pre.target. May 8 00:45:16.420829 systemd[1]: Stopped target sysinit.target. May 8 00:45:16.421815 systemd[1]: Stopped target local-fs.target. May 8 00:45:16.424112 systemd[1]: Stopped target local-fs-pre.target. May 8 00:45:16.427000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.425214 systemd[1]: Stopped target swap.target. May 8 00:45:16.434409 kernel: audit: type=1131 audit(1746665116.427:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.425584 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 8 00:45:16.439551 kernel: audit: type=1131 audit(1746665116.434:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.434000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.425736 systemd[1]: Stopped dracut-pre-mount.service. May 8 00:45:16.439000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.429001 systemd[1]: Stopped target cryptsetup.target. May 8 00:45:16.433708 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 8 00:45:16.433830 systemd[1]: Stopped dracut-initqueue.service. May 8 00:45:16.435481 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 8 00:45:16.435620 systemd[1]: Stopped ignition-fetch-offline.service. May 8 00:45:16.440707 systemd[1]: Stopped target paths.target. May 8 00:45:16.442275 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 8 00:45:16.446734 systemd[1]: Stopped systemd-ask-password-console.path. May 8 00:45:16.450262 systemd[1]: Stopped target slices.target. May 8 00:45:16.451918 systemd[1]: Stopped target sockets.target. May 8 00:45:16.453533 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 8 00:45:16.454794 systemd[1]: Stopped initrd-setup-root-after-ignition.service. May 8 00:45:16.456000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.456956 systemd[1]: ignition-files.service: Deactivated successfully. May 8 00:45:16.457957 systemd[1]: Stopped ignition-files.service. May 8 00:45:16.459000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.460480 systemd[1]: Stopping ignition-mount.service... May 8 00:45:16.462209 systemd[1]: Stopping iscsid.service... May 8 00:45:16.463860 iscsid[721]: iscsid shutting down. May 8 00:45:16.465834 systemd[1]: Stopping sysroot-boot.service... May 8 00:45:16.466798 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 8 00:45:16.468000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.469739 ignition[871]: INFO : Ignition 2.14.0 May 8 00:45:16.469739 ignition[871]: INFO : Stage: umount May 8 00:45:16.469739 ignition[871]: INFO : no configs at "/usr/lib/ignition/base.d" May 8 00:45:16.469739 ignition[871]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 8 00:45:16.470000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.473000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.467020 systemd[1]: Stopped systemd-udev-trigger.service. May 8 00:45:16.476000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.477141 ignition[871]: INFO : umount: umount passed May 8 00:45:16.477141 ignition[871]: INFO : Ignition finished successfully May 8 00:45:16.468804 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 8 00:45:16.480000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.468953 systemd[1]: Stopped dracut-pre-trigger.service. May 8 00:45:16.481000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.472945 systemd[1]: iscsid.service: Deactivated successfully. May 8 00:45:16.483000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.473062 systemd[1]: Stopped iscsid.service. May 8 00:45:16.473806 systemd[1]: ignition-mount.service: Deactivated successfully. May 8 00:45:16.486000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.486000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.473923 systemd[1]: Stopped ignition-mount.service. May 8 00:45:16.476610 systemd[1]: iscsid.socket: Deactivated successfully. May 8 00:45:16.489000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.476697 systemd[1]: Closed iscsid.socket. May 8 00:45:16.477857 systemd[1]: ignition-disks.service: Deactivated successfully. May 8 00:45:16.477928 systemd[1]: Stopped ignition-disks.service. May 8 00:45:16.480255 systemd[1]: ignition-kargs.service: Deactivated successfully. May 8 00:45:16.480301 systemd[1]: Stopped ignition-kargs.service. May 8 00:45:16.482071 systemd[1]: ignition-setup.service: Deactivated successfully. May 8 00:45:16.482132 systemd[1]: Stopped ignition-setup.service. May 8 00:45:16.483662 systemd[1]: Stopping iscsiuio.service... May 8 00:45:16.485448 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 8 00:45:16.485557 systemd[1]: Finished initrd-cleanup.service. May 8 00:45:16.487782 systemd[1]: iscsiuio.service: Deactivated successfully. May 8 00:45:16.487913 systemd[1]: Stopped iscsiuio.service. May 8 00:45:16.489461 systemd[1]: Stopped target network.target. May 8 00:45:16.504000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.491054 systemd[1]: iscsiuio.socket: Deactivated successfully. May 8 00:45:16.491096 systemd[1]: Closed iscsiuio.socket. May 8 00:45:16.491845 systemd[1]: Stopping systemd-networkd.service... May 8 00:45:16.493568 systemd[1]: Stopping systemd-resolved.service... May 8 00:45:16.511000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.500966 systemd-networkd[714]: eth0: DHCPv6 lease lost May 8 00:45:16.512000 audit: BPF prog-id=9 op=UNLOAD May 8 00:45:16.513000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.502791 systemd[1]: systemd-networkd.service: Deactivated successfully. May 8 00:45:16.514000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.502870 systemd[1]: Stopped systemd-networkd.service. May 8 00:45:16.505472 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 8 00:45:16.505502 systemd[1]: Closed systemd-networkd.socket. May 8 00:45:16.507560 systemd[1]: Stopping network-cleanup.service... May 8 00:45:16.509153 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 8 00:45:16.509207 systemd[1]: Stopped parse-ip-for-networkd.service. May 8 00:45:16.511387 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 8 00:45:16.511425 systemd[1]: Stopped systemd-sysctl.service. May 8 00:45:16.514392 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 8 00:45:16.514431 systemd[1]: Stopped systemd-modules-load.service. May 8 00:45:16.527000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.519975 systemd[1]: Stopping systemd-udevd.service... May 8 00:45:16.523727 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 8 00:45:16.528000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.523800 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 8 00:45:16.525963 systemd[1]: systemd-resolved.service: Deactivated successfully. May 8 00:45:16.536000 audit: BPF prog-id=6 op=UNLOAD May 8 00:45:16.526090 systemd[1]: Stopped systemd-resolved.service. May 8 00:45:16.528151 systemd[1]: systemd-udevd.service: Deactivated successfully. May 8 00:45:16.528306 systemd[1]: Stopped systemd-udevd.service. May 8 00:45:16.529418 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 8 00:45:16.529458 systemd[1]: Closed systemd-udevd-control.socket. May 8 00:45:16.529723 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 8 00:45:16.529753 systemd[1]: Closed systemd-udevd-kernel.socket. May 8 00:45:16.536054 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 8 00:45:16.536155 systemd[1]: Stopped dracut-pre-udev.service. May 8 00:45:16.544000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.544467 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 8 00:45:16.546000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.544506 systemd[1]: Stopped dracut-cmdline.service. May 8 00:45:16.546304 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 8 00:45:16.547059 systemd[1]: Stopped dracut-cmdline-ask.service. May 8 00:45:16.549000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.550957 systemd[1]: Starting initrd-udevadm-cleanup-db.service... May 8 00:45:16.552717 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 8 00:45:16.554000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.552763 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service. May 8 00:45:16.554927 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 8 00:45:16.557000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.555758 systemd[1]: Stopped kmod-static-nodes.service. May 8 00:45:16.557652 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 8 00:45:16.557689 systemd[1]: Stopped systemd-vconsole-setup.service. May 8 00:45:16.560000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.562069 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. May 8 00:45:16.564878 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 8 00:45:16.566088 systemd[1]: Finished initrd-udevadm-cleanup-db.service. May 8 00:45:16.567000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.567000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.568185 systemd[1]: network-cleanup.service: Deactivated successfully. May 8 00:45:16.569146 systemd[1]: Stopped network-cleanup.service. May 8 00:45:16.569000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.597451 systemd[1]: sysroot-boot.service: Deactivated successfully. May 8 00:45:16.597562 systemd[1]: Stopped sysroot-boot.service. May 8 00:45:16.599000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.599451 systemd[1]: Reached target initrd-switch-root.target. May 8 00:45:16.600962 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 8 00:45:16.602000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:16.601006 systemd[1]: Stopped initrd-setup-root.service. May 8 00:45:16.603643 systemd[1]: Starting initrd-switch-root.service... May 8 00:45:16.611454 systemd[1]: Switching root. May 8 00:45:16.612000 audit: BPF prog-id=5 op=UNLOAD May 8 00:45:16.612000 audit: BPF prog-id=4 op=UNLOAD May 8 00:45:16.612000 audit: BPF prog-id=3 op=UNLOAD May 8 00:45:16.613000 audit: BPF prog-id=8 op=UNLOAD May 8 00:45:16.613000 audit: BPF prog-id=7 op=UNLOAD May 8 00:45:16.630170 systemd-journald[197]: Journal stopped May 8 00:45:21.599673 systemd-journald[197]: Received SIGTERM from PID 1 (systemd). May 8 00:45:21.599731 kernel: SELinux: Class mctp_socket not defined in policy. May 8 00:45:21.599744 kernel: SELinux: Class anon_inode not defined in policy. May 8 00:45:21.599755 kernel: SELinux: the above unknown classes and permissions will be allowed May 8 00:45:21.599765 kernel: SELinux: policy capability network_peer_controls=1 May 8 00:45:21.599774 kernel: SELinux: policy capability open_perms=1 May 8 00:45:21.599784 kernel: SELinux: policy capability extended_socket_class=1 May 8 00:45:21.599796 kernel: SELinux: policy capability always_check_network=0 May 8 00:45:21.599810 kernel: SELinux: policy capability cgroup_seclabel=1 May 8 00:45:21.599828 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 8 00:45:21.599839 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 8 00:45:21.599848 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 8 00:45:21.599859 systemd[1]: Successfully loaded SELinux policy in 86.357ms. May 8 00:45:21.599876 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 6.857ms. May 8 00:45:21.599906 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) May 8 00:45:21.599926 systemd[1]: Detected virtualization kvm. May 8 00:45:21.599954 systemd[1]: Detected architecture x86-64. May 8 00:45:21.599970 systemd[1]: Detected first boot. May 8 00:45:21.599981 systemd[1]: Initializing machine ID from VM UUID. May 8 00:45:21.599991 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). May 8 00:45:21.600002 systemd[1]: Populated /etc with preset unit settings. May 8 00:45:21.600012 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 8 00:45:21.600031 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 8 00:45:21.600043 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 8 00:45:21.600062 systemd[1]: Queued start job for default target multi-user.target. May 8 00:45:21.600072 systemd[1]: Unnecessary job was removed for dev-vda6.device. May 8 00:45:21.600083 systemd[1]: Created slice system-addon\x2dconfig.slice. May 8 00:45:21.600093 systemd[1]: Created slice system-addon\x2drun.slice. May 8 00:45:21.600112 systemd[1]: Created slice system-getty.slice. May 8 00:45:21.600123 systemd[1]: Created slice system-modprobe.slice. May 8 00:45:21.600138 systemd[1]: Created slice system-serial\x2dgetty.slice. May 8 00:45:21.600149 systemd[1]: Created slice system-system\x2dcloudinit.slice. May 8 00:45:21.600160 systemd[1]: Created slice system-systemd\x2dfsck.slice. May 8 00:45:21.600171 systemd[1]: Created slice user.slice. May 8 00:45:21.600181 systemd[1]: Started systemd-ask-password-console.path. May 8 00:45:21.600192 systemd[1]: Started systemd-ask-password-wall.path. May 8 00:45:21.600202 systemd[1]: Set up automount boot.automount. May 8 00:45:21.600214 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. May 8 00:45:21.600225 systemd[1]: Reached target integritysetup.target. May 8 00:45:21.600239 systemd[1]: Reached target remote-cryptsetup.target. May 8 00:45:21.600250 systemd[1]: Reached target remote-fs.target. May 8 00:45:21.600260 systemd[1]: Reached target slices.target. May 8 00:45:21.600270 systemd[1]: Reached target swap.target. May 8 00:45:21.600280 systemd[1]: Reached target torcx.target. May 8 00:45:21.600290 systemd[1]: Reached target veritysetup.target. May 8 00:45:21.600301 systemd[1]: Listening on systemd-coredump.socket. May 8 00:45:21.600310 systemd[1]: Listening on systemd-initctl.socket. May 8 00:45:21.600320 kernel: kauditd_printk_skb: 48 callbacks suppressed May 8 00:45:21.600334 systemd[1]: Listening on systemd-journald-audit.socket. May 8 00:45:21.600349 kernel: audit: type=1400 audit(1746665121.418:86): avc: denied { audit_read } for pid=1 comm="systemd" capability=37 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 May 8 00:45:21.600360 kernel: audit: type=1335 audit(1746665121.418:87): pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 May 8 00:45:21.600370 systemd[1]: Listening on systemd-journald-dev-log.socket. May 8 00:45:21.600381 systemd[1]: Listening on systemd-journald.socket. May 8 00:45:21.600391 systemd[1]: Listening on systemd-networkd.socket. May 8 00:45:21.600402 systemd[1]: Listening on systemd-udevd-control.socket. May 8 00:45:21.600412 systemd[1]: Listening on systemd-udevd-kernel.socket. May 8 00:45:21.600423 systemd[1]: Listening on systemd-userdbd.socket. May 8 00:45:21.600437 systemd[1]: Mounting dev-hugepages.mount... May 8 00:45:21.600448 systemd[1]: Mounting dev-mqueue.mount... May 8 00:45:21.600458 systemd[1]: Mounting media.mount... May 8 00:45:21.600468 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 8 00:45:21.600479 systemd[1]: Mounting sys-kernel-debug.mount... May 8 00:45:21.600489 systemd[1]: Mounting sys-kernel-tracing.mount... May 8 00:45:21.600499 systemd[1]: Mounting tmp.mount... May 8 00:45:21.600510 systemd[1]: Starting flatcar-tmpfiles.service... May 8 00:45:21.600521 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 8 00:45:21.600536 systemd[1]: Starting kmod-static-nodes.service... May 8 00:45:21.600546 systemd[1]: Starting modprobe@configfs.service... May 8 00:45:21.600556 systemd[1]: Starting modprobe@dm_mod.service... May 8 00:45:21.600566 systemd[1]: Starting modprobe@drm.service... May 8 00:45:21.600580 systemd[1]: Starting modprobe@efi_pstore.service... May 8 00:45:21.600590 systemd[1]: Starting modprobe@fuse.service... May 8 00:45:21.600603 systemd[1]: Starting modprobe@loop.service... May 8 00:45:21.600618 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 8 00:45:21.600651 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. May 8 00:45:21.600667 systemd[1]: (This warning is only shown for the first unit using IP firewalling.) May 8 00:45:21.600678 systemd[1]: Starting systemd-journald.service... May 8 00:45:21.600688 kernel: fuse: init (API version 7.34) May 8 00:45:21.600698 systemd[1]: Starting systemd-modules-load.service... May 8 00:45:21.600707 kernel: loop: module loaded May 8 00:45:21.600718 systemd[1]: Starting systemd-network-generator.service... May 8 00:45:21.600728 systemd[1]: Starting systemd-remount-fs.service... May 8 00:45:21.600739 systemd[1]: Starting systemd-udev-trigger.service... May 8 00:45:21.600755 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 8 00:45:21.600766 systemd[1]: Mounted dev-hugepages.mount. May 8 00:45:21.600776 systemd[1]: Mounted dev-mqueue.mount. May 8 00:45:21.600786 systemd[1]: Mounted media.mount. May 8 00:45:21.600796 systemd[1]: Mounted sys-kernel-debug.mount. May 8 00:45:21.600806 systemd[1]: Mounted sys-kernel-tracing.mount. May 8 00:45:21.600824 systemd[1]: Mounted tmp.mount. May 8 00:45:21.600836 kernel: audit: type=1305 audit(1746665121.597:88): op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 May 8 00:45:21.600851 systemd-journald[1019]: Journal started May 8 00:45:21.600908 systemd-journald[1019]: Runtime Journal (/run/log/journal/31e67fdc4ddf4ee889a1c11818c6b221) is 6.0M, max 48.4M, 42.4M free. May 8 00:45:21.418000 audit[1]: AVC avc: denied { audit_read } for pid=1 comm="systemd" capability=37 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 May 8 00:45:21.418000 audit[1]: EVENT_LISTENER pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 May 8 00:45:21.597000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 May 8 00:45:21.597000 audit[1019]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=6 a1=7fff988a3c80 a2=4000 a3=7fff988a3d1c items=0 ppid=1 pid=1019 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:21.606034 kernel: audit: type=1300 audit(1746665121.597:88): arch=c000003e syscall=46 success=yes exit=60 a0=6 a1=7fff988a3c80 a2=4000 a3=7fff988a3d1c items=0 ppid=1 pid=1019 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:21.606073 systemd[1]: Started systemd-journald.service. May 8 00:45:21.606089 kernel: audit: type=1327 audit(1746665121.597:88): proctitle="/usr/lib/systemd/systemd-journald" May 8 00:45:21.597000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" May 8 00:45:21.608000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.610161 systemd[1]: Finished flatcar-tmpfiles.service. May 8 00:45:21.656922 kernel: audit: type=1130 audit(1746665121.608:89): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.657049 kernel: audit: type=1130 audit(1746665121.656:90): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.656000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.657961 systemd[1]: Finished kmod-static-nodes.service. May 8 00:45:21.661000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.662212 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 8 00:45:21.662363 systemd[1]: Finished modprobe@configfs.service. May 8 00:45:21.702119 kernel: audit: type=1130 audit(1746665121.661:91): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.702253 kernel: audit: type=1130 audit(1746665121.701:92): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.701000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.702343 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 8 00:45:21.702476 systemd[1]: Finished modprobe@dm_mod.service. May 8 00:45:21.701000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.781189 kernel: audit: type=1131 audit(1746665121.701:93): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.781000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.781000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.782884 systemd[1]: modprobe@drm.service: Deactivated successfully. May 8 00:45:21.783214 systemd[1]: Finished modprobe@drm.service. May 8 00:45:21.783000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.783000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.784731 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 8 00:45:21.784977 systemd[1]: Finished modprobe@efi_pstore.service. May 8 00:45:21.785000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.785000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.786568 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 8 00:45:21.786791 systemd[1]: Finished modprobe@fuse.service. May 8 00:45:21.786000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.786000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.788155 systemd[1]: modprobe@loop.service: Deactivated successfully. May 8 00:45:21.788385 systemd[1]: Finished modprobe@loop.service. May 8 00:45:21.788000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.788000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.789929 systemd[1]: Finished systemd-modules-load.service. May 8 00:45:21.790000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.791645 systemd[1]: Finished systemd-network-generator.service. May 8 00:45:21.791000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.793188 systemd[1]: Finished systemd-remount-fs.service. May 8 00:45:21.793000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.794959 systemd[1]: Finished systemd-udev-trigger.service. May 8 00:45:21.795000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:21.796483 systemd[1]: Reached target network-pre.target. May 8 00:45:21.799118 systemd[1]: Mounting sys-fs-fuse-connections.mount... May 8 00:45:21.801282 systemd[1]: Mounting sys-kernel-config.mount... May 8 00:45:21.854339 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 8 00:45:21.856537 systemd[1]: Starting systemd-hwdb-update.service... May 8 00:45:21.859477 systemd[1]: Starting systemd-journal-flush.service... May 8 00:45:21.861181 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 8 00:45:21.863014 systemd[1]: Starting systemd-random-seed.service... May 8 00:45:21.864253 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 8 00:45:21.866220 systemd-journald[1019]: Time spent on flushing to /var/log/journal/31e67fdc4ddf4ee889a1c11818c6b221 is 18.694ms for 1117 entries. May 8 00:45:21.866220 systemd-journald[1019]: System Journal (/var/log/journal/31e67fdc4ddf4ee889a1c11818c6b221) is 8.0M, max 195.6M, 187.6M free. May 8 00:45:21.983050 systemd-journald[1019]: Received client request to flush runtime journal. May 8 00:45:21.866139 systemd[1]: Starting systemd-sysctl.service... May 8 00:45:21.871263 systemd[1]: Starting systemd-sysusers.service... May 8 00:45:21.907074 systemd[1]: Starting systemd-udev-settle.service... May 8 00:45:21.910037 systemd[1]: Mounted sys-fs-fuse-connections.mount. May 8 00:45:21.983968 udevadm[1055]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. May 8 00:45:21.911127 systemd[1]: Mounted sys-kernel-config.mount. May 8 00:45:21.929849 systemd[1]: Finished systemd-sysusers.service. May 8 00:45:22.031000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:22.033244 systemd[1]: Finished systemd-journal-flush.service. May 8 00:45:22.033000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:22.034823 systemd[1]: Finished systemd-random-seed.service. May 8 00:45:22.034000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:22.036177 systemd[1]: Finished systemd-sysctl.service. May 8 00:45:22.036000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:22.037293 systemd[1]: Reached target first-boot-complete.target. May 8 00:45:22.039540 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... May 8 00:45:22.059228 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. May 8 00:45:22.060000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:22.711515 systemd[1]: Finished systemd-hwdb-update.service. May 8 00:45:22.711000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:22.713906 systemd[1]: Starting systemd-udevd.service... May 8 00:45:22.733253 systemd-udevd[1067]: Using default interface naming scheme 'v252'. May 8 00:45:22.752107 systemd[1]: Started systemd-udevd.service. May 8 00:45:22.752000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:22.756546 systemd[1]: Starting systemd-networkd.service... May 8 00:45:22.766734 systemd[1]: Starting systemd-userdbd.service... May 8 00:45:22.791611 systemd[1]: Found device dev-ttyS0.device. May 8 00:45:22.826000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:22.826566 systemd[1]: Started systemd-userdbd.service. May 8 00:45:22.858918 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 May 8 00:45:22.877921 kernel: ACPI: button: Power Button [PWRF] May 8 00:45:22.892309 systemd-networkd[1072]: lo: Link UP May 8 00:45:22.892326 systemd-networkd[1072]: lo: Gained carrier May 8 00:45:22.893016 systemd-networkd[1072]: Enumeration completed May 8 00:45:22.873000 audit[1068]: AVC avc: denied { confidentiality } for pid=1068 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 May 8 00:45:22.893241 systemd[1]: Started systemd-networkd.service. May 8 00:45:22.894000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:22.873000 audit[1068]: SYSCALL arch=c000003e syscall=175 success=yes exit=0 a0=55b3a3b54730 a1=338ac a2=7f2f0ed3fbc5 a3=5 items=110 ppid=1067 pid=1068 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="(udev-worker)" exe="/usr/bin/udevadm" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:22.873000 audit: CWD cwd="/" May 8 00:45:22.873000 audit: PATH item=0 name=(null) inode=1040 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=1 name=(null) inode=15584 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=2 name=(null) inode=15584 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=3 name=(null) inode=15585 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=4 name=(null) inode=15584 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=5 name=(null) inode=15586 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=6 name=(null) inode=15584 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=7 name=(null) inode=15587 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=8 name=(null) inode=15587 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=9 name=(null) inode=15588 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=10 name=(null) inode=15587 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=11 name=(null) inode=15589 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=12 name=(null) inode=15587 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=13 name=(null) inode=15590 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=14 name=(null) inode=15587 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=15 name=(null) inode=15591 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=16 name=(null) inode=15587 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=17 name=(null) inode=15592 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=18 name=(null) inode=15584 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=19 name=(null) inode=15593 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=20 name=(null) inode=15593 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=21 name=(null) inode=15594 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=22 name=(null) inode=15593 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=23 name=(null) inode=15595 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=24 name=(null) inode=15593 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=25 name=(null) inode=15596 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=26 name=(null) inode=15593 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=27 name=(null) inode=15597 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=28 name=(null) inode=15593 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=29 name=(null) inode=15598 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=30 name=(null) inode=15584 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=31 name=(null) inode=15599 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=32 name=(null) inode=15599 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=33 name=(null) inode=15600 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=34 name=(null) inode=15599 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=35 name=(null) inode=15601 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=36 name=(null) inode=15599 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=37 name=(null) inode=15602 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=38 name=(null) inode=15599 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=39 name=(null) inode=15603 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=40 name=(null) inode=15599 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=41 name=(null) inode=15604 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=42 name=(null) inode=15584 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=43 name=(null) inode=15605 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=44 name=(null) inode=15605 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=45 name=(null) inode=15606 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=46 name=(null) inode=15605 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=47 name=(null) inode=15607 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=48 name=(null) inode=15605 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=49 name=(null) inode=15608 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=50 name=(null) inode=15605 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=51 name=(null) inode=15609 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=52 name=(null) inode=15605 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=53 name=(null) inode=15610 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=54 name=(null) inode=1040 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=55 name=(null) inode=15611 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=56 name=(null) inode=15611 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=57 name=(null) inode=15612 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=58 name=(null) inode=15611 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=59 name=(null) inode=15613 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=60 name=(null) inode=15611 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=61 name=(null) inode=15614 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=62 name=(null) inode=15614 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=63 name=(null) inode=15615 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=64 name=(null) inode=15614 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=65 name=(null) inode=15616 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=66 name=(null) inode=15614 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=67 name=(null) inode=15617 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=68 name=(null) inode=15614 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=69 name=(null) inode=15618 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=70 name=(null) inode=15614 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=71 name=(null) inode=15619 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=72 name=(null) inode=15611 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=73 name=(null) inode=15620 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=74 name=(null) inode=15620 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=75 name=(null) inode=15621 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=76 name=(null) inode=15620 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=77 name=(null) inode=15622 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=78 name=(null) inode=15620 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=79 name=(null) inode=15623 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=80 name=(null) inode=15620 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=81 name=(null) inode=15624 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=82 name=(null) inode=15620 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=83 name=(null) inode=15625 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=84 name=(null) inode=15611 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=85 name=(null) inode=15626 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=86 name=(null) inode=15626 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=87 name=(null) inode=15627 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=88 name=(null) inode=15626 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=89 name=(null) inode=15628 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=90 name=(null) inode=15626 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=91 name=(null) inode=15629 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=92 name=(null) inode=15626 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=93 name=(null) inode=15630 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=94 name=(null) inode=15626 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=95 name=(null) inode=15631 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=96 name=(null) inode=15611 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=97 name=(null) inode=15632 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=98 name=(null) inode=15632 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=99 name=(null) inode=15633 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=100 name=(null) inode=15632 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=101 name=(null) inode=15634 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=102 name=(null) inode=15632 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=103 name=(null) inode=15635 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=104 name=(null) inode=15632 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=105 name=(null) inode=15636 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=106 name=(null) inode=15632 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=107 name=(null) inode=15637 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=108 name=(null) inode=1 dev=00:07 mode=040700 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PATH item=109 name=(null) inode=15638 dev=00:07 mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:45:22.873000 audit: PROCTITLE proctitle="(udev-worker)" May 8 00:45:22.900203 systemd-networkd[1072]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 8 00:45:22.901331 systemd-networkd[1072]: eth0: Link UP May 8 00:45:22.901336 systemd-networkd[1072]: eth0: Gained carrier May 8 00:45:22.908987 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. May 8 00:45:22.909362 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device May 8 00:45:22.923547 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt May 8 00:45:22.923678 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 May 8 00:45:22.923700 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) May 8 00:45:22.923831 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD May 8 00:45:22.923947 kernel: mousedev: PS/2 mouse device common for all mice May 8 00:45:22.915020 systemd-networkd[1072]: eth0: DHCPv4 address 10.0.0.82/16, gateway 10.0.0.1 acquired from 10.0.0.1 May 8 00:45:22.967258 kernel: kvm: Nested Virtualization enabled May 8 00:45:22.967356 kernel: SVM: kvm: Nested Paging enabled May 8 00:45:22.968649 kernel: SVM: Virtual VMLOAD VMSAVE supported May 8 00:45:22.968698 kernel: SVM: Virtual GIF supported May 8 00:45:23.004915 kernel: EDAC MC: Ver: 3.0.0 May 8 00:45:23.030429 systemd[1]: Finished systemd-udev-settle.service. May 8 00:45:23.030000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.032699 systemd[1]: Starting lvm2-activation-early.service... May 8 00:45:23.040521 lvm[1104]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 8 00:45:23.070933 systemd[1]: Finished lvm2-activation-early.service. May 8 00:45:23.080000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.081009 systemd[1]: Reached target cryptsetup.target. May 8 00:45:23.083024 systemd[1]: Starting lvm2-activation.service... May 8 00:45:23.092489 lvm[1106]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 8 00:45:23.122181 systemd[1]: Finished lvm2-activation.service. May 8 00:45:23.122000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.123448 systemd[1]: Reached target local-fs-pre.target. May 8 00:45:23.124510 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 8 00:45:23.124533 systemd[1]: Reached target local-fs.target. May 8 00:45:23.125502 systemd[1]: Reached target machines.target. May 8 00:45:23.128213 systemd[1]: Starting ldconfig.service... May 8 00:45:23.129464 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 8 00:45:23.129519 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 8 00:45:23.130853 systemd[1]: Starting systemd-boot-update.service... May 8 00:45:23.133282 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... May 8 00:45:23.136456 systemd[1]: Starting systemd-machine-id-commit.service... May 8 00:45:23.139011 systemd[1]: Starting systemd-sysext.service... May 8 00:45:23.141314 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1109 (bootctl) May 8 00:45:23.146155 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... May 8 00:45:23.147993 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. May 8 00:45:23.148000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.153326 systemd[1]: Unmounting usr-share-oem.mount... May 8 00:45:23.157686 systemd[1]: usr-share-oem.mount: Deactivated successfully. May 8 00:45:23.157990 systemd[1]: Unmounted usr-share-oem.mount. May 8 00:45:23.168961 kernel: loop0: detected capacity change from 0 to 210664 May 8 00:45:23.605911 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 8 00:45:23.627016 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 8 00:45:23.627742 systemd[1]: Finished systemd-machine-id-commit.service. May 8 00:45:23.626000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.630243 systemd-fsck[1119]: fsck.fat 4.2 (2021-01-31) May 8 00:45:23.630243 systemd-fsck[1119]: /dev/vda1: 791 files, 120730/258078 clusters May 8 00:45:23.632758 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. May 8 00:45:23.633000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.635651 systemd[1]: Mounting boot.mount... May 8 00:45:23.638946 kernel: loop1: detected capacity change from 0 to 210664 May 8 00:45:23.644739 systemd[1]: Mounted boot.mount. May 8 00:45:23.646675 (sd-sysext)[1128]: Using extensions 'kubernetes'. May 8 00:45:23.647112 (sd-sysext)[1128]: Merged extensions into '/usr'. May 8 00:45:23.657187 systemd[1]: Finished systemd-boot-update.service. May 8 00:45:23.657000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.670545 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 8 00:45:23.672468 systemd[1]: Mounting usr-share-oem.mount... May 8 00:45:23.673622 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 8 00:45:23.674985 systemd[1]: Starting modprobe@dm_mod.service... May 8 00:45:23.677238 systemd[1]: Starting modprobe@efi_pstore.service... May 8 00:45:23.679277 systemd[1]: Starting modprobe@loop.service... May 8 00:45:23.680265 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 8 00:45:23.680403 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 8 00:45:23.680523 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 8 00:45:23.683927 systemd[1]: Mounted usr-share-oem.mount. May 8 00:45:23.685413 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 8 00:45:23.685661 systemd[1]: Finished modprobe@dm_mod.service. May 8 00:45:23.686000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.686000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.687535 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 8 00:45:23.687790 systemd[1]: Finished modprobe@efi_pstore.service. May 8 00:45:23.688000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.688000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.689566 systemd[1]: modprobe@loop.service: Deactivated successfully. May 8 00:45:23.689745 systemd[1]: Finished modprobe@loop.service. May 8 00:45:23.690000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.690000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.691680 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 8 00:45:23.691827 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 8 00:45:23.693162 systemd[1]: Finished systemd-sysext.service. May 8 00:45:23.693000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.696128 systemd[1]: Starting ensure-sysext.service... May 8 00:45:23.698288 systemd[1]: Starting systemd-tmpfiles-setup.service... May 8 00:45:23.703465 systemd[1]: Reloading. May 8 00:45:23.710240 systemd-tmpfiles[1145]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. May 8 00:45:23.712255 systemd-tmpfiles[1145]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 8 00:45:23.714439 systemd-tmpfiles[1145]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 8 00:45:23.726129 ldconfig[1108]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 8 00:45:23.759636 /usr/lib/systemd/system-generators/torcx-generator[1166]: time="2025-05-08T00:45:23Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 8 00:45:23.760056 /usr/lib/systemd/system-generators/torcx-generator[1166]: time="2025-05-08T00:45:23Z" level=info msg="torcx already run" May 8 00:45:23.829568 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 8 00:45:23.829589 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 8 00:45:23.849595 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 8 00:45:23.901497 systemd[1]: Finished ldconfig.service. May 8 00:45:23.902000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ldconfig comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.902734 systemd[1]: Finished systemd-tmpfiles-setup.service. May 8 00:45:23.902000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.907016 systemd[1]: Starting audit-rules.service... May 8 00:45:23.909301 systemd[1]: Starting clean-ca-certificates.service... May 8 00:45:23.911833 systemd[1]: Starting systemd-journal-catalog-update.service... May 8 00:45:23.914862 systemd[1]: Starting systemd-resolved.service... May 8 00:45:23.917656 systemd[1]: Starting systemd-timesyncd.service... May 8 00:45:23.919990 systemd[1]: Starting systemd-update-utmp.service... May 8 00:45:23.921729 systemd[1]: Finished clean-ca-certificates.service. May 8 00:45:23.922000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.926319 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 8 00:45:23.926000 audit[1227]: SYSTEM_BOOT pid=1227 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' May 8 00:45:23.932055 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 8 00:45:23.933851 systemd[1]: Starting modprobe@dm_mod.service... May 8 00:45:23.936083 systemd[1]: Starting modprobe@efi_pstore.service... May 8 00:45:23.938448 systemd[1]: Starting modprobe@loop.service... May 8 00:45:23.939372 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 8 00:45:23.939554 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 8 00:45:23.939726 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 8 00:45:23.941014 systemd[1]: Finished systemd-journal-catalog-update.service. May 8 00:45:23.941000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.943000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.943000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.942716 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 8 00:45:23.944000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.944000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.946000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.946000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.942942 systemd[1]: Finished modprobe@dm_mod.service. May 8 00:45:23.944512 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 8 00:45:23.944691 systemd[1]: Finished modprobe@efi_pstore.service. May 8 00:45:23.946460 systemd[1]: modprobe@loop.service: Deactivated successfully. May 8 00:45:23.946665 systemd[1]: Finished modprobe@loop.service. May 8 00:45:23.948054 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 8 00:45:23.948210 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 8 00:45:23.950327 systemd[1]: Starting systemd-update-done.service... May 8 00:45:23.953117 systemd[1]: Finished systemd-update-utmp.service. May 8 00:45:23.953000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.958125 systemd[1]: Finished systemd-update-done.service. May 8 00:45:23.958000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-done comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:23.960054 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 8 00:45:23.962273 systemd[1]: Starting modprobe@dm_mod.service... May 8 00:45:23.964808 systemd[1]: Starting modprobe@efi_pstore.service... May 8 00:45:23.967310 augenrules[1248]: No rules May 8 00:45:23.967225 systemd[1]: Starting modprobe@loop.service... May 8 00:45:23.968100 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 8 00:45:23.968228 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 8 00:45:23.968387 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 8 00:45:23.969587 systemd[1]: Finished audit-rules.service. May 8 00:45:23.964000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 May 8 00:45:23.964000 audit[1248]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffd64ae0aa0 a2=420 a3=0 items=0 ppid=1215 pid=1248 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:23.964000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 May 8 00:45:23.970914 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 8 00:45:23.971102 systemd[1]: Finished modprobe@dm_mod.service. May 8 00:45:23.972481 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 8 00:45:23.972627 systemd[1]: Finished modprobe@efi_pstore.service. May 8 00:45:23.974060 systemd[1]: modprobe@loop.service: Deactivated successfully. May 8 00:45:23.974243 systemd[1]: Finished modprobe@loop.service. May 8 00:45:23.978486 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 8 00:45:23.978574 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 8 00:45:23.981010 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 8 00:45:23.982203 systemd[1]: Starting modprobe@dm_mod.service... May 8 00:45:23.983983 systemd[1]: Starting modprobe@drm.service... May 8 00:45:23.986464 systemd[1]: Starting modprobe@efi_pstore.service... May 8 00:45:23.988883 systemd[1]: Starting modprobe@loop.service... May 8 00:45:23.990096 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 8 00:45:23.990215 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 8 00:45:23.992224 systemd[1]: Starting systemd-networkd-wait-online.service... May 8 00:45:23.994999 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 8 00:45:23.996350 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 8 00:45:23.996518 systemd[1]: Finished modprobe@dm_mod.service. May 8 00:45:23.997965 systemd[1]: modprobe@drm.service: Deactivated successfully. May 8 00:45:23.998125 systemd[1]: Finished modprobe@drm.service. May 8 00:45:23.999640 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 8 00:45:23.999782 systemd[1]: Finished modprobe@efi_pstore.service. May 8 00:45:24.001098 systemd[1]: modprobe@loop.service: Deactivated successfully. May 8 00:45:24.001261 systemd[1]: Finished modprobe@loop.service. May 8 00:45:24.002906 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 8 00:45:24.002998 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 8 00:45:24.003997 systemd[1]: Finished ensure-sysext.service. May 8 00:45:24.009138 systemd[1]: Started systemd-timesyncd.service. May 8 00:45:23.592637 systemd-timesyncd[1222]: Contacted time server 10.0.0.1:123 (10.0.0.1). May 8 00:45:23.618320 systemd-journald[1019]: Time jumped backwards, rotating. May 8 00:45:23.592699 systemd-timesyncd[1222]: Initial clock synchronization to Thu 2025-05-08 00:45:23.592563 UTC. May 8 00:45:23.593041 systemd[1]: Reached target time-set.target. May 8 00:45:23.618364 systemd-resolved[1221]: Positive Trust Anchors: May 8 00:45:23.618373 systemd-resolved[1221]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 8 00:45:23.618400 systemd-resolved[1221]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test May 8 00:45:23.624987 systemd-resolved[1221]: Defaulting to hostname 'linux'. May 8 00:45:23.626445 systemd[1]: Started systemd-resolved.service. May 8 00:45:23.627516 systemd[1]: Reached target network.target. May 8 00:45:23.628363 systemd[1]: Reached target nss-lookup.target. May 8 00:45:23.629240 systemd[1]: Reached target sysinit.target. May 8 00:45:23.630205 systemd[1]: Started motdgen.path. May 8 00:45:23.630986 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. May 8 00:45:23.632379 systemd[1]: Started logrotate.timer. May 8 00:45:23.633295 systemd[1]: Started mdadm.timer. May 8 00:45:23.634089 systemd[1]: Started systemd-tmpfiles-clean.timer. May 8 00:45:23.635064 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 8 00:45:23.635093 systemd[1]: Reached target paths.target. May 8 00:45:23.635912 systemd[1]: Reached target timers.target. May 8 00:45:23.637124 systemd[1]: Listening on dbus.socket. May 8 00:45:23.639119 systemd[1]: Starting docker.socket... May 8 00:45:23.641074 systemd[1]: Listening on sshd.socket. May 8 00:45:23.641998 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 8 00:45:23.642354 systemd[1]: Listening on docker.socket. May 8 00:45:23.643145 systemd[1]: Reached target sockets.target. May 8 00:45:23.643948 systemd[1]: Reached target basic.target. May 8 00:45:23.644835 systemd[1]: System is tainted: cgroupsv1 May 8 00:45:23.644876 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. May 8 00:45:23.644894 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. May 8 00:45:23.645910 systemd[1]: Starting containerd.service... May 8 00:45:23.647717 systemd[1]: Starting dbus.service... May 8 00:45:23.649984 systemd[1]: Starting enable-oem-cloudinit.service... May 8 00:45:23.652058 systemd[1]: Starting extend-filesystems.service... May 8 00:45:23.653827 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). May 8 00:45:23.655914 jq[1279]: false May 8 00:45:23.655595 systemd[1]: Starting motdgen.service... May 8 00:45:23.657853 systemd[1]: Starting prepare-helm.service... May 8 00:45:23.663225 extend-filesystems[1280]: Found loop1 May 8 00:45:23.663225 extend-filesystems[1280]: Found sr0 May 8 00:45:23.663225 extend-filesystems[1280]: Found vda May 8 00:45:23.663225 extend-filesystems[1280]: Found vda1 May 8 00:45:23.663225 extend-filesystems[1280]: Found vda2 May 8 00:45:23.663225 extend-filesystems[1280]: Found vda3 May 8 00:45:23.663225 extend-filesystems[1280]: Found usr May 8 00:45:23.663225 extend-filesystems[1280]: Found vda4 May 8 00:45:23.663225 extend-filesystems[1280]: Found vda6 May 8 00:45:23.663225 extend-filesystems[1280]: Found vda7 May 8 00:45:23.663225 extend-filesystems[1280]: Found vda9 May 8 00:45:23.663225 extend-filesystems[1280]: Checking size of /dev/vda9 May 8 00:45:23.660426 systemd[1]: Starting ssh-key-proc-cmdline.service... May 8 00:45:23.714951 extend-filesystems[1280]: Resized partition /dev/vda9 May 8 00:45:23.669447 dbus-daemon[1277]: [system] SELinux support is enabled May 8 00:45:23.663317 systemd[1]: Starting sshd-keygen.service... May 8 00:45:23.719440 extend-filesystems[1322]: resize2fs 1.46.5 (30-Dec-2021) May 8 00:45:23.723040 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks May 8 00:45:23.668470 systemd[1]: Starting systemd-logind.service... May 8 00:45:23.669537 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 8 00:45:23.669637 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 8 00:45:23.671460 systemd[1]: Starting update-engine.service... May 8 00:45:23.726324 jq[1302]: true May 8 00:45:23.675421 systemd[1]: Starting update-ssh-keys-after-ignition.service... May 8 00:45:23.677727 systemd[1]: Started dbus.service. May 8 00:45:23.728000 tar[1306]: linux-amd64/helm May 8 00:45:23.683598 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 8 00:45:23.728370 jq[1310]: true May 8 00:45:23.683956 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. May 8 00:45:23.728563 env[1311]: time="2025-05-08T00:45:23.727725385Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 May 8 00:45:23.684869 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 8 00:45:23.685152 systemd[1]: Finished ssh-key-proc-cmdline.service. May 8 00:45:23.687619 systemd[1]: motdgen.service: Deactivated successfully. May 8 00:45:23.687916 systemd[1]: Finished motdgen.service. May 8 00:45:23.691005 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 8 00:45:23.691061 systemd[1]: Reached target system-config.target. May 8 00:45:23.692630 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 8 00:45:23.692656 systemd[1]: Reached target user-config.target. May 8 00:45:23.735493 update_engine[1297]: I0508 00:45:23.734468 1297 main.cc:92] Flatcar Update Engine starting May 8 00:45:23.737072 update_engine[1297]: I0508 00:45:23.736774 1297 update_check_scheduler.cc:74] Next update check in 4m15s May 8 00:45:23.736682 systemd[1]: Started update-engine.service. May 8 00:45:23.739644 systemd[1]: Started locksmithd.service. May 8 00:45:23.742975 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 8 00:45:23.743012 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 8 00:45:23.752303 env[1311]: time="2025-05-08T00:45:23.752244906Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 May 8 00:45:23.752662 env[1311]: time="2025-05-08T00:45:23.752640228Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 May 8 00:45:23.753448 systemd-logind[1291]: Watching system buttons on /dev/input/event1 (Power Button) May 8 00:45:23.753473 systemd-logind[1291]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) May 8 00:45:23.754045 systemd-logind[1291]: New seat seat0. May 8 00:45:23.755499 env[1311]: time="2025-05-08T00:45:23.754413945Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.180-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 May 8 00:45:23.755499 env[1311]: time="2025-05-08T00:45:23.754442629Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 May 8 00:45:23.756018 env[1311]: time="2025-05-08T00:45:23.755997635Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 May 8 00:45:23.756514 env[1311]: time="2025-05-08T00:45:23.756488236Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 May 8 00:45:23.756612 env[1311]: time="2025-05-08T00:45:23.756585197Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" May 8 00:45:23.756685 systemd[1]: Started systemd-logind.service. May 8 00:45:23.756900 env[1311]: time="2025-05-08T00:45:23.756674385Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 May 8 00:45:23.756900 env[1311]: time="2025-05-08T00:45:23.756783680Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 May 8 00:45:23.757162 env[1311]: time="2025-05-08T00:45:23.757101225Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 May 8 00:45:23.757481 env[1311]: time="2025-05-08T00:45:23.757449779Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 May 8 00:45:23.757582 env[1311]: time="2025-05-08T00:45:23.757556940Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 May 8 00:45:23.757727 env[1311]: time="2025-05-08T00:45:23.757700890Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" May 8 00:45:23.757818 env[1311]: time="2025-05-08T00:45:23.757798674Z" level=info msg="metadata content store policy set" policy=shared May 8 00:45:23.764150 kernel: EXT4-fs (vda9): resized filesystem to 1864699 May 8 00:45:23.793180 extend-filesystems[1322]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required May 8 00:45:23.793180 extend-filesystems[1322]: old_desc_blocks = 1, new_desc_blocks = 1 May 8 00:45:23.793180 extend-filesystems[1322]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. May 8 00:45:23.813128 extend-filesystems[1280]: Resized filesystem in /dev/vda9 May 8 00:45:23.814089 env[1311]: time="2025-05-08T00:45:23.808921936Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 May 8 00:45:23.814089 env[1311]: time="2025-05-08T00:45:23.808994472Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 May 8 00:45:23.814089 env[1311]: time="2025-05-08T00:45:23.809015021Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 May 8 00:45:23.814089 env[1311]: time="2025-05-08T00:45:23.809076666Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 May 8 00:45:23.814089 env[1311]: time="2025-05-08T00:45:23.809092677Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 May 8 00:45:23.814089 env[1311]: time="2025-05-08T00:45:23.809105120Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 May 8 00:45:23.814089 env[1311]: time="2025-05-08T00:45:23.809133132Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 May 8 00:45:23.814089 env[1311]: time="2025-05-08T00:45:23.809146608Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 May 8 00:45:23.814089 env[1311]: time="2025-05-08T00:45:23.809158089Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 May 8 00:45:23.814089 env[1311]: time="2025-05-08T00:45:23.809170813Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 May 8 00:45:23.814089 env[1311]: time="2025-05-08T00:45:23.809182635Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 May 8 00:45:23.814089 env[1311]: time="2025-05-08T00:45:23.809199166Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 May 8 00:45:23.814089 env[1311]: time="2025-05-08T00:45:23.809336023Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 May 8 00:45:23.814089 env[1311]: time="2025-05-08T00:45:23.809415783Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 May 8 00:45:23.793555 systemd[1]: extend-filesystems.service: Deactivated successfully. May 8 00:45:23.814533 bash[1338]: Updated "/home/core/.ssh/authorized_keys" May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.809876226Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.809924597Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.809945937Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.810029263Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.810051295Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.810073226Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.810093374Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.810133108Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.810154398Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.810171069Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.810189273Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.810206175Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.810329436Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.810343332Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 May 8 00:45:23.814644 env[1311]: time="2025-05-08T00:45:23.810355195Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 May 8 00:45:23.793864 systemd[1]: Finished extend-filesystems.service. May 8 00:45:23.815006 env[1311]: time="2025-05-08T00:45:23.810365254Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 May 8 00:45:23.815006 env[1311]: time="2025-05-08T00:45:23.810378839Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 May 8 00:45:23.815006 env[1311]: time="2025-05-08T00:45:23.810389168Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 May 8 00:45:23.815006 env[1311]: time="2025-05-08T00:45:23.810420467Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" May 8 00:45:23.815006 env[1311]: time="2025-05-08T00:45:23.810460713Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 May 8 00:45:23.806887 systemd[1]: Finished update-ssh-keys-after-ignition.service. May 8 00:45:23.813626 systemd[1]: Started containerd.service. May 8 00:45:23.815189 env[1311]: time="2025-05-08T00:45:23.810760735Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" May 8 00:45:23.815189 env[1311]: time="2025-05-08T00:45:23.810840846Z" level=info msg="Connect containerd service" May 8 00:45:23.815189 env[1311]: time="2025-05-08T00:45:23.810899796Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" May 8 00:45:23.815189 env[1311]: time="2025-05-08T00:45:23.811834529Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 8 00:45:23.815189 env[1311]: time="2025-05-08T00:45:23.812341280Z" level=info msg="Start subscribing containerd event" May 8 00:45:23.815189 env[1311]: time="2025-05-08T00:45:23.812423685Z" level=info msg="Start recovering state" May 8 00:45:23.815189 env[1311]: time="2025-05-08T00:45:23.812536476Z" level=info msg="Start event monitor" May 8 00:45:23.815189 env[1311]: time="2025-05-08T00:45:23.812571903Z" level=info msg="Start snapshots syncer" May 8 00:45:23.815189 env[1311]: time="2025-05-08T00:45:23.812590998Z" level=info msg="Start cni network conf syncer for default" May 8 00:45:23.815189 env[1311]: time="2025-05-08T00:45:23.812601969Z" level=info msg="Start streaming server" May 8 00:45:23.815189 env[1311]: time="2025-05-08T00:45:23.813299237Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 8 00:45:23.815189 env[1311]: time="2025-05-08T00:45:23.813340274Z" level=info msg=serving... address=/run/containerd/containerd.sock May 8 00:45:23.820691 env[1311]: time="2025-05-08T00:45:23.820655685Z" level=info msg="containerd successfully booted in 0.093653s" May 8 00:45:23.829064 locksmithd[1339]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 8 00:45:24.166819 tar[1306]: linux-amd64/LICENSE May 8 00:45:24.166973 tar[1306]: linux-amd64/README.md May 8 00:45:24.171234 systemd[1]: Finished prepare-helm.service. May 8 00:45:24.246375 systemd-networkd[1072]: eth0: Gained IPv6LL May 8 00:45:24.248667 systemd[1]: Finished systemd-networkd-wait-online.service. May 8 00:45:24.250474 systemd[1]: Reached target network-online.target. May 8 00:45:24.253320 systemd[1]: Starting kubelet.service... May 8 00:45:25.226377 sshd_keygen[1303]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 8 00:45:25.248625 systemd[1]: Finished sshd-keygen.service. May 8 00:45:25.252077 systemd[1]: Starting issuegen.service... May 8 00:45:25.259124 systemd[1]: issuegen.service: Deactivated successfully. May 8 00:45:25.259386 systemd[1]: Finished issuegen.service. May 8 00:45:25.262289 systemd[1]: Starting systemd-user-sessions.service... May 8 00:45:25.269851 systemd[1]: Finished systemd-user-sessions.service. May 8 00:45:25.272736 systemd[1]: Started getty@tty1.service. May 8 00:45:25.276899 systemd[1]: Started serial-getty@ttyS0.service. May 8 00:45:25.278472 systemd[1]: Reached target getty.target. May 8 00:45:25.616135 systemd[1]: Started kubelet.service. May 8 00:45:25.618087 systemd[1]: Reached target multi-user.target. May 8 00:45:25.621100 systemd[1]: Starting systemd-update-utmp-runlevel.service... May 8 00:45:25.629618 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. May 8 00:45:25.629907 systemd[1]: Finished systemd-update-utmp-runlevel.service. May 8 00:45:25.632316 systemd[1]: Startup finished in 6.779s (kernel) + 9.375s (userspace) = 16.155s. May 8 00:45:26.473920 kubelet[1380]: E0508 00:45:26.473826 1380 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 00:45:26.476254 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 00:45:26.476458 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 00:45:33.224654 systemd[1]: Created slice system-sshd.slice. May 8 00:45:33.225837 systemd[1]: Started sshd@0-10.0.0.82:22-10.0.0.1:56178.service. May 8 00:45:33.270912 sshd[1391]: Accepted publickey for core from 10.0.0.1 port 56178 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:45:33.272671 sshd[1391]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:45:33.283553 systemd-logind[1291]: New session 1 of user core. May 8 00:45:33.284666 systemd[1]: Created slice user-500.slice. May 8 00:45:33.285834 systemd[1]: Starting user-runtime-dir@500.service... May 8 00:45:33.295681 systemd[1]: Finished user-runtime-dir@500.service. May 8 00:45:33.297556 systemd[1]: Starting user@500.service... May 8 00:45:33.300792 (systemd)[1396]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 8 00:45:33.372827 systemd[1396]: Queued start job for default target default.target. May 8 00:45:33.373050 systemd[1396]: Reached target paths.target. May 8 00:45:33.373065 systemd[1396]: Reached target sockets.target. May 8 00:45:33.373077 systemd[1396]: Reached target timers.target. May 8 00:45:33.373088 systemd[1396]: Reached target basic.target. May 8 00:45:33.373139 systemd[1396]: Reached target default.target. May 8 00:45:33.373163 systemd[1396]: Startup finished in 66ms. May 8 00:45:33.373302 systemd[1]: Started user@500.service. May 8 00:45:33.374332 systemd[1]: Started session-1.scope. May 8 00:45:33.424645 systemd[1]: Started sshd@1-10.0.0.82:22-10.0.0.1:56194.service. May 8 00:45:33.466489 sshd[1405]: Accepted publickey for core from 10.0.0.1 port 56194 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:45:33.467885 sshd[1405]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:45:33.471956 systemd-logind[1291]: New session 2 of user core. May 8 00:45:33.472743 systemd[1]: Started session-2.scope. May 8 00:45:33.529248 sshd[1405]: pam_unix(sshd:session): session closed for user core May 8 00:45:33.532931 systemd[1]: Started sshd@2-10.0.0.82:22-10.0.0.1:56200.service. May 8 00:45:33.533541 systemd[1]: sshd@1-10.0.0.82:22-10.0.0.1:56194.service: Deactivated successfully. May 8 00:45:33.534652 systemd[1]: session-2.scope: Deactivated successfully. May 8 00:45:33.535219 systemd-logind[1291]: Session 2 logged out. Waiting for processes to exit. May 8 00:45:33.536210 systemd-logind[1291]: Removed session 2. May 8 00:45:33.577030 sshd[1411]: Accepted publickey for core from 10.0.0.1 port 56200 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:45:33.578421 sshd[1411]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:45:33.582060 systemd-logind[1291]: New session 3 of user core. May 8 00:45:33.582973 systemd[1]: Started session-3.scope. May 8 00:45:33.633350 sshd[1411]: pam_unix(sshd:session): session closed for user core May 8 00:45:33.636049 systemd[1]: Started sshd@3-10.0.0.82:22-10.0.0.1:56210.service. May 8 00:45:33.636539 systemd[1]: sshd@2-10.0.0.82:22-10.0.0.1:56200.service: Deactivated successfully. May 8 00:45:33.638426 systemd-logind[1291]: Session 3 logged out. Waiting for processes to exit. May 8 00:45:33.638489 systemd[1]: session-3.scope: Deactivated successfully. May 8 00:45:33.639505 systemd-logind[1291]: Removed session 3. May 8 00:45:33.678089 sshd[1418]: Accepted publickey for core from 10.0.0.1 port 56210 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:45:33.679536 sshd[1418]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:45:33.683160 systemd-logind[1291]: New session 4 of user core. May 8 00:45:33.683868 systemd[1]: Started session-4.scope. May 8 00:45:33.738805 sshd[1418]: pam_unix(sshd:session): session closed for user core May 8 00:45:33.741419 systemd[1]: Started sshd@4-10.0.0.82:22-10.0.0.1:56224.service. May 8 00:45:33.742041 systemd[1]: sshd@3-10.0.0.82:22-10.0.0.1:56210.service: Deactivated successfully. May 8 00:45:33.742758 systemd-logind[1291]: Session 4 logged out. Waiting for processes to exit. May 8 00:45:33.742817 systemd[1]: session-4.scope: Deactivated successfully. May 8 00:45:33.743927 systemd-logind[1291]: Removed session 4. May 8 00:45:33.783905 sshd[1424]: Accepted publickey for core from 10.0.0.1 port 56224 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:45:33.785387 sshd[1424]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:45:33.789724 systemd-logind[1291]: New session 5 of user core. May 8 00:45:33.790476 systemd[1]: Started session-5.scope. May 8 00:45:33.848299 sudo[1430]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 May 8 00:45:33.848579 sudo[1430]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) May 8 00:45:33.859393 dbus-daemon[1277]: н\u0017G\xd6U: received setenforce notice (enforcing=-598081792) May 8 00:45:33.861671 sudo[1430]: pam_unix(sudo:session): session closed for user root May 8 00:45:33.863850 sshd[1424]: pam_unix(sshd:session): session closed for user core May 8 00:45:33.866803 systemd[1]: Started sshd@5-10.0.0.82:22-10.0.0.1:56236.service. May 8 00:45:33.867297 systemd[1]: sshd@4-10.0.0.82:22-10.0.0.1:56224.service: Deactivated successfully. May 8 00:45:33.868405 systemd[1]: session-5.scope: Deactivated successfully. May 8 00:45:33.868478 systemd-logind[1291]: Session 5 logged out. Waiting for processes to exit. May 8 00:45:33.869543 systemd-logind[1291]: Removed session 5. May 8 00:45:33.974648 sshd[1432]: Accepted publickey for core from 10.0.0.1 port 56236 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:45:33.976125 sshd[1432]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:45:33.979881 systemd-logind[1291]: New session 6 of user core. May 8 00:45:33.980802 systemd[1]: Started session-6.scope. May 8 00:45:34.035801 sudo[1439]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules May 8 00:45:34.036100 sudo[1439]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) May 8 00:45:34.040280 sudo[1439]: pam_unix(sudo:session): session closed for user root May 8 00:45:34.045562 sudo[1438]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules May 8 00:45:34.045823 sudo[1438]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) May 8 00:45:34.055672 systemd[1]: Stopping audit-rules.service... May 8 00:45:34.056000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 May 8 00:45:34.057657 auditctl[1442]: No rules May 8 00:45:34.062561 kernel: kauditd_printk_skb: 167 callbacks suppressed May 8 00:45:34.062714 kernel: audit: type=1305 audit(1746665134.056:146): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=remove_rule key=(null) list=5 res=1 May 8 00:45:34.062788 systemd[1]: audit-rules.service: Deactivated successfully. May 8 00:45:34.063080 systemd[1]: Stopped audit-rules.service. May 8 00:45:34.056000 audit[1442]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffe73f27b20 a2=420 a3=0 items=0 ppid=1 pid=1442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:34.065014 systemd[1]: Starting audit-rules.service... May 8 00:45:34.070460 kernel: audit: type=1300 audit(1746665134.056:146): arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffe73f27b20 a2=420 a3=0 items=0 ppid=1 pid=1442 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:34.070556 kernel: audit: type=1327 audit(1746665134.056:146): proctitle=2F7362696E2F617564697463746C002D44 May 8 00:45:34.056000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D44 May 8 00:45:34.062000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:34.075900 kernel: audit: type=1131 audit(1746665134.062:147): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:34.091211 augenrules[1460]: No rules May 8 00:45:34.092004 systemd[1]: Finished audit-rules.service. May 8 00:45:34.091000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:34.094335 sudo[1438]: pam_unix(sudo:session): session closed for user root May 8 00:45:34.093000 audit[1438]: USER_END pid=1438 uid=500 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 8 00:45:34.096348 sshd[1432]: pam_unix(sshd:session): session closed for user core May 8 00:45:34.099528 systemd[1]: Started sshd@6-10.0.0.82:22-10.0.0.1:56242.service. May 8 00:45:34.100796 systemd[1]: sshd@5-10.0.0.82:22-10.0.0.1:56236.service: Deactivated successfully. May 8 00:45:34.102151 systemd[1]: session-6.scope: Deactivated successfully. May 8 00:45:34.102751 systemd-logind[1291]: Session 6 logged out. Waiting for processes to exit. May 8 00:45:34.103610 systemd-logind[1291]: Removed session 6. May 8 00:45:34.106046 kernel: audit: type=1130 audit(1746665134.091:148): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=audit-rules comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:34.106270 kernel: audit: type=1106 audit(1746665134.093:149): pid=1438 uid=500 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 8 00:45:34.106299 kernel: audit: type=1104 audit(1746665134.093:150): pid=1438 uid=500 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 8 00:45:34.093000 audit[1438]: CRED_DISP pid=1438 uid=500 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 8 00:45:34.109921 kernel: audit: type=1106 audit(1746665134.098:151): pid=1432 uid=0 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:45:34.098000 audit[1432]: USER_END pid=1432 uid=0 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:45:34.098000 audit[1432]: CRED_DISP pid=1432 uid=0 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:45:34.146171 kernel: audit: type=1104 audit(1746665134.098:152): pid=1432 uid=0 auid=500 ses=6 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:45:34.146352 kernel: audit: type=1130 audit(1746665134.098:153): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-10.0.0.82:22-10.0.0.1:56242 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:34.098000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-10.0.0.82:22-10.0.0.1:56242 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:34.099000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@5-10.0.0.82:22-10.0.0.1:56236 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:34.158000 audit[1465]: USER_ACCT pid=1465 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:45:34.160346 sshd[1465]: Accepted publickey for core from 10.0.0.1 port 56242 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:45:34.160000 audit[1465]: CRED_ACQ pid=1465 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:45:34.160000 audit[1465]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffcf1761e20 a2=3 a3=0 items=0 ppid=1 pid=1465 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=7 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:34.160000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:45:34.161817 sshd[1465]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:45:34.165823 systemd-logind[1291]: New session 7 of user core. May 8 00:45:34.166598 systemd[1]: Started session-7.scope. May 8 00:45:34.169000 audit[1465]: USER_START pid=1465 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:45:34.171000 audit[1470]: CRED_ACQ pid=1470 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:45:34.221000 audit[1471]: USER_ACCT pid=1471 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 8 00:45:34.221000 audit[1471]: CRED_REFR pid=1471 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 8 00:45:34.222307 sudo[1471]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 8 00:45:34.222520 sudo[1471]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) May 8 00:45:34.223000 audit[1471]: USER_START pid=1471 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 8 00:45:34.273491 systemd[1]: Starting docker.service... May 8 00:45:34.342371 env[1483]: time="2025-05-08T00:45:34.342211537Z" level=info msg="Starting up" May 8 00:45:34.347125 env[1483]: time="2025-05-08T00:45:34.347076382Z" level=info msg="parsed scheme: \"unix\"" module=grpc May 8 00:45:34.347125 env[1483]: time="2025-05-08T00:45:34.347119222Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc May 8 00:45:34.347213 env[1483]: time="2025-05-08T00:45:34.347142546Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc May 8 00:45:34.347213 env[1483]: time="2025-05-08T00:45:34.347153266Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc May 8 00:45:34.349303 env[1483]: time="2025-05-08T00:45:34.349278883Z" level=info msg="parsed scheme: \"unix\"" module=grpc May 8 00:45:34.349303 env[1483]: time="2025-05-08T00:45:34.349300203Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc May 8 00:45:34.349388 env[1483]: time="2025-05-08T00:45:34.349319399Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc May 8 00:45:34.349388 env[1483]: time="2025-05-08T00:45:34.349328646Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc May 8 00:45:35.901137 env[1483]: time="2025-05-08T00:45:35.901054781Z" level=warning msg="Your kernel does not support cgroup blkio weight" May 8 00:45:35.901137 env[1483]: time="2025-05-08T00:45:35.901100677Z" level=warning msg="Your kernel does not support cgroup blkio weight_device" May 8 00:45:35.901666 env[1483]: time="2025-05-08T00:45:35.901414596Z" level=info msg="Loading containers: start." May 8 00:45:36.006000 audit[1517]: NETFILTER_CFG table=nat:2 family=2 entries=2 op=nft_register_chain pid=1517 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.006000 audit[1517]: SYSCALL arch=c000003e syscall=46 success=yes exit=116 a0=3 a1=7ffc34f7e8d0 a2=0 a3=7ffc34f7e8bc items=0 ppid=1483 pid=1517 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.006000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D4E00444F434B4552 May 8 00:45:36.008000 audit[1519]: NETFILTER_CFG table=filter:3 family=2 entries=2 op=nft_register_chain pid=1519 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.008000 audit[1519]: SYSCALL arch=c000003e syscall=46 success=yes exit=124 a0=3 a1=7ffdaff5d560 a2=0 a3=7ffdaff5d54c items=0 ppid=1483 pid=1519 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.008000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B4552 May 8 00:45:36.010000 audit[1521]: NETFILTER_CFG table=filter:4 family=2 entries=1 op=nft_register_chain pid=1521 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.010000 audit[1521]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7fff9901da60 a2=0 a3=7fff9901da4c items=0 ppid=1483 pid=1521 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.010000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D31 May 8 00:45:36.012000 audit[1523]: NETFILTER_CFG table=filter:5 family=2 entries=1 op=nft_register_chain pid=1523 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.012000 audit[1523]: SYSCALL arch=c000003e syscall=46 success=yes exit=112 a0=3 a1=7fffc5c34bc0 a2=0 a3=7fffc5c34bac items=0 ppid=1483 pid=1523 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.012000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D49534F4C4154494F4E2D53544147452D32 May 8 00:45:36.015000 audit[1525]: NETFILTER_CFG table=filter:6 family=2 entries=1 op=nft_register_rule pid=1525 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.015000 audit[1525]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7ffd61ce7400 a2=0 a3=7ffd61ce73ec items=0 ppid=1483 pid=1525 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.015000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4100444F434B45522D49534F4C4154494F4E2D53544147452D31002D6A0052455455524E May 8 00:45:36.036000 audit[1530]: NETFILTER_CFG table=filter:7 family=2 entries=1 op=nft_register_rule pid=1530 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.036000 audit[1530]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7ffe0575dd30 a2=0 a3=7ffe0575dd1c items=0 ppid=1483 pid=1530 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.036000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4100444F434B45522D49534F4C4154494F4E2D53544147452D32002D6A0052455455524E May 8 00:45:36.170000 audit[1532]: NETFILTER_CFG table=filter:8 family=2 entries=1 op=nft_register_chain pid=1532 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.170000 audit[1532]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffe019a2100 a2=0 a3=7ffe019a20ec items=0 ppid=1483 pid=1532 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.170000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4E00444F434B45522D55534552 May 8 00:45:36.173000 audit[1534]: NETFILTER_CFG table=filter:9 family=2 entries=1 op=nft_register_rule pid=1534 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.173000 audit[1534]: SYSCALL arch=c000003e syscall=46 success=yes exit=212 a0=3 a1=7ffedfb15ff0 a2=0 a3=7ffedfb15fdc items=0 ppid=1483 pid=1534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.173000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4100444F434B45522D55534552002D6A0052455455524E May 8 00:45:36.175000 audit[1536]: NETFILTER_CFG table=filter:10 family=2 entries=2 op=nft_register_chain pid=1536 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.175000 audit[1536]: SYSCALL arch=c000003e syscall=46 success=yes exit=308 a0=3 a1=7ffe737fce40 a2=0 a3=7ffe737fce2c items=0 ppid=1483 pid=1536 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.175000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 May 8 00:45:36.197000 audit[1540]: NETFILTER_CFG table=filter:11 family=2 entries=1 op=nft_unregister_rule pid=1540 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.197000 audit[1540]: SYSCALL arch=c000003e syscall=46 success=yes exit=216 a0=3 a1=7ffe16745a30 a2=0 a3=7ffe16745a1c items=0 ppid=1483 pid=1540 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.197000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4400464F5257415244002D6A00444F434B45522D55534552 May 8 00:45:36.202000 audit[1541]: NETFILTER_CFG table=filter:12 family=2 entries=1 op=nft_register_rule pid=1541 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.202000 audit[1541]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7ffebbea0450 a2=0 a3=7ffebbea043c items=0 ppid=1483 pid=1541 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.202000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 May 8 00:45:36.213334 kernel: Initializing XFRM netlink socket May 8 00:45:36.273888 env[1483]: time="2025-05-08T00:45:36.273303143Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" May 8 00:45:36.291000 audit[1549]: NETFILTER_CFG table=nat:13 family=2 entries=2 op=nft_register_chain pid=1549 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.291000 audit[1549]: SYSCALL arch=c000003e syscall=46 success=yes exit=492 a0=3 a1=7fff887171f0 a2=0 a3=7fff887171dc items=0 ppid=1483 pid=1549 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.291000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D4900504F5354524F5554494E47002D73003137322E31372E302E302F31360000002D6F00646F636B657230002D6A004D415351554552414445 May 8 00:45:36.302000 audit[1552]: NETFILTER_CFG table=nat:14 family=2 entries=1 op=nft_register_rule pid=1552 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.302000 audit[1552]: SYSCALL arch=c000003e syscall=46 success=yes exit=288 a0=3 a1=7ffef42c33e0 a2=0 a3=7ffef42c33cc items=0 ppid=1483 pid=1552 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.302000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D4900444F434B4552002D6900646F636B657230002D6A0052455455524E May 8 00:45:36.305000 audit[1555]: NETFILTER_CFG table=filter:15 family=2 entries=1 op=nft_register_rule pid=1555 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.305000 audit[1555]: SYSCALL arch=c000003e syscall=46 success=yes exit=376 a0=3 a1=7fff7cfaba70 a2=0 a3=7fff7cfaba5c items=0 ppid=1483 pid=1555 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.305000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6900646F636B657230002D6F00646F636B657230002D6A00414343455054 May 8 00:45:36.307000 audit[1557]: NETFILTER_CFG table=filter:16 family=2 entries=1 op=nft_register_rule pid=1557 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.307000 audit[1557]: SYSCALL arch=c000003e syscall=46 success=yes exit=376 a0=3 a1=7fff3cb963c0 a2=0 a3=7fff3cb963ac items=0 ppid=1483 pid=1557 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.307000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6900646F636B6572300000002D6F00646F636B657230002D6A00414343455054 May 8 00:45:36.368000 audit[1559]: NETFILTER_CFG table=nat:17 family=2 entries=2 op=nft_register_chain pid=1559 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.368000 audit[1559]: SYSCALL arch=c000003e syscall=46 success=yes exit=356 a0=3 a1=7ffde1be0400 a2=0 a3=7ffde1be03ec items=0 ppid=1483 pid=1559 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.368000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D4100505245524F5554494E47002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B4552 May 8 00:45:36.371000 audit[1561]: NETFILTER_CFG table=nat:18 family=2 entries=2 op=nft_register_chain pid=1561 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.371000 audit[1561]: SYSCALL arch=c000003e syscall=46 success=yes exit=444 a0=3 a1=7fff2c611f70 a2=0 a3=7fff2c611f5c items=0 ppid=1483 pid=1561 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.371000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D74006E6174002D41004F5554505554002D6D006164647274797065002D2D6473742D74797065004C4F43414C002D6A00444F434B45520000002D2D647374003132372E302E302E302F38 May 8 00:45:36.373000 audit[1563]: NETFILTER_CFG table=filter:19 family=2 entries=1 op=nft_register_rule pid=1563 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.373000 audit[1563]: SYSCALL arch=c000003e syscall=46 success=yes exit=304 a0=3 a1=7ffe452d9090 a2=0 a3=7ffe452d907c items=0 ppid=1483 pid=1563 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.373000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6F00646F636B657230002D6A00444F434B4552 May 8 00:45:36.381000 audit[1566]: NETFILTER_CFG table=filter:20 family=2 entries=1 op=nft_register_rule pid=1566 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.381000 audit[1566]: SYSCALL arch=c000003e syscall=46 success=yes exit=508 a0=3 a1=7fffea48b190 a2=0 a3=7fffea48b17c items=0 ppid=1483 pid=1566 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.381000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6F00646F636B657230002D6D00636F6E6E747261636B002D2D637473746174650052454C415445442C45535441424C4953484544002D6A00414343455054 May 8 00:45:36.383000 audit[1568]: NETFILTER_CFG table=filter:21 family=2 entries=1 op=nft_register_rule pid=1568 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.383000 audit[1568]: SYSCALL arch=c000003e syscall=46 success=yes exit=240 a0=3 a1=7fff877bcd30 a2=0 a3=7fff877bcd1c items=0 ppid=1483 pid=1568 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.383000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D31 May 8 00:45:36.385000 audit[1570]: NETFILTER_CFG table=filter:22 family=2 entries=1 op=nft_register_rule pid=1570 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.385000 audit[1570]: SYSCALL arch=c000003e syscall=46 success=yes exit=428 a0=3 a1=7ffff1812c90 a2=0 a3=7ffff1812c7c items=0 ppid=1483 pid=1570 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.385000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4900444F434B45522D49534F4C4154494F4E2D53544147452D31002D6900646F636B6572300000002D6F00646F636B657230002D6A00444F434B45522D49534F4C4154494F4E2D53544147452D32 May 8 00:45:36.387000 audit[1572]: NETFILTER_CFG table=filter:23 family=2 entries=1 op=nft_register_rule pid=1572 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.387000 audit[1572]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffefaed31b0 a2=0 a3=7ffefaed319c items=0 ppid=1483 pid=1572 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.387000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D740066696C746572002D4900444F434B45522D49534F4C4154494F4E2D53544147452D32002D6F00646F636B657230002D6A0044524F50 May 8 00:45:36.389539 systemd-networkd[1072]: docker0: Link UP May 8 00:45:36.400000 audit[1576]: NETFILTER_CFG table=filter:24 family=2 entries=1 op=nft_unregister_rule pid=1576 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.400000 audit[1576]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7fff942add10 a2=0 a3=7fff942adcfc items=0 ppid=1483 pid=1576 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.400000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4400464F5257415244002D6A00444F434B45522D55534552 May 8 00:45:36.410000 audit[1577]: NETFILTER_CFG table=filter:25 family=2 entries=1 op=nft_register_rule pid=1577 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:45:36.410000 audit[1577]: SYSCALL arch=c000003e syscall=46 success=yes exit=224 a0=3 a1=7ffdeec14d90 a2=0 a3=7ffdeec14d7c items=0 ppid=1483 pid=1577 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:45:36.410000 audit: PROCTITLE proctitle=2F7573722F7362696E2F69707461626C6573002D2D77616974002D4900464F5257415244002D6A00444F434B45522D55534552 May 8 00:45:36.412026 env[1483]: time="2025-05-08T00:45:36.411970353Z" level=info msg="Loading containers: done." May 8 00:45:36.429958 env[1483]: time="2025-05-08T00:45:36.429834972Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 8 00:45:36.430133 env[1483]: time="2025-05-08T00:45:36.430057699Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 May 8 00:45:36.430204 env[1483]: time="2025-05-08T00:45:36.430179919Z" level=info msg="Daemon has completed initialization" May 8 00:45:36.449805 systemd[1]: Started docker.service. May 8 00:45:36.448000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=docker comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:36.453787 env[1483]: time="2025-05-08T00:45:36.453717909Z" level=info msg="API listen on /run/docker.sock" May 8 00:45:36.494788 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 8 00:45:36.494975 systemd[1]: Stopped kubelet.service. May 8 00:45:36.493000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:36.493000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:36.496483 systemd[1]: Starting kubelet.service... May 8 00:45:36.593000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:36.593967 systemd[1]: Started kubelet.service. May 8 00:45:36.743669 kubelet[1617]: E0508 00:45:36.743480 1617 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 00:45:36.746528 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 00:45:36.746686 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 00:45:36.745000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' May 8 00:45:37.655168 env[1311]: time="2025-05-08T00:45:37.655079119Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.12\"" May 8 00:45:39.365678 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3878847839.mount: Deactivated successfully. May 8 00:45:41.887034 env[1311]: time="2025-05-08T00:45:41.886944730Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:41.968562 env[1311]: time="2025-05-08T00:45:41.968502018Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:e113c59aa22f0650435e2a3ed64aadb01e87f3d2835aa3825fe078cd39699bfb,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:42.099091 env[1311]: time="2025-05-08T00:45:42.099004804Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:42.139008 env[1311]: time="2025-05-08T00:45:42.138860462Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:4878682f7a044274d42399a6316ef452c5411aafd4ad99cc57de7235ca490e4e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:42.140073 env[1311]: time="2025-05-08T00:45:42.140006982Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.12\" returns image reference \"sha256:e113c59aa22f0650435e2a3ed64aadb01e87f3d2835aa3825fe078cd39699bfb\"" May 8 00:45:42.157688 env[1311]: time="2025-05-08T00:45:42.157647320Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.12\"" May 8 00:45:45.848064 env[1311]: time="2025-05-08T00:45:45.847987183Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:45.850894 env[1311]: time="2025-05-08T00:45:45.850803115Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:70742b7b7d90a618a1fa06d89248dbe2c291c19d7f75f4ad60a69d0454dbbac8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:45.853929 env[1311]: time="2025-05-08T00:45:45.853841414Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:45.858827 env[1311]: time="2025-05-08T00:45:45.858760872Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:3a36711d0409d565b370a18d0c19339e93d4f1b1f2b3fd382eb31c714c463b74,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:45.859731 env[1311]: time="2025-05-08T00:45:45.859666280Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.12\" returns image reference \"sha256:70742b7b7d90a618a1fa06d89248dbe2c291c19d7f75f4ad60a69d0454dbbac8\"" May 8 00:45:45.874495 env[1311]: time="2025-05-08T00:45:45.874439480Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.12\"" May 8 00:45:46.994885 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. May 8 00:45:46.994000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:46.995085 systemd[1]: Stopped kubelet.service. May 8 00:45:46.996733 systemd[1]: Starting kubelet.service... May 8 00:45:46.998262 kernel: kauditd_printk_skb: 88 callbacks suppressed May 8 00:45:46.998410 kernel: audit: type=1130 audit(1746665146.994:192): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:46.994000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:47.005819 kernel: audit: type=1131 audit(1746665146.994:193): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:47.085000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:47.085994 systemd[1]: Started kubelet.service. May 8 00:45:47.101230 kernel: audit: type=1130 audit(1746665147.085:194): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:47.199347 kubelet[1658]: E0508 00:45:47.199216 1658 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 00:45:47.201256 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 00:45:47.201438 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 00:45:47.200000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' May 8 00:45:47.206165 kernel: audit: type=1131 audit(1746665147.200:195): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' May 8 00:45:51.789061 env[1311]: time="2025-05-08T00:45:51.788989055Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:51.794180 env[1311]: time="2025-05-08T00:45:51.794092617Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:c0b91cfea9f9a1c09fc5d056f3a015e52604fd0d63671ff5bf31e642402ef05d,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:51.797426 env[1311]: time="2025-05-08T00:45:51.797375265Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:51.800001 env[1311]: time="2025-05-08T00:45:51.799965824Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:521c843d01025be7d4e246ddee8cde74556eb9813c606d6db9f0f03236f6d029,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:51.800791 env[1311]: time="2025-05-08T00:45:51.800754303Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.12\" returns image reference \"sha256:c0b91cfea9f9a1c09fc5d056f3a015e52604fd0d63671ff5bf31e642402ef05d\"" May 8 00:45:51.812445 env[1311]: time="2025-05-08T00:45:51.812385019Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.12\"" May 8 00:45:54.057216 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3386824878.mount: Deactivated successfully. May 8 00:45:54.735780 env[1311]: time="2025-05-08T00:45:54.735661279Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:54.740923 env[1311]: time="2025-05-08T00:45:54.740861433Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:c9356fea5d151501039907c3ba870272461396117eabc74063632616f4e31b2b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:54.743711 env[1311]: time="2025-05-08T00:45:54.743637951Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:54.745601 env[1311]: time="2025-05-08T00:45:54.745557953Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:ea8c7d5392acf6b0c11ebba78301e1a6c2dc6abcd7544102ed578e49d1c82f15,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:54.746055 env[1311]: time="2025-05-08T00:45:54.746001435Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.12\" returns image reference \"sha256:c9356fea5d151501039907c3ba870272461396117eabc74063632616f4e31b2b\"" May 8 00:45:54.755311 env[1311]: time="2025-05-08T00:45:54.755273025Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" May 8 00:45:55.268732 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2367180559.mount: Deactivated successfully. May 8 00:45:56.563657 env[1311]: time="2025-05-08T00:45:56.563441655Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.11.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:56.568086 env[1311]: time="2025-05-08T00:45:56.568009835Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:56.570212 env[1311]: time="2025-05-08T00:45:56.570171218Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.11.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:56.574354 env[1311]: time="2025-05-08T00:45:56.574280843Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:56.575048 env[1311]: time="2025-05-08T00:45:56.574991956Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" May 8 00:45:56.585484 env[1311]: time="2025-05-08T00:45:56.585427684Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" May 8 00:45:57.244950 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. May 8 00:45:57.244000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:57.245190 systemd[1]: Stopped kubelet.service. May 8 00:45:57.246882 systemd[1]: Starting kubelet.service... May 8 00:45:57.244000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:57.251892 kernel: audit: type=1130 audit(1746665157.244:196): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:57.252121 kernel: audit: type=1131 audit(1746665157.244:197): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:57.324000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:57.325691 systemd[1]: Started kubelet.service. May 8 00:45:57.334148 kernel: audit: type=1130 audit(1746665157.324:198): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:45:57.392754 kubelet[1694]: E0508 00:45:57.392674 1694 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 8 00:45:57.394875 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 8 00:45:57.395141 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 8 00:45:57.394000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' May 8 00:45:57.400147 kernel: audit: type=1131 audit(1746665157.394:199): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed' May 8 00:45:57.808407 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3666726597.mount: Deactivated successfully. May 8 00:45:57.814569 env[1311]: time="2025-05-08T00:45:57.814448319Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:57.817120 env[1311]: time="2025-05-08T00:45:57.817064923Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:57.818837 env[1311]: time="2025-05-08T00:45:57.818798223Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:57.820487 env[1311]: time="2025-05-08T00:45:57.820449266Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:45:57.821077 env[1311]: time="2025-05-08T00:45:57.820990509Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\"" May 8 00:45:57.831439 env[1311]: time="2025-05-08T00:45:57.831377528Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" May 8 00:45:59.046727 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2242688739.mount: Deactivated successfully. May 8 00:46:03.318240 env[1311]: time="2025-05-08T00:46:03.318165590Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.12-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:03.511421 env[1311]: time="2025-05-08T00:46:03.511350135Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:03.601990 env[1311]: time="2025-05-08T00:46:03.601785456Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.12-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:03.724737 env[1311]: time="2025-05-08T00:46:03.724643196Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:03.726152 env[1311]: time="2025-05-08T00:46:03.726101553Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\"" May 8 00:46:06.409939 systemd[1]: Stopped kubelet.service. May 8 00:46:06.409000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:06.412067 systemd[1]: Starting kubelet.service... May 8 00:46:06.409000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:06.418269 kernel: audit: type=1130 audit(1746665166.409:200): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:06.418336 kernel: audit: type=1131 audit(1746665166.409:201): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:06.433628 systemd[1]: Reloading. May 8 00:46:06.502264 /usr/lib/systemd/system-generators/torcx-generator[1812]: time="2025-05-08T00:46:06Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 8 00:46:06.502802 /usr/lib/systemd/system-generators/torcx-generator[1812]: time="2025-05-08T00:46:06Z" level=info msg="torcx already run" May 8 00:46:06.911394 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 8 00:46:06.911423 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 8 00:46:06.937350 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 8 00:46:07.027293 systemd[1]: Started kubelet.service. May 8 00:46:07.027000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:07.031399 systemd[1]: Stopping kubelet.service... May 8 00:46:07.032157 kernel: audit: type=1130 audit(1746665167.027:202): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:07.032733 systemd[1]: kubelet.service: Deactivated successfully. May 8 00:46:07.032962 systemd[1]: Stopped kubelet.service. May 8 00:46:07.032000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:07.034881 systemd[1]: Starting kubelet.service... May 8 00:46:07.037157 kernel: audit: type=1131 audit(1746665167.032:203): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:07.132731 systemd[1]: Started kubelet.service. May 8 00:46:07.134000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:07.139159 kernel: audit: type=1130 audit(1746665167.134:204): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:07.183732 kubelet[1877]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 8 00:46:07.183732 kubelet[1877]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 8 00:46:07.183732 kubelet[1877]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 8 00:46:07.185435 kubelet[1877]: I0508 00:46:07.185375 1877 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 8 00:46:07.480341 kubelet[1877]: I0508 00:46:07.480203 1877 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" May 8 00:46:07.480341 kubelet[1877]: I0508 00:46:07.480237 1877 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 8 00:46:07.480525 kubelet[1877]: I0508 00:46:07.480467 1877 server.go:927] "Client rotation is on, will bootstrap in background" May 8 00:46:07.520266 kubelet[1877]: I0508 00:46:07.520206 1877 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 8 00:46:07.522869 kubelet[1877]: E0508 00:46:07.522839 1877 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.82:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:07.546507 kubelet[1877]: I0508 00:46:07.546465 1877 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 8 00:46:07.553397 kubelet[1877]: I0508 00:46:07.553338 1877 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 8 00:46:07.553623 kubelet[1877]: I0508 00:46:07.553389 1877 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} May 8 00:46:07.554276 kubelet[1877]: I0508 00:46:07.554253 1877 topology_manager.go:138] "Creating topology manager with none policy" May 8 00:46:07.554276 kubelet[1877]: I0508 00:46:07.554273 1877 container_manager_linux.go:301] "Creating device plugin manager" May 8 00:46:07.554431 kubelet[1877]: I0508 00:46:07.554410 1877 state_mem.go:36] "Initialized new in-memory state store" May 8 00:46:07.556082 kubelet[1877]: I0508 00:46:07.556054 1877 kubelet.go:400] "Attempting to sync node with API server" May 8 00:46:07.556082 kubelet[1877]: I0508 00:46:07.556080 1877 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" May 8 00:46:07.556172 kubelet[1877]: I0508 00:46:07.556130 1877 kubelet.go:312] "Adding apiserver pod source" May 8 00:46:07.556310 kubelet[1877]: I0508 00:46:07.556286 1877 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 8 00:46:07.556720 kubelet[1877]: W0508 00:46:07.556639 1877 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.82:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:07.556720 kubelet[1877]: E0508 00:46:07.556716 1877 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.82:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:07.557057 kubelet[1877]: W0508 00:46:07.556998 1877 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.82:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:07.557057 kubelet[1877]: E0508 00:46:07.557055 1877 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.82:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:07.587273 kubelet[1877]: I0508 00:46:07.587221 1877 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" May 8 00:46:07.599380 kubelet[1877]: I0508 00:46:07.599328 1877 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 8 00:46:07.599529 kubelet[1877]: W0508 00:46:07.599434 1877 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 8 00:46:07.600291 kubelet[1877]: I0508 00:46:07.600267 1877 server.go:1264] "Started kubelet" May 8 00:46:07.600810 kubelet[1877]: I0508 00:46:07.600704 1877 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 8 00:46:07.600997 kubelet[1877]: I0508 00:46:07.600881 1877 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 8 00:46:07.601294 kubelet[1877]: I0508 00:46:07.601264 1877 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 8 00:46:07.616261 kubelet[1877]: E0508 00:46:07.616220 1877 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 8 00:46:07.616722 kubelet[1877]: I0508 00:46:07.616693 1877 server.go:455] "Adding debug handlers to kubelet server" May 8 00:46:07.620000 audit[1877]: AVC avc: denied { mac_admin } for pid=1877 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:46:07.620962 kubelet[1877]: I0508 00:46:07.620845 1877 kubelet.go:1419] "Unprivileged containerized plugins might not work, could not set selinux context on plugin registration dir" path="/var/lib/kubelet/plugins_registry" err="setxattr /var/lib/kubelet/plugins_registry: invalid argument" May 8 00:46:07.620962 kubelet[1877]: I0508 00:46:07.620891 1877 kubelet.go:1423] "Unprivileged containerized plugins might not work, could not set selinux context on plugins dir" path="/var/lib/kubelet/plugins" err="setxattr /var/lib/kubelet/plugins: invalid argument" May 8 00:46:07.621045 kubelet[1877]: I0508 00:46:07.620990 1877 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 8 00:46:07.625170 kubelet[1877]: E0508 00:46:07.623001 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:07.625260 kubelet[1877]: I0508 00:46:07.623871 1877 volume_manager.go:291] "Starting Kubelet Volume Manager" May 8 00:46:07.625260 kubelet[1877]: I0508 00:46:07.623906 1877 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 8 00:46:07.625260 kubelet[1877]: W0508 00:46:07.624417 1877 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.82:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:07.625260 kubelet[1877]: E0508 00:46:07.625260 1877 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.82:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:07.625438 kubelet[1877]: E0508 00:46:07.624961 1877 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.82:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.82:6443: connect: connection refused" interval="200ms" May 8 00:46:07.625438 kubelet[1877]: I0508 00:46:07.625361 1877 reconciler.go:26] "Reconciler: start to sync state" May 8 00:46:07.625619 kubelet[1877]: I0508 00:46:07.625562 1877 factory.go:221] Registration of the systemd container factory successfully May 8 00:46:07.625708 kubelet[1877]: I0508 00:46:07.625640 1877 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 8 00:46:07.626739 kubelet[1877]: I0508 00:46:07.626718 1877 factory.go:221] Registration of the containerd container factory successfully May 8 00:46:07.643128 kernel: audit: type=1400 audit(1746665167.620:205): avc: denied { mac_admin } for pid=1877 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:46:07.643277 kernel: audit: type=1401 audit(1746665167.620:205): op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 8 00:46:07.620000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 8 00:46:07.620000 audit[1877]: SYSCALL arch=c000003e syscall=188 success=no exit=-22 a0=c000988ea0 a1=c0009f03d8 a2=c000988e70 a3=25 items=0 ppid=1 pid=1877 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.647032 kubelet[1877]: E0508 00:46:07.646876 1877 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.82:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.82:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.183d66b7385c1359 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-05-08 00:46:07.600227161 +0000 UTC m=+0.463552869,LastTimestamp:2025-05-08 00:46:07.600227161 +0000 UTC m=+0.463552869,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" May 8 00:46:07.620000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 8 00:46:07.655892 kubelet[1877]: I0508 00:46:07.655824 1877 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 8 00:46:07.657739 kubelet[1877]: I0508 00:46:07.657712 1877 cpu_manager.go:214] "Starting CPU manager" policy="none" May 8 00:46:07.657848 kubelet[1877]: I0508 00:46:07.657827 1877 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 8 00:46:07.657958 kubelet[1877]: I0508 00:46:07.657795 1877 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 8 00:46:07.658030 kubelet[1877]: I0508 00:46:07.657997 1877 status_manager.go:217] "Starting to sync pod status with apiserver" May 8 00:46:07.658063 kubelet[1877]: I0508 00:46:07.658038 1877 kubelet.go:2337] "Starting kubelet main sync loop" May 8 00:46:07.658145 kubelet[1877]: I0508 00:46:07.658126 1877 state_mem.go:36] "Initialized new in-memory state store" May 8 00:46:07.658274 kernel: audit: type=1300 audit(1746665167.620:205): arch=c000003e syscall=188 success=no exit=-22 a0=c000988ea0 a1=c0009f03d8 a2=c000988e70 a3=25 items=0 ppid=1 pid=1877 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.658333 kernel: audit: type=1327 audit(1746665167.620:205): proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 8 00:46:07.658361 kernel: audit: type=1400 audit(1746665167.620:206): avc: denied { mac_admin } for pid=1877 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:46:07.620000 audit[1877]: AVC avc: denied { mac_admin } for pid=1877 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:46:07.660084 kubelet[1877]: W0508 00:46:07.660035 1877 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.82:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:07.660214 kubelet[1877]: E0508 00:46:07.660193 1877 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.82:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:07.660310 kubelet[1877]: E0508 00:46:07.658105 1877 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 8 00:46:07.620000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 8 00:46:07.620000 audit[1877]: SYSCALL arch=c000003e syscall=188 success=no exit=-22 a0=c0009d44c0 a1=c0009f03f0 a2=c000988f30 a3=25 items=0 ppid=1 pid=1877 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.620000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 8 00:46:07.623000 audit[1889]: NETFILTER_CFG table=mangle:26 family=2 entries=2 op=nft_register_chain pid=1889 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:07.623000 audit[1889]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffd91d49900 a2=0 a3=7ffd91d498ec items=0 ppid=1877 pid=1889 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.623000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 May 8 00:46:07.624000 audit[1890]: NETFILTER_CFG table=filter:27 family=2 entries=1 op=nft_register_chain pid=1890 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:07.624000 audit[1890]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fffafc0ed70 a2=0 a3=7fffafc0ed5c items=0 ppid=1877 pid=1890 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.624000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 May 8 00:46:07.627000 audit[1892]: NETFILTER_CFG table=filter:28 family=2 entries=2 op=nft_register_chain pid=1892 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:07.627000 audit[1892]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffe11b078b0 a2=0 a3=7ffe11b0789c items=0 ppid=1877 pid=1892 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.627000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C May 8 00:46:07.629000 audit[1894]: NETFILTER_CFG table=filter:29 family=2 entries=2 op=nft_register_chain pid=1894 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:07.629000 audit[1894]: SYSCALL arch=c000003e syscall=46 success=yes exit=312 a0=3 a1=7ffe1aad0f80 a2=0 a3=7ffe1aad0f6c items=0 ppid=1877 pid=1894 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.629000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C May 8 00:46:07.655000 audit[1902]: NETFILTER_CFG table=filter:30 family=2 entries=1 op=nft_register_rule pid=1902 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:07.655000 audit[1902]: SYSCALL arch=c000003e syscall=46 success=yes exit=924 a0=3 a1=7ffc1f360db0 a2=0 a3=7ffc1f360d9c items=0 ppid=1877 pid=1902 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.655000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D41004B5542452D4649524557414C4C002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E7400626C6F636B20696E636F6D696E67206C6F63616C6E657420636F6E6E656374696F6E73002D2D647374003132372E302E302E302F38 May 8 00:46:07.657000 audit[1905]: NETFILTER_CFG table=mangle:31 family=2 entries=1 op=nft_register_chain pid=1905 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:07.657000 audit[1905]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fff3a12ded0 a2=0 a3=7fff3a12debc items=0 ppid=1877 pid=1905 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.657000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 May 8 00:46:07.657000 audit[1904]: NETFILTER_CFG table=mangle:32 family=10 entries=2 op=nft_register_chain pid=1904 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:07.657000 audit[1904]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffc45397fb0 a2=0 a3=7ffc45397f9c items=0 ppid=1877 pid=1904 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.657000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D49505441424C45532D48494E54002D74006D616E676C65 May 8 00:46:07.659000 audit[1906]: NETFILTER_CFG table=nat:33 family=2 entries=1 op=nft_register_chain pid=1906 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:07.659000 audit[1906]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffc6e0d8810 a2=0 a3=7ffc6e0d87fc items=0 ppid=1877 pid=1906 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.659000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 May 8 00:46:07.659000 audit[1907]: NETFILTER_CFG table=mangle:34 family=10 entries=1 op=nft_register_chain pid=1907 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:07.659000 audit[1907]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffd35f83400 a2=0 a3=7ffd35f833ec items=0 ppid=1877 pid=1907 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.659000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006D616E676C65 May 8 00:46:07.660000 audit[1909]: NETFILTER_CFG table=nat:35 family=10 entries=2 op=nft_register_chain pid=1909 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:07.660000 audit[1909]: SYSCALL arch=c000003e syscall=46 success=yes exit=128 a0=3 a1=7fff1d0a0fd0 a2=0 a3=7fff1d0a0fbc items=0 ppid=1877 pid=1909 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.660000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D74006E6174 May 8 00:46:07.660000 audit[1908]: NETFILTER_CFG table=filter:36 family=2 entries=1 op=nft_register_chain pid=1908 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:07.660000 audit[1908]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffd7da04fb0 a2=0 a3=7ffd7da04f9c items=0 ppid=1877 pid=1908 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.660000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 May 8 00:46:07.661000 audit[1910]: NETFILTER_CFG table=filter:37 family=10 entries=2 op=nft_register_chain pid=1910 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:07.661000 audit[1910]: SYSCALL arch=c000003e syscall=46 success=yes exit=136 a0=3 a1=7ffcd9e72c70 a2=0 a3=7ffcd9e72c5c items=0 ppid=1877 pid=1910 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:07.661000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4B5542454C45542D43414E415259002D740066696C746572 May 8 00:46:07.726626 kubelet[1877]: I0508 00:46:07.726575 1877 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 8 00:46:07.727087 kubelet[1877]: E0508 00:46:07.727045 1877 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.82:6443/api/v1/nodes\": dial tcp 10.0.0.82:6443: connect: connection refused" node="localhost" May 8 00:46:07.761237 kubelet[1877]: E0508 00:46:07.761183 1877 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 8 00:46:07.826127 kubelet[1877]: E0508 00:46:07.826065 1877 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.82:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.82:6443: connect: connection refused" interval="400ms" May 8 00:46:07.928894 kubelet[1877]: I0508 00:46:07.928847 1877 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 8 00:46:07.929210 kubelet[1877]: E0508 00:46:07.929178 1877 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.82:6443/api/v1/nodes\": dial tcp 10.0.0.82:6443: connect: connection refused" node="localhost" May 8 00:46:07.962390 kubelet[1877]: E0508 00:46:07.962330 1877 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 8 00:46:08.227579 kubelet[1877]: E0508 00:46:08.227426 1877 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.82:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.82:6443: connect: connection refused" interval="800ms" May 8 00:46:08.331430 kubelet[1877]: I0508 00:46:08.331382 1877 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 8 00:46:08.331883 kubelet[1877]: E0508 00:46:08.331825 1877 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.82:6443/api/v1/nodes\": dial tcp 10.0.0.82:6443: connect: connection refused" node="localhost" May 8 00:46:08.362982 kubelet[1877]: E0508 00:46:08.362935 1877 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 8 00:46:08.732160 kubelet[1877]: W0508 00:46:08.732090 1877 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.82:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:08.732160 kubelet[1877]: E0508 00:46:08.732167 1877 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.82:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:08.824316 kubelet[1877]: I0508 00:46:08.824243 1877 policy_none.go:49] "None policy: Start" May 8 00:46:08.825358 kubelet[1877]: I0508 00:46:08.825319 1877 memory_manager.go:170] "Starting memorymanager" policy="None" May 8 00:46:08.825358 kubelet[1877]: I0508 00:46:08.825350 1877 state_mem.go:35] "Initializing new in-memory state store" May 8 00:46:08.939271 kubelet[1877]: W0508 00:46:08.939179 1877 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.82:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:08.939271 kubelet[1877]: E0508 00:46:08.939262 1877 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.82:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:08.983288 kubelet[1877]: I0508 00:46:08.982791 1877 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 8 00:46:08.983288 kubelet[1877]: I0508 00:46:08.982884 1877 server.go:88] "Unprivileged containerized plugins might not work. Could not set selinux context on socket dir" path="/var/lib/kubelet/device-plugins/" err="setxattr /var/lib/kubelet/device-plugins/: invalid argument" May 8 00:46:08.983288 kubelet[1877]: I0508 00:46:08.983028 1877 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 8 00:46:08.983288 kubelet[1877]: I0508 00:46:08.983256 1877 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 8 00:46:08.982000 audit[1877]: AVC avc: denied { mac_admin } for pid=1877 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:46:08.982000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 8 00:46:08.982000 audit[1877]: SYSCALL arch=c000003e syscall=188 success=no exit=-22 a0=c000a6a300 a1=c000f3a8d0 a2=c000a6a2d0 a3=25 items=0 ppid=1 pid=1877 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:08.982000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 8 00:46:08.985216 kubelet[1877]: E0508 00:46:08.985194 1877 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" May 8 00:46:08.989647 kubelet[1877]: W0508 00:46:08.989593 1877 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.82:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:08.989647 kubelet[1877]: E0508 00:46:08.989647 1877 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.82:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:09.028743 kubelet[1877]: E0508 00:46:09.028681 1877 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.82:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.82:6443: connect: connection refused" interval="1.6s" May 8 00:46:09.097738 kubelet[1877]: W0508 00:46:09.097631 1877 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.82:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:09.097738 kubelet[1877]: E0508 00:46:09.097719 1877 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.82:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:09.135239 kubelet[1877]: I0508 00:46:09.135179 1877 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 8 00:46:09.135625 kubelet[1877]: E0508 00:46:09.135586 1877 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.82:6443/api/v1/nodes\": dial tcp 10.0.0.82:6443: connect: connection refused" node="localhost" May 8 00:46:09.164020 kubelet[1877]: I0508 00:46:09.163901 1877 topology_manager.go:215] "Topology Admit Handler" podUID="fab5ac8a586de009b6fecf503f384500" podNamespace="kube-system" podName="kube-apiserver-localhost" May 8 00:46:09.165753 kubelet[1877]: I0508 00:46:09.165677 1877 topology_manager.go:215] "Topology Admit Handler" podUID="b20b39a8540dba87b5883a6f0f602dba" podNamespace="kube-system" podName="kube-controller-manager-localhost" May 8 00:46:09.167288 kubelet[1877]: I0508 00:46:09.167221 1877 topology_manager.go:215] "Topology Admit Handler" podUID="6ece95f10dbffa04b25ec3439a115512" podNamespace="kube-system" podName="kube-scheduler-localhost" May 8 00:46:09.234046 kubelet[1877]: I0508 00:46:09.233789 1877 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fab5ac8a586de009b6fecf503f384500-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"fab5ac8a586de009b6fecf503f384500\") " pod="kube-system/kube-apiserver-localhost" May 8 00:46:09.234046 kubelet[1877]: I0508 00:46:09.233864 1877 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fab5ac8a586de009b6fecf503f384500-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"fab5ac8a586de009b6fecf503f384500\") " pod="kube-system/kube-apiserver-localhost" May 8 00:46:09.234046 kubelet[1877]: I0508 00:46:09.233895 1877 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fab5ac8a586de009b6fecf503f384500-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"fab5ac8a586de009b6fecf503f384500\") " pod="kube-system/kube-apiserver-localhost" May 8 00:46:09.234046 kubelet[1877]: I0508 00:46:09.233920 1877 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 00:46:09.234046 kubelet[1877]: I0508 00:46:09.233950 1877 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6ece95f10dbffa04b25ec3439a115512-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6ece95f10dbffa04b25ec3439a115512\") " pod="kube-system/kube-scheduler-localhost" May 8 00:46:09.234659 kubelet[1877]: I0508 00:46:09.233973 1877 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 00:46:09.234659 kubelet[1877]: I0508 00:46:09.234000 1877 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 00:46:09.234659 kubelet[1877]: I0508 00:46:09.234021 1877 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 00:46:09.234659 kubelet[1877]: I0508 00:46:09.234043 1877 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 00:46:09.472896 kubelet[1877]: E0508 00:46:09.472831 1877 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:09.472896 kubelet[1877]: E0508 00:46:09.472871 1877 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:09.473298 kubelet[1877]: E0508 00:46:09.473279 1877 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:09.473957 env[1311]: time="2025-05-08T00:46:09.473914951Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6ece95f10dbffa04b25ec3439a115512,Namespace:kube-system,Attempt:0,}" May 8 00:46:09.473957 env[1311]: time="2025-05-08T00:46:09.473922926Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b20b39a8540dba87b5883a6f0f602dba,Namespace:kube-system,Attempt:0,}" May 8 00:46:09.474304 env[1311]: time="2025-05-08T00:46:09.473951361Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:fab5ac8a586de009b6fecf503f384500,Namespace:kube-system,Attempt:0,}" May 8 00:46:09.511283 update_engine[1297]: I0508 00:46:09.511188 1297 update_attempter.cc:509] Updating boot flags... May 8 00:46:09.704190 kubelet[1877]: E0508 00:46:09.704104 1877 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.82:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:10.385303 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2534110834.mount: Deactivated successfully. May 8 00:46:10.396500 env[1311]: time="2025-05-08T00:46:10.396420646Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:10.402063 env[1311]: time="2025-05-08T00:46:10.402002991Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:10.403603 env[1311]: time="2025-05-08T00:46:10.403566919Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:10.405778 env[1311]: time="2025-05-08T00:46:10.405734033Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:10.407520 env[1311]: time="2025-05-08T00:46:10.407483413Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:10.409150 env[1311]: time="2025-05-08T00:46:10.409079173Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:10.410559 env[1311]: time="2025-05-08T00:46:10.410489118Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:10.412164 env[1311]: time="2025-05-08T00:46:10.412128039Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:10.413979 env[1311]: time="2025-05-08T00:46:10.413922766Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:10.415588 env[1311]: time="2025-05-08T00:46:10.415521519Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:10.418493 env[1311]: time="2025-05-08T00:46:10.418444378Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:10.426866 env[1311]: time="2025-05-08T00:46:10.426801769Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:10.459497 env[1311]: time="2025-05-08T00:46:10.459256650Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 8 00:46:10.459497 env[1311]: time="2025-05-08T00:46:10.459320040Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 8 00:46:10.459497 env[1311]: time="2025-05-08T00:46:10.459333024Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 8 00:46:10.459852 env[1311]: time="2025-05-08T00:46:10.459784331Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/ea493220747fb75f5ff309bbdc7dbcef26e8718c8c1d2d3d196e1acc1290ebc0 pid=1933 runtime=io.containerd.runc.v2 May 8 00:46:10.466621 env[1311]: time="2025-05-08T00:46:10.466336588Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 8 00:46:10.466621 env[1311]: time="2025-05-08T00:46:10.466387063Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 8 00:46:10.466621 env[1311]: time="2025-05-08T00:46:10.466401050Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 8 00:46:10.466839 env[1311]: time="2025-05-08T00:46:10.466790910Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/a1b781de35928ea5a5fedf8e523327f50fddf4fc26b5f2349d1778659779ca96 pid=1946 runtime=io.containerd.runc.v2 May 8 00:46:10.473816 env[1311]: time="2025-05-08T00:46:10.473465148Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 8 00:46:10.473816 env[1311]: time="2025-05-08T00:46:10.473566159Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 8 00:46:10.473816 env[1311]: time="2025-05-08T00:46:10.473580657Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 8 00:46:10.474365 env[1311]: time="2025-05-08T00:46:10.473876368Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/ee01f0ea38d072570c32230d5cc252d9ed88a618628f174e29bd8338786d5d32 pid=1967 runtime=io.containerd.runc.v2 May 8 00:46:10.482276 kubelet[1877]: W0508 00:46:10.482223 1877 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.82:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:10.482276 kubelet[1877]: E0508 00:46:10.482275 1877 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.82:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:10.629894 kubelet[1877]: E0508 00:46:10.629840 1877 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.82:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.82:6443: connect: connection refused" interval="3.2s" May 8 00:46:10.675289 env[1311]: time="2025-05-08T00:46:10.674916776Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b20b39a8540dba87b5883a6f0f602dba,Namespace:kube-system,Attempt:0,} returns sandbox id \"ea493220747fb75f5ff309bbdc7dbcef26e8718c8c1d2d3d196e1acc1290ebc0\"" May 8 00:46:10.676171 kubelet[1877]: E0508 00:46:10.676138 1877 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:10.679020 env[1311]: time="2025-05-08T00:46:10.678972222Z" level=info msg="CreateContainer within sandbox \"ea493220747fb75f5ff309bbdc7dbcef26e8718c8c1d2d3d196e1acc1290ebc0\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 8 00:46:10.686491 env[1311]: time="2025-05-08T00:46:10.686453893Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:fab5ac8a586de009b6fecf503f384500,Namespace:kube-system,Attempt:0,} returns sandbox id \"a1b781de35928ea5a5fedf8e523327f50fddf4fc26b5f2349d1778659779ca96\"" May 8 00:46:10.687151 kubelet[1877]: E0508 00:46:10.687126 1877 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:10.689059 env[1311]: time="2025-05-08T00:46:10.689025614Z" level=info msg="CreateContainer within sandbox \"a1b781de35928ea5a5fedf8e523327f50fddf4fc26b5f2349d1778659779ca96\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 8 00:46:10.702809 env[1311]: time="2025-05-08T00:46:10.702758830Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6ece95f10dbffa04b25ec3439a115512,Namespace:kube-system,Attempt:0,} returns sandbox id \"ee01f0ea38d072570c32230d5cc252d9ed88a618628f174e29bd8338786d5d32\"" May 8 00:46:10.703521 kubelet[1877]: E0508 00:46:10.703464 1877 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:10.705242 env[1311]: time="2025-05-08T00:46:10.705214291Z" level=info msg="CreateContainer within sandbox \"ee01f0ea38d072570c32230d5cc252d9ed88a618628f174e29bd8338786d5d32\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 8 00:46:10.737319 kubelet[1877]: I0508 00:46:10.737280 1877 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 8 00:46:10.737658 kubelet[1877]: E0508 00:46:10.737624 1877 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.82:6443/api/v1/nodes\": dial tcp 10.0.0.82:6443: connect: connection refused" node="localhost" May 8 00:46:10.942697 env[1311]: time="2025-05-08T00:46:10.942547754Z" level=info msg="CreateContainer within sandbox \"ea493220747fb75f5ff309bbdc7dbcef26e8718c8c1d2d3d196e1acc1290ebc0\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"1de3f6c99bb0b2947a1e04918feefa433d4d67db28ad9ae2b7c192c7d9076056\"" May 8 00:46:10.943675 env[1311]: time="2025-05-08T00:46:10.943610301Z" level=info msg="StartContainer for \"1de3f6c99bb0b2947a1e04918feefa433d4d67db28ad9ae2b7c192c7d9076056\"" May 8 00:46:10.949980 env[1311]: time="2025-05-08T00:46:10.949929344Z" level=info msg="CreateContainer within sandbox \"a1b781de35928ea5a5fedf8e523327f50fddf4fc26b5f2349d1778659779ca96\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"84feb454bc178e95d73a553a438936e241a94838d8d934fa581963d84aa43fbf\"" May 8 00:46:10.950715 env[1311]: time="2025-05-08T00:46:10.950670701Z" level=info msg="StartContainer for \"84feb454bc178e95d73a553a438936e241a94838d8d934fa581963d84aa43fbf\"" May 8 00:46:10.951781 env[1311]: time="2025-05-08T00:46:10.951728619Z" level=info msg="CreateContainer within sandbox \"ee01f0ea38d072570c32230d5cc252d9ed88a618628f174e29bd8338786d5d32\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"9f30e5b9faadefb12a5d25aec6b941abb091c259c5b5671afa38c611c67ee8af\"" May 8 00:46:10.952300 env[1311]: time="2025-05-08T00:46:10.952263745Z" level=info msg="StartContainer for \"9f30e5b9faadefb12a5d25aec6b941abb091c259c5b5671afa38c611c67ee8af\"" May 8 00:46:11.033440 kubelet[1877]: W0508 00:46:11.033382 1877 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.82:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:11.033440 kubelet[1877]: E0508 00:46:11.033435 1877 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.82:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.82:6443: connect: connection refused May 8 00:46:11.084228 env[1311]: time="2025-05-08T00:46:11.080123680Z" level=info msg="StartContainer for \"1de3f6c99bb0b2947a1e04918feefa433d4d67db28ad9ae2b7c192c7d9076056\" returns successfully" May 8 00:46:11.097058 env[1311]: time="2025-05-08T00:46:11.097012051Z" level=info msg="StartContainer for \"9f30e5b9faadefb12a5d25aec6b941abb091c259c5b5671afa38c611c67ee8af\" returns successfully" May 8 00:46:11.098343 env[1311]: time="2025-05-08T00:46:11.098312518Z" level=info msg="StartContainer for \"84feb454bc178e95d73a553a438936e241a94838d8d934fa581963d84aa43fbf\" returns successfully" May 8 00:46:11.668425 kubelet[1877]: E0508 00:46:11.668382 1877 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:11.670032 kubelet[1877]: E0508 00:46:11.670004 1877 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:11.672924 kubelet[1877]: E0508 00:46:11.672884 1877 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:12.674870 kubelet[1877]: E0508 00:46:12.674812 1877 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:13.489325 kubelet[1877]: E0508 00:46:13.489272 1877 csi_plugin.go:308] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "localhost" not found May 8 00:46:13.676380 kubelet[1877]: E0508 00:46:13.676338 1877 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:13.939575 kubelet[1877]: I0508 00:46:13.939541 1877 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 8 00:46:14.123637 kubelet[1877]: I0508 00:46:14.123568 1877 kubelet_node_status.go:76] "Successfully registered node" node="localhost" May 8 00:46:14.157184 kubelet[1877]: E0508 00:46:14.157147 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:14.258144 kubelet[1877]: E0508 00:46:14.258087 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:14.358762 kubelet[1877]: E0508 00:46:14.358706 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:14.459342 kubelet[1877]: E0508 00:46:14.459281 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:14.559918 kubelet[1877]: E0508 00:46:14.559757 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:14.660156 kubelet[1877]: E0508 00:46:14.660071 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:14.760287 kubelet[1877]: E0508 00:46:14.760224 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:14.860991 kubelet[1877]: E0508 00:46:14.860843 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:14.961340 kubelet[1877]: E0508 00:46:14.961264 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:15.062510 kubelet[1877]: E0508 00:46:15.062415 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:15.163251 kubelet[1877]: E0508 00:46:15.163142 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:15.263904 kubelet[1877]: E0508 00:46:15.263839 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:15.318199 kubelet[1877]: E0508 00:46:15.318154 1877 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:15.364167 kubelet[1877]: E0508 00:46:15.364096 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:15.464639 kubelet[1877]: E0508 00:46:15.464445 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:15.565562 kubelet[1877]: E0508 00:46:15.565496 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:15.642126 systemd[1]: Reloading. May 8 00:46:15.666655 kubelet[1877]: E0508 00:46:15.666613 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:15.704311 /usr/lib/systemd/system-generators/torcx-generator[2190]: time="2025-05-08T00:46:15Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 8 00:46:15.704336 /usr/lib/systemd/system-generators/torcx-generator[2190]: time="2025-05-08T00:46:15Z" level=info msg="torcx already run" May 8 00:46:15.765253 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 8 00:46:15.765271 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 8 00:46:15.767160 kubelet[1877]: E0508 00:46:15.767092 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:15.786042 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 8 00:46:15.867760 kubelet[1877]: E0508 00:46:15.867715 1877 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 8 00:46:15.870259 systemd[1]: Stopping kubelet.service... May 8 00:46:15.891670 systemd[1]: kubelet.service: Deactivated successfully. May 8 00:46:15.891965 systemd[1]: Stopped kubelet.service. May 8 00:46:15.890000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:15.894035 systemd[1]: Starting kubelet.service... May 8 00:46:15.899020 kernel: kauditd_printk_skb: 43 callbacks suppressed May 8 00:46:15.899133 kernel: audit: type=1131 audit(1746665175.890:220): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:16.051128 systemd[1]: Started kubelet.service. May 8 00:46:16.050000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:16.058142 kernel: audit: type=1130 audit(1746665176.050:221): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kubelet comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:16.115832 kubelet[2246]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 8 00:46:16.115832 kubelet[2246]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 8 00:46:16.115832 kubelet[2246]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 8 00:46:16.116398 kubelet[2246]: I0508 00:46:16.115895 2246 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 8 00:46:16.120926 kubelet[2246]: I0508 00:46:16.120884 2246 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" May 8 00:46:16.120926 kubelet[2246]: I0508 00:46:16.120917 2246 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 8 00:46:16.123797 kubelet[2246]: I0508 00:46:16.121131 2246 server.go:927] "Client rotation is on, will bootstrap in background" May 8 00:46:16.124930 kubelet[2246]: I0508 00:46:16.124904 2246 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". May 8 00:46:16.126051 kubelet[2246]: I0508 00:46:16.125941 2246 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 8 00:46:16.135283 kubelet[2246]: I0508 00:46:16.135258 2246 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 8 00:46:16.135780 kubelet[2246]: I0508 00:46:16.135714 2246 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 8 00:46:16.135965 kubelet[2246]: I0508 00:46:16.135755 2246 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} May 8 00:46:16.135965 kubelet[2246]: I0508 00:46:16.135948 2246 topology_manager.go:138] "Creating topology manager with none policy" May 8 00:46:16.135965 kubelet[2246]: I0508 00:46:16.135959 2246 container_manager_linux.go:301] "Creating device plugin manager" May 8 00:46:16.136101 kubelet[2246]: I0508 00:46:16.135993 2246 state_mem.go:36] "Initialized new in-memory state store" May 8 00:46:16.136101 kubelet[2246]: I0508 00:46:16.136073 2246 kubelet.go:400] "Attempting to sync node with API server" May 8 00:46:16.136101 kubelet[2246]: I0508 00:46:16.136084 2246 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" May 8 00:46:16.136226 kubelet[2246]: I0508 00:46:16.136105 2246 kubelet.go:312] "Adding apiserver pod source" May 8 00:46:16.136226 kubelet[2246]: I0508 00:46:16.136129 2246 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 8 00:46:16.136945 kubelet[2246]: I0508 00:46:16.136920 2246 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" May 8 00:46:16.137087 kubelet[2246]: I0508 00:46:16.137064 2246 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 8 00:46:16.137436 kubelet[2246]: I0508 00:46:16.137415 2246 server.go:1264] "Started kubelet" May 8 00:46:16.137000 audit[2246]: AVC avc: denied { mac_admin } for pid=2246 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:46:16.150309 kubelet[2246]: I0508 00:46:16.138870 2246 kubelet.go:1419] "Unprivileged containerized plugins might not work, could not set selinux context on plugin registration dir" path="/var/lib/kubelet/plugins_registry" err="setxattr /var/lib/kubelet/plugins_registry: invalid argument" May 8 00:46:16.150309 kubelet[2246]: I0508 00:46:16.138914 2246 kubelet.go:1423] "Unprivileged containerized plugins might not work, could not set selinux context on plugins dir" path="/var/lib/kubelet/plugins" err="setxattr /var/lib/kubelet/plugins: invalid argument" May 8 00:46:16.150309 kubelet[2246]: I0508 00:46:16.138940 2246 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 8 00:46:16.150309 kubelet[2246]: I0508 00:46:16.141802 2246 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 8 00:46:16.150309 kubelet[2246]: I0508 00:46:16.143149 2246 server.go:455] "Adding debug handlers to kubelet server" May 8 00:46:16.150309 kubelet[2246]: I0508 00:46:16.145529 2246 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 8 00:46:16.150309 kubelet[2246]: I0508 00:46:16.145738 2246 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 8 00:46:16.150950 kubelet[2246]: I0508 00:46:16.150921 2246 volume_manager.go:291] "Starting Kubelet Volume Manager" May 8 00:46:16.151283 kubelet[2246]: I0508 00:46:16.151269 2246 reconciler.go:26] "Reconciler: start to sync state" May 8 00:46:16.151403 kubelet[2246]: I0508 00:46:16.151387 2246 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 8 00:46:16.157178 kubelet[2246]: I0508 00:46:16.157159 2246 factory.go:221] Registration of the containerd container factory successfully May 8 00:46:16.157297 kubelet[2246]: I0508 00:46:16.157282 2246 factory.go:221] Registration of the systemd container factory successfully May 8 00:46:16.157434 kubelet[2246]: I0508 00:46:16.157414 2246 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 8 00:46:16.159570 kubelet[2246]: E0508 00:46:16.159542 2246 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 8 00:46:16.163789 kubelet[2246]: I0508 00:46:16.163750 2246 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 8 00:46:16.164619 kubelet[2246]: I0508 00:46:16.164595 2246 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 8 00:46:16.164702 kubelet[2246]: I0508 00:46:16.164627 2246 status_manager.go:217] "Starting to sync pod status with apiserver" May 8 00:46:16.164702 kubelet[2246]: I0508 00:46:16.164649 2246 kubelet.go:2337] "Starting kubelet main sync loop" May 8 00:46:16.164702 kubelet[2246]: E0508 00:46:16.164690 2246 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 8 00:46:16.172899 kernel: audit: type=1400 audit(1746665176.137:222): avc: denied { mac_admin } for pid=2246 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:46:16.173027 kernel: audit: type=1401 audit(1746665176.137:222): op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 8 00:46:16.137000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 8 00:46:16.137000 audit[2246]: SYSCALL arch=c000003e syscall=188 success=no exit=-22 a0=c000ce6630 a1=c000c670e0 a2=c000ce6600 a3=25 items=0 ppid=1 pid=2246 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:16.137000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 8 00:46:16.183009 kernel: audit: type=1300 audit(1746665176.137:222): arch=c000003e syscall=188 success=no exit=-22 a0=c000ce6630 a1=c000c670e0 a2=c000ce6600 a3=25 items=0 ppid=1 pid=2246 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:16.183104 kernel: audit: type=1327 audit(1746665176.137:222): proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 8 00:46:16.183144 kernel: audit: type=1400 audit(1746665176.137:223): avc: denied { mac_admin } for pid=2246 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:46:16.137000 audit[2246]: AVC avc: denied { mac_admin } for pid=2246 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:46:16.186049 kernel: audit: type=1401 audit(1746665176.137:223): op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 8 00:46:16.137000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 8 00:46:16.187630 kernel: audit: type=1300 audit(1746665176.137:223): arch=c000003e syscall=188 success=no exit=-22 a0=c00082b8c0 a1=c000c670f8 a2=c000ce66c0 a3=25 items=0 ppid=1 pid=2246 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:16.137000 audit[2246]: SYSCALL arch=c000003e syscall=188 success=no exit=-22 a0=c00082b8c0 a1=c000c670f8 a2=c000ce66c0 a3=25 items=0 ppid=1 pid=2246 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:16.196235 kernel: audit: type=1327 audit(1746665176.137:223): proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 8 00:46:16.137000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 8 00:46:16.205501 kubelet[2246]: I0508 00:46:16.205410 2246 cpu_manager.go:214] "Starting CPU manager" policy="none" May 8 00:46:16.205501 kubelet[2246]: I0508 00:46:16.205430 2246 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 8 00:46:16.205501 kubelet[2246]: I0508 00:46:16.205457 2246 state_mem.go:36] "Initialized new in-memory state store" May 8 00:46:16.205750 kubelet[2246]: I0508 00:46:16.205640 2246 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 8 00:46:16.205750 kubelet[2246]: I0508 00:46:16.205650 2246 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 8 00:46:16.205750 kubelet[2246]: I0508 00:46:16.205668 2246 policy_none.go:49] "None policy: Start" May 8 00:46:16.207539 kubelet[2246]: I0508 00:46:16.207517 2246 memory_manager.go:170] "Starting memorymanager" policy="None" May 8 00:46:16.207669 kubelet[2246]: I0508 00:46:16.207652 2246 state_mem.go:35] "Initializing new in-memory state store" May 8 00:46:16.208004 kubelet[2246]: I0508 00:46:16.207983 2246 state_mem.go:75] "Updated machine memory state" May 8 00:46:16.210331 kubelet[2246]: I0508 00:46:16.210307 2246 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 8 00:46:16.209000 audit[2246]: AVC avc: denied { mac_admin } for pid=2246 comm="kubelet" capability=33 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:46:16.209000 audit: SELINUX_ERR op=setxattr invalid_context="system_u:object_r:container_file_t:s0" May 8 00:46:16.209000 audit[2246]: SYSCALL arch=c000003e syscall=188 success=no exit=-22 a0=c000c52f90 a1=c0006fc9d8 a2=c000c52f60 a3=25 items=0 ppid=1 pid=2246 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="kubelet" exe="/usr/bin/kubelet" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:16.209000 audit: PROCTITLE proctitle=2F7573722F62696E2F6B7562656C6574002D2D626F6F7473747261702D6B756265636F6E6669673D2F6574632F6B756265726E657465732F626F6F7473747261702D6B7562656C65742E636F6E66002D2D6B756265636F6E6669673D2F6574632F6B756265726E657465732F6B7562656C65742E636F6E66002D2D636F6E6669 May 8 00:46:16.210600 kubelet[2246]: I0508 00:46:16.210369 2246 server.go:88] "Unprivileged containerized plugins might not work. Could not set selinux context on socket dir" path="/var/lib/kubelet/device-plugins/" err="setxattr /var/lib/kubelet/device-plugins/: invalid argument" May 8 00:46:16.210600 kubelet[2246]: I0508 00:46:16.210507 2246 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 8 00:46:16.210600 kubelet[2246]: I0508 00:46:16.210586 2246 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 8 00:46:16.255324 kubelet[2246]: I0508 00:46:16.255271 2246 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 8 00:46:16.265571 kubelet[2246]: I0508 00:46:16.265514 2246 topology_manager.go:215] "Topology Admit Handler" podUID="fab5ac8a586de009b6fecf503f384500" podNamespace="kube-system" podName="kube-apiserver-localhost" May 8 00:46:16.265718 kubelet[2246]: I0508 00:46:16.265687 2246 topology_manager.go:215] "Topology Admit Handler" podUID="b20b39a8540dba87b5883a6f0f602dba" podNamespace="kube-system" podName="kube-controller-manager-localhost" May 8 00:46:16.265769 kubelet[2246]: I0508 00:46:16.265754 2246 topology_manager.go:215] "Topology Admit Handler" podUID="6ece95f10dbffa04b25ec3439a115512" podNamespace="kube-system" podName="kube-scheduler-localhost" May 8 00:46:16.452443 kubelet[2246]: I0508 00:46:16.452314 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 00:46:16.452443 kubelet[2246]: I0508 00:46:16.452356 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 00:46:16.452443 kubelet[2246]: I0508 00:46:16.452375 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 00:46:16.452443 kubelet[2246]: I0508 00:46:16.452398 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fab5ac8a586de009b6fecf503f384500-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"fab5ac8a586de009b6fecf503f384500\") " pod="kube-system/kube-apiserver-localhost" May 8 00:46:16.452443 kubelet[2246]: I0508 00:46:16.452413 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fab5ac8a586de009b6fecf503f384500-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"fab5ac8a586de009b6fecf503f384500\") " pod="kube-system/kube-apiserver-localhost" May 8 00:46:16.452738 kubelet[2246]: I0508 00:46:16.452425 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 00:46:16.452738 kubelet[2246]: I0508 00:46:16.452441 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 8 00:46:16.452738 kubelet[2246]: I0508 00:46:16.452454 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6ece95f10dbffa04b25ec3439a115512-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6ece95f10dbffa04b25ec3439a115512\") " pod="kube-system/kube-scheduler-localhost" May 8 00:46:16.452738 kubelet[2246]: I0508 00:46:16.452467 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fab5ac8a586de009b6fecf503f384500-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"fab5ac8a586de009b6fecf503f384500\") " pod="kube-system/kube-apiserver-localhost" May 8 00:46:16.466176 kubelet[2246]: I0508 00:46:16.465997 2246 kubelet_node_status.go:112] "Node was previously registered" node="localhost" May 8 00:46:16.466176 kubelet[2246]: I0508 00:46:16.466090 2246 kubelet_node_status.go:76] "Successfully registered node" node="localhost" May 8 00:46:16.806925 kubelet[2246]: E0508 00:46:16.806876 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:16.806925 kubelet[2246]: E0508 00:46:16.806927 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:16.807377 kubelet[2246]: E0508 00:46:16.807358 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:17.137209 kubelet[2246]: I0508 00:46:17.137011 2246 apiserver.go:52] "Watching apiserver" May 8 00:46:17.152613 kubelet[2246]: I0508 00:46:17.152554 2246 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 8 00:46:17.182981 kubelet[2246]: E0508 00:46:17.179771 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:17.182981 kubelet[2246]: E0508 00:46:17.180429 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:17.191804 kubelet[2246]: E0508 00:46:17.191712 2246 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" May 8 00:46:17.192256 kubelet[2246]: E0508 00:46:17.192231 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:17.213543 kubelet[2246]: I0508 00:46:17.213451 2246 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.213407445 podStartE2EDuration="1.213407445s" podCreationTimestamp="2025-05-08 00:46:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 00:46:17.205916859 +0000 UTC m=+1.146731330" watchObservedRunningTime="2025-05-08 00:46:17.213407445 +0000 UTC m=+1.154221926" May 8 00:46:17.256764 kubelet[2246]: I0508 00:46:17.256690 2246 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.2566687349999999 podStartE2EDuration="1.256668735s" podCreationTimestamp="2025-05-08 00:46:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 00:46:17.256381602 +0000 UTC m=+1.197196093" watchObservedRunningTime="2025-05-08 00:46:17.256668735 +0000 UTC m=+1.197483206" May 8 00:46:17.257037 kubelet[2246]: I0508 00:46:17.256797 2246 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.256789343 podStartE2EDuration="1.256789343s" podCreationTimestamp="2025-05-08 00:46:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 00:46:17.23919936 +0000 UTC m=+1.180013831" watchObservedRunningTime="2025-05-08 00:46:17.256789343 +0000 UTC m=+1.197603844" May 8 00:46:18.180938 kubelet[2246]: E0508 00:46:18.180893 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:18.515355 kubelet[2246]: E0508 00:46:18.515274 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:19.173782 kubelet[2246]: E0508 00:46:19.173736 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:19.183572 kubelet[2246]: E0508 00:46:19.183537 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:22.435860 sudo[1471]: pam_unix(sudo:session): session closed for user root May 8 00:46:22.434000 audit[1471]: USER_END pid=1471 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 8 00:46:22.436897 kernel: kauditd_printk_skb: 4 callbacks suppressed May 8 00:46:22.436958 kernel: audit: type=1106 audit(1746665182.434:225): pid=1471 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_limits,pam_env,pam_unix,pam_permit,pam_systemd acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 8 00:46:22.434000 audit[1471]: CRED_DISP pid=1471 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 8 00:46:22.444712 kernel: audit: type=1104 audit(1746665182.434:226): pid=1471 uid=500 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="root" exe="/usr/bin/sudo" hostname=? addr=? terminal=? res=success' May 8 00:46:22.451212 sshd[1465]: pam_unix(sshd:session): session closed for user core May 8 00:46:22.451000 audit[1465]: USER_END pid=1465 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:22.454592 systemd[1]: sshd@6-10.0.0.82:22-10.0.0.1:56242.service: Deactivated successfully. May 8 00:46:22.456097 systemd[1]: session-7.scope: Deactivated successfully. May 8 00:46:22.456164 systemd-logind[1291]: Session 7 logged out. Waiting for processes to exit. May 8 00:46:22.457289 systemd-logind[1291]: Removed session 7. May 8 00:46:22.451000 audit[1465]: CRED_DISP pid=1465 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:22.461900 kernel: audit: type=1106 audit(1746665182.451:227): pid=1465 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:22.461961 kernel: audit: type=1104 audit(1746665182.451:228): pid=1465 uid=0 auid=500 ses=7 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:22.461998 kernel: audit: type=1131 audit(1746665182.451:229): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-10.0.0.82:22-10.0.0.1:56242 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:22.451000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@6-10.0.0.82:22-10.0.0.1:56242 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:25.245751 kubelet[2246]: E0508 00:46:25.245686 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:26.195704 kubelet[2246]: E0508 00:46:26.195664 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:27.197567 kubelet[2246]: E0508 00:46:27.197502 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:28.519556 kubelet[2246]: E0508 00:46:28.519500 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:29.178963 kubelet[2246]: E0508 00:46:29.178925 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:30.179355 kubelet[2246]: I0508 00:46:30.179305 2246 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 8 00:46:30.179833 kubelet[2246]: I0508 00:46:30.179734 2246 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 8 00:46:30.179896 env[1311]: time="2025-05-08T00:46:30.179604557Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 8 00:46:30.241637 kubelet[2246]: I0508 00:46:30.241571 2246 topology_manager.go:215] "Topology Admit Handler" podUID="994e834b-b433-4935-805d-e1055df5f070" podNamespace="tigera-operator" podName="tigera-operator-797db67f8-wvrpq" May 8 00:46:30.443236 kubelet[2246]: I0508 00:46:30.443077 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/994e834b-b433-4935-805d-e1055df5f070-var-lib-calico\") pod \"tigera-operator-797db67f8-wvrpq\" (UID: \"994e834b-b433-4935-805d-e1055df5f070\") " pod="tigera-operator/tigera-operator-797db67f8-wvrpq" May 8 00:46:30.443236 kubelet[2246]: I0508 00:46:30.443159 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r24nt\" (UniqueName: \"kubernetes.io/projected/994e834b-b433-4935-805d-e1055df5f070-kube-api-access-r24nt\") pod \"tigera-operator-797db67f8-wvrpq\" (UID: \"994e834b-b433-4935-805d-e1055df5f070\") " pod="tigera-operator/tigera-operator-797db67f8-wvrpq" May 8 00:46:30.446743 kubelet[2246]: I0508 00:46:30.446701 2246 topology_manager.go:215] "Topology Admit Handler" podUID="4e2293ba-429d-4ed0-a9d3-c55493a33556" podNamespace="kube-system" podName="kube-proxy-h4c2z" May 8 00:46:30.644778 kubelet[2246]: I0508 00:46:30.644676 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/4e2293ba-429d-4ed0-a9d3-c55493a33556-kube-proxy\") pod \"kube-proxy-h4c2z\" (UID: \"4e2293ba-429d-4ed0-a9d3-c55493a33556\") " pod="kube-system/kube-proxy-h4c2z" May 8 00:46:30.644778 kubelet[2246]: I0508 00:46:30.644736 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4e2293ba-429d-4ed0-a9d3-c55493a33556-lib-modules\") pod \"kube-proxy-h4c2z\" (UID: \"4e2293ba-429d-4ed0-a9d3-c55493a33556\") " pod="kube-system/kube-proxy-h4c2z" May 8 00:46:30.644778 kubelet[2246]: I0508 00:46:30.644755 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6bqr\" (UniqueName: \"kubernetes.io/projected/4e2293ba-429d-4ed0-a9d3-c55493a33556-kube-api-access-p6bqr\") pod \"kube-proxy-h4c2z\" (UID: \"4e2293ba-429d-4ed0-a9d3-c55493a33556\") " pod="kube-system/kube-proxy-h4c2z" May 8 00:46:30.644778 kubelet[2246]: I0508 00:46:30.644774 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/4e2293ba-429d-4ed0-a9d3-c55493a33556-xtables-lock\") pod \"kube-proxy-h4c2z\" (UID: \"4e2293ba-429d-4ed0-a9d3-c55493a33556\") " pod="kube-system/kube-proxy-h4c2z" May 8 00:46:30.847472 env[1311]: time="2025-05-08T00:46:30.847424213Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-797db67f8-wvrpq,Uid:994e834b-b433-4935-805d-e1055df5f070,Namespace:tigera-operator,Attempt:0,}" May 8 00:46:30.869559 env[1311]: time="2025-05-08T00:46:30.869451515Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 8 00:46:30.869559 env[1311]: time="2025-05-08T00:46:30.869512761Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 8 00:46:30.869559 env[1311]: time="2025-05-08T00:46:30.869525575Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 8 00:46:30.869766 env[1311]: time="2025-05-08T00:46:30.869726663Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/56ee6c12bb09607ada2c780138ec646ea88bb60199002b37d130f4ff09189391 pid=2342 runtime=io.containerd.runc.v2 May 8 00:46:30.920988 env[1311]: time="2025-05-08T00:46:30.920913790Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-797db67f8-wvrpq,Uid:994e834b-b433-4935-805d-e1055df5f070,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"56ee6c12bb09607ada2c780138ec646ea88bb60199002b37d130f4ff09189391\"" May 8 00:46:30.923198 env[1311]: time="2025-05-08T00:46:30.923153554Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\"" May 8 00:46:31.052037 kubelet[2246]: E0508 00:46:31.051968 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:31.052835 env[1311]: time="2025-05-08T00:46:31.052769562Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-h4c2z,Uid:4e2293ba-429d-4ed0-a9d3-c55493a33556,Namespace:kube-system,Attempt:0,}" May 8 00:46:31.075747 env[1311]: time="2025-05-08T00:46:31.075649169Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 8 00:46:31.075747 env[1311]: time="2025-05-08T00:46:31.075705285Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 8 00:46:31.075747 env[1311]: time="2025-05-08T00:46:31.075717768Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 8 00:46:31.075984 env[1311]: time="2025-05-08T00:46:31.075912134Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/bf29100e4456b65a684fa9643669344df6820cbaae370f90e60ffa84894a1497 pid=2382 runtime=io.containerd.runc.v2 May 8 00:46:31.110521 env[1311]: time="2025-05-08T00:46:31.110376257Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-h4c2z,Uid:4e2293ba-429d-4ed0-a9d3-c55493a33556,Namespace:kube-system,Attempt:0,} returns sandbox id \"bf29100e4456b65a684fa9643669344df6820cbaae370f90e60ffa84894a1497\"" May 8 00:46:31.111622 kubelet[2246]: E0508 00:46:31.111586 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:31.113684 env[1311]: time="2025-05-08T00:46:31.113627645Z" level=info msg="CreateContainer within sandbox \"bf29100e4456b65a684fa9643669344df6820cbaae370f90e60ffa84894a1497\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 8 00:46:31.153016 env[1311]: time="2025-05-08T00:46:31.152927265Z" level=info msg="CreateContainer within sandbox \"bf29100e4456b65a684fa9643669344df6820cbaae370f90e60ffa84894a1497\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"87027267003366f5731485bcc056f9d64ed6ba33d0c8c3953a5cea6b5d2e657a\"" May 8 00:46:31.153794 env[1311]: time="2025-05-08T00:46:31.153756075Z" level=info msg="StartContainer for \"87027267003366f5731485bcc056f9d64ed6ba33d0c8c3953a5cea6b5d2e657a\"" May 8 00:46:31.208575 env[1311]: time="2025-05-08T00:46:31.208492529Z" level=info msg="StartContainer for \"87027267003366f5731485bcc056f9d64ed6ba33d0c8c3953a5cea6b5d2e657a\" returns successfully" May 8 00:46:31.264000 audit[2477]: NETFILTER_CFG table=mangle:38 family=10 entries=1 op=nft_register_chain pid=2477 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.264000 audit[2477]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffc50d174e0 a2=0 a3=7ffc50d174cc items=0 ppid=2433 pid=2477 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.274358 kernel: audit: type=1325 audit(1746665191.264:230): table=mangle:38 family=10 entries=1 op=nft_register_chain pid=2477 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.274551 kernel: audit: type=1300 audit(1746665191.264:230): arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffc50d174e0 a2=0 a3=7ffc50d174cc items=0 ppid=2433 pid=2477 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.274577 kernel: audit: type=1327 audit(1746665191.264:230): proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 May 8 00:46:31.264000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 May 8 00:46:31.264000 audit[2476]: NETFILTER_CFG table=mangle:39 family=2 entries=1 op=nft_register_chain pid=2476 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.279091 kernel: audit: type=1325 audit(1746665191.264:231): table=mangle:39 family=2 entries=1 op=nft_register_chain pid=2476 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.279287 kernel: audit: type=1300 audit(1746665191.264:231): arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffc3dd0d170 a2=0 a3=7ffc3dd0d15c items=0 ppid=2433 pid=2476 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.264000 audit[2476]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffc3dd0d170 a2=0 a3=7ffc3dd0d15c items=0 ppid=2433 pid=2476 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.264000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 May 8 00:46:31.286054 kernel: audit: type=1327 audit(1746665191.264:231): proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006D616E676C65 May 8 00:46:31.286127 kernel: audit: type=1325 audit(1746665191.265:232): table=nat:40 family=10 entries=1 op=nft_register_chain pid=2478 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.265000 audit[2478]: NETFILTER_CFG table=nat:40 family=10 entries=1 op=nft_register_chain pid=2478 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.265000 audit[2478]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe3911d200 a2=0 a3=7ffe3911d1ec items=0 ppid=2433 pid=2478 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.293181 kernel: audit: type=1300 audit(1746665191.265:232): arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffe3911d200 a2=0 a3=7ffe3911d1ec items=0 ppid=2433 pid=2478 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.293287 kernel: audit: type=1327 audit(1746665191.265:232): proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 May 8 00:46:31.265000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 May 8 00:46:31.265000 audit[2479]: NETFILTER_CFG table=nat:41 family=2 entries=1 op=nft_register_chain pid=2479 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.297659 kernel: audit: type=1325 audit(1746665191.265:233): table=nat:41 family=2 entries=1 op=nft_register_chain pid=2479 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.265000 audit[2479]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffde75f4700 a2=0 a3=7ffde75f46ec items=0 ppid=2433 pid=2479 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.265000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D74006E6174 May 8 00:46:31.266000 audit[2480]: NETFILTER_CFG table=filter:42 family=10 entries=1 op=nft_register_chain pid=2480 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.266000 audit[2480]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7fff957e76f0 a2=0 a3=7fff957e76dc items=0 ppid=2433 pid=2480 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.266000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 May 8 00:46:31.266000 audit[2481]: NETFILTER_CFG table=filter:43 family=2 entries=1 op=nft_register_chain pid=2481 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.266000 audit[2481]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffd52e44320 a2=0 a3=7ffd52e4430c items=0 ppid=2433 pid=2481 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.266000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D43414E415259002D740066696C746572 May 8 00:46:31.371000 audit[2482]: NETFILTER_CFG table=filter:44 family=2 entries=1 op=nft_register_chain pid=2482 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.371000 audit[2482]: SYSCALL arch=c000003e syscall=46 success=yes exit=108 a0=3 a1=7fff89d0be50 a2=0 a3=7fff89d0be3c items=0 ppid=2433 pid=2482 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.371000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 May 8 00:46:31.374000 audit[2484]: NETFILTER_CFG table=filter:45 family=2 entries=1 op=nft_register_rule pid=2484 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.374000 audit[2484]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffcd484d9b0 a2=0 a3=7ffcd484d99c items=0 ppid=2433 pid=2484 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.374000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C652073657276696365 May 8 00:46:31.378000 audit[2487]: NETFILTER_CFG table=filter:46 family=2 entries=1 op=nft_register_rule pid=2487 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.378000 audit[2487]: SYSCALL arch=c000003e syscall=46 success=yes exit=752 a0=3 a1=7ffdd2735d60 a2=0 a3=7ffdd2735d4c items=0 ppid=2433 pid=2487 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.378000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C65207365727669 May 8 00:46:31.379000 audit[2488]: NETFILTER_CFG table=filter:47 family=2 entries=1 op=nft_register_chain pid=2488 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.379000 audit[2488]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffea2c53ef0 a2=0 a3=7ffea2c53edc items=0 ppid=2433 pid=2488 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.379000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D4E4F4445504F525453002D740066696C746572 May 8 00:46:31.382000 audit[2490]: NETFILTER_CFG table=filter:48 family=2 entries=1 op=nft_register_rule pid=2490 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.382000 audit[2490]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7fff3cd99980 a2=0 a3=7fff3cd9996c items=0 ppid=2433 pid=2490 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.382000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 May 8 00:46:31.383000 audit[2491]: NETFILTER_CFG table=filter:49 family=2 entries=1 op=nft_register_chain pid=2491 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.383000 audit[2491]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff0ce3f770 a2=0 a3=7fff0ce3f75c items=0 ppid=2433 pid=2491 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.383000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D740066696C746572 May 8 00:46:31.386000 audit[2493]: NETFILTER_CFG table=filter:50 family=2 entries=1 op=nft_register_rule pid=2493 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.386000 audit[2493]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7fff899a1f00 a2=0 a3=7fff899a1eec items=0 ppid=2433 pid=2493 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.386000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D May 8 00:46:31.389000 audit[2496]: NETFILTER_CFG table=filter:51 family=2 entries=1 op=nft_register_rule pid=2496 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.389000 audit[2496]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffd2d087a80 a2=0 a3=7ffd2d087a6c items=0 ppid=2433 pid=2496 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.389000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D53 May 8 00:46:31.390000 audit[2497]: NETFILTER_CFG table=filter:52 family=2 entries=1 op=nft_register_chain pid=2497 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.390000 audit[2497]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff7b84e1e0 a2=0 a3=7fff7b84e1cc items=0 ppid=2433 pid=2497 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.390000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D464F5257415244002D740066696C746572 May 8 00:46:31.393000 audit[2499]: NETFILTER_CFG table=filter:53 family=2 entries=1 op=nft_register_rule pid=2499 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.393000 audit[2499]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffd8321c370 a2=0 a3=7ffd8321c35c items=0 ppid=2433 pid=2499 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.393000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 May 8 00:46:31.394000 audit[2500]: NETFILTER_CFG table=filter:54 family=2 entries=1 op=nft_register_chain pid=2500 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.394000 audit[2500]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffdd2d97f70 a2=0 a3=7ffdd2d97f5c items=0 ppid=2433 pid=2500 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.394000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 May 8 00:46:31.396000 audit[2502]: NETFILTER_CFG table=filter:55 family=2 entries=1 op=nft_register_rule pid=2502 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.396000 audit[2502]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7fffb620f9c0 a2=0 a3=7fffb620f9ac items=0 ppid=2433 pid=2502 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.396000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A May 8 00:46:31.399000 audit[2505]: NETFILTER_CFG table=filter:56 family=2 entries=1 op=nft_register_rule pid=2505 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.399000 audit[2505]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffc5151ff90 a2=0 a3=7ffc5151ff7c items=0 ppid=2433 pid=2505 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.399000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A May 8 00:46:31.403000 audit[2508]: NETFILTER_CFG table=filter:57 family=2 entries=1 op=nft_register_rule pid=2508 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.403000 audit[2508]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffe8df36b90 a2=0 a3=7ffe8df36b7c items=0 ppid=2433 pid=2508 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.403000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D May 8 00:46:31.405000 audit[2509]: NETFILTER_CFG table=nat:58 family=2 entries=1 op=nft_register_chain pid=2509 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.405000 audit[2509]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffd7d1ea9a0 a2=0 a3=7ffd7d1ea98c items=0 ppid=2433 pid=2509 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.405000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D74006E6174 May 8 00:46:31.407000 audit[2511]: NETFILTER_CFG table=nat:59 family=2 entries=1 op=nft_register_rule pid=2511 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.407000 audit[2511]: SYSCALL arch=c000003e syscall=46 success=yes exit=524 a0=3 a1=7ffe8f3381b0 a2=0 a3=7ffe8f33819c items=0 ppid=2433 pid=2511 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.407000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 May 8 00:46:31.411000 audit[2514]: NETFILTER_CFG table=nat:60 family=2 entries=1 op=nft_register_rule pid=2514 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.411000 audit[2514]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffdca3d89b0 a2=0 a3=7ffdca3d899c items=0 ppid=2433 pid=2514 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.411000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 May 8 00:46:31.412000 audit[2515]: NETFILTER_CFG table=nat:61 family=2 entries=1 op=nft_register_chain pid=2515 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.412000 audit[2515]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffeaba8cf50 a2=0 a3=7ffeaba8cf3c items=0 ppid=2433 pid=2515 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.412000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4E004B5542452D504F5354524F5554494E47002D74006E6174 May 8 00:46:31.415000 audit[2517]: NETFILTER_CFG table=nat:62 family=2 entries=1 op=nft_register_rule pid=2517 subj=system_u:system_r:kernel_t:s0 comm="iptables" May 8 00:46:31.415000 audit[2517]: SYSCALL arch=c000003e syscall=46 success=yes exit=532 a0=3 a1=7fff7a51b950 a2=0 a3=7fff7a51b93c items=0 ppid=2433 pid=2517 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.415000 audit: PROCTITLE proctitle=69707461626C6573002D770035002D5700313030303030002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 May 8 00:46:31.436000 audit[2523]: NETFILTER_CFG table=filter:63 family=2 entries=8 op=nft_register_rule pid=2523 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:31.436000 audit[2523]: SYSCALL arch=c000003e syscall=46 success=yes exit=5164 a0=3 a1=7ffc5a704300 a2=0 a3=7ffc5a7042ec items=0 ppid=2433 pid=2523 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.436000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:31.447000 audit[2523]: NETFILTER_CFG table=nat:64 family=2 entries=14 op=nft_register_chain pid=2523 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:31.447000 audit[2523]: SYSCALL arch=c000003e syscall=46 success=yes exit=5508 a0=3 a1=7ffc5a704300 a2=0 a3=7ffc5a7042ec items=0 ppid=2433 pid=2523 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.447000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:31.449000 audit[2529]: NETFILTER_CFG table=filter:65 family=10 entries=1 op=nft_register_chain pid=2529 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.449000 audit[2529]: SYSCALL arch=c000003e syscall=46 success=yes exit=108 a0=3 a1=7ffed2adc0f0 a2=0 a3=7ffed2adc0dc items=0 ppid=2433 pid=2529 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.449000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D45585445524E414C2D5345525649434553002D740066696C746572 May 8 00:46:31.452000 audit[2531]: NETFILTER_CFG table=filter:66 family=10 entries=2 op=nft_register_chain pid=2531 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.452000 audit[2531]: SYSCALL arch=c000003e syscall=46 success=yes exit=836 a0=3 a1=7fff41917240 a2=0 a3=7fff4191722c items=0 ppid=2433 pid=2531 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.452000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C6520736572766963 May 8 00:46:31.456000 audit[2534]: NETFILTER_CFG table=filter:67 family=10 entries=2 op=nft_register_chain pid=2534 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.456000 audit[2534]: SYSCALL arch=c000003e syscall=46 success=yes exit=836 a0=3 a1=7ffe365e0140 a2=0 a3=7ffe365e012c items=0 ppid=2433 pid=2534 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.456000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E657465732065787465726E616C6C792D76697369626C652073657276 May 8 00:46:31.457000 audit[2535]: NETFILTER_CFG table=filter:68 family=10 entries=1 op=nft_register_chain pid=2535 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.457000 audit[2535]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffefcf2ecd0 a2=0 a3=7ffefcf2ecbc items=0 ppid=2433 pid=2535 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.457000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4E4F4445504F525453002D740066696C746572 May 8 00:46:31.460000 audit[2537]: NETFILTER_CFG table=filter:69 family=10 entries=1 op=nft_register_rule pid=2537 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.460000 audit[2537]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffc68970c90 a2=0 a3=7ffc68970c7c items=0 ppid=2433 pid=2537 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.460000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206865616C746820636865636B207365727669636520706F727473002D6A004B5542452D4E4F4445504F525453 May 8 00:46:31.461000 audit[2538]: NETFILTER_CFG table=filter:70 family=10 entries=1 op=nft_register_chain pid=2538 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.461000 audit[2538]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7fff0b552ad0 a2=0 a3=7fff0b552abc items=0 ppid=2433 pid=2538 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.461000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D740066696C746572 May 8 00:46:31.464000 audit[2540]: NETFILTER_CFG table=filter:71 family=10 entries=1 op=nft_register_rule pid=2540 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.464000 audit[2540]: SYSCALL arch=c000003e syscall=46 success=yes exit=744 a0=3 a1=7ffdfcd588a0 a2=0 a3=7ffdfcd5888c items=0 ppid=2433 pid=2540 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.464000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B554245 May 8 00:46:31.467000 audit[2543]: NETFILTER_CFG table=filter:72 family=10 entries=2 op=nft_register_chain pid=2543 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.467000 audit[2543]: SYSCALL arch=c000003e syscall=46 success=yes exit=828 a0=3 a1=7fff09a3ea90 a2=0 a3=7fff09a3ea7c items=0 ppid=2433 pid=2543 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.467000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D May 8 00:46:31.468000 audit[2544]: NETFILTER_CFG table=filter:73 family=10 entries=1 op=nft_register_chain pid=2544 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.468000 audit[2544]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffd21cf9260 a2=0 a3=7ffd21cf924c items=0 ppid=2433 pid=2544 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.468000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D464F5257415244002D740066696C746572 May 8 00:46:31.470000 audit[2546]: NETFILTER_CFG table=filter:74 family=10 entries=1 op=nft_register_rule pid=2546 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.470000 audit[2546]: SYSCALL arch=c000003e syscall=46 success=yes exit=528 a0=3 a1=7ffcbfe616a0 a2=0 a3=7ffcbfe6168c items=0 ppid=2433 pid=2546 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.470000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320666F7277617264696E672072756C6573002D6A004B5542452D464F5257415244 May 8 00:46:31.471000 audit[2547]: NETFILTER_CFG table=filter:75 family=10 entries=1 op=nft_register_chain pid=2547 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.471000 audit[2547]: SYSCALL arch=c000003e syscall=46 success=yes exit=104 a0=3 a1=7ffebd5e7260 a2=0 a3=7ffebd5e724c items=0 ppid=2433 pid=2547 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.471000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D50524F58592D4649524557414C4C002D740066696C746572 May 8 00:46:31.474000 audit[2549]: NETFILTER_CFG table=filter:76 family=10 entries=1 op=nft_register_rule pid=2549 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.474000 audit[2549]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7fff494ef090 a2=0 a3=7fff494ef07c items=0 ppid=2433 pid=2549 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.474000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D6A May 8 00:46:31.478000 audit[2552]: NETFILTER_CFG table=filter:77 family=10 entries=1 op=nft_register_rule pid=2552 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.478000 audit[2552]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffe438b4ee0 a2=0 a3=7ffe438b4ecc items=0 ppid=2433 pid=2552 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.478000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C002D May 8 00:46:31.481000 audit[2555]: NETFILTER_CFG table=filter:78 family=10 entries=1 op=nft_register_rule pid=2555 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.481000 audit[2555]: SYSCALL arch=c000003e syscall=46 success=yes exit=748 a0=3 a1=7ffc3c0ba5c0 a2=0 a3=7ffc3c0ba5ac items=0 ppid=2433 pid=2555 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.481000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900464F5257415244002D740066696C746572002D6D00636F6E6E747261636B002D2D63747374617465004E4557002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573206C6F61642062616C616E636572206669726577616C6C May 8 00:46:31.482000 audit[2556]: NETFILTER_CFG table=nat:79 family=10 entries=1 op=nft_register_chain pid=2556 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.482000 audit[2556]: SYSCALL arch=c000003e syscall=46 success=yes exit=96 a0=3 a1=7ffe0c1d48b0 a2=0 a3=7ffe0c1d489c items=0 ppid=2433 pid=2556 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.482000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D5345525649434553002D74006E6174 May 8 00:46:31.485000 audit[2558]: NETFILTER_CFG table=nat:80 family=10 entries=2 op=nft_register_chain pid=2558 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.485000 audit[2558]: SYSCALL arch=c000003e syscall=46 success=yes exit=600 a0=3 a1=7fff165a44a0 a2=0 a3=7fff165a448c items=0 ppid=2433 pid=2558 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.485000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 May 8 00:46:31.488000 audit[2561]: NETFILTER_CFG table=nat:81 family=10 entries=2 op=nft_register_chain pid=2561 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.488000 audit[2561]: SYSCALL arch=c000003e syscall=46 success=yes exit=608 a0=3 a1=7fffb29cc430 a2=0 a3=7fffb29cc41c items=0 ppid=2433 pid=2561 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.488000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900505245524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E65746573207365727669636520706F7274616C73002D6A004B5542452D5345525649434553 May 8 00:46:31.489000 audit[2562]: NETFILTER_CFG table=nat:82 family=10 entries=1 op=nft_register_chain pid=2562 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.489000 audit[2562]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffef1a9a5e0 a2=0 a3=7ffef1a9a5cc items=0 ppid=2433 pid=2562 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.489000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D504F5354524F5554494E47002D74006E6174 May 8 00:46:31.492000 audit[2564]: NETFILTER_CFG table=nat:83 family=10 entries=2 op=nft_register_chain pid=2564 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.492000 audit[2564]: SYSCALL arch=c000003e syscall=46 success=yes exit=612 a0=3 a1=7fff3abc1180 a2=0 a3=7fff3abc116c items=0 ppid=2433 pid=2564 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.492000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900504F5354524F5554494E47002D74006E6174002D6D00636F6D6D656E74002D2D636F6D6D656E74006B756265726E6574657320706F7374726F7574696E672072756C6573002D6A004B5542452D504F5354524F5554494E47 May 8 00:46:31.493000 audit[2565]: NETFILTER_CFG table=filter:84 family=10 entries=1 op=nft_register_chain pid=2565 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.493000 audit[2565]: SYSCALL arch=c000003e syscall=46 success=yes exit=100 a0=3 a1=7ffcb048da50 a2=0 a3=7ffcb048da3c items=0 ppid=2433 pid=2565 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.493000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4E004B5542452D4649524557414C4C002D740066696C746572 May 8 00:46:31.496000 audit[2567]: NETFILTER_CFG table=filter:85 family=10 entries=1 op=nft_register_rule pid=2567 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.496000 audit[2567]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7ffe43876950 a2=0 a3=7ffe4387693c items=0 ppid=2433 pid=2567 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.496000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D4900494E505554002D740066696C746572002D6A004B5542452D4649524557414C4C May 8 00:46:31.499000 audit[2570]: NETFILTER_CFG table=filter:86 family=10 entries=1 op=nft_register_rule pid=2570 subj=system_u:system_r:kernel_t:s0 comm="ip6tables" May 8 00:46:31.499000 audit[2570]: SYSCALL arch=c000003e syscall=46 success=yes exit=228 a0=3 a1=7fff4079d0d0 a2=0 a3=7fff4079d0bc items=0 ppid=2433 pid=2570 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.499000 audit: PROCTITLE proctitle=6970367461626C6573002D770035002D5700313030303030002D49004F5554505554002D740066696C746572002D6A004B5542452D4649524557414C4C May 8 00:46:31.503000 audit[2572]: NETFILTER_CFG table=filter:87 family=10 entries=3 op=nft_register_rule pid=2572 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" May 8 00:46:31.503000 audit[2572]: SYSCALL arch=c000003e syscall=46 success=yes exit=2004 a0=3 a1=7ffcdc69db30 a2=0 a3=7ffcdc69db1c items=0 ppid=2433 pid=2572 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.503000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:31.503000 audit[2572]: NETFILTER_CFG table=nat:88 family=10 entries=7 op=nft_register_chain pid=2572 subj=system_u:system_r:kernel_t:s0 comm="ip6tables-resto" May 8 00:46:31.503000 audit[2572]: SYSCALL arch=c000003e syscall=46 success=yes exit=2056 a0=3 a1=7ffcdc69db30 a2=0 a3=7ffcdc69db1c items=0 ppid=2433 pid=2572 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="ip6tables-resto" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:31.503000 audit: PROCTITLE proctitle=6970367461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:32.213329 kubelet[2246]: E0508 00:46:32.213291 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:32.283573 kubelet[2246]: I0508 00:46:32.283493 2246 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-h4c2z" podStartSLOduration=2.283468838 podStartE2EDuration="2.283468838s" podCreationTimestamp="2025-05-08 00:46:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 00:46:32.283354563 +0000 UTC m=+16.224169044" watchObservedRunningTime="2025-05-08 00:46:32.283468838 +0000 UTC m=+16.224283309" May 8 00:46:33.058192 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1246788584.mount: Deactivated successfully. May 8 00:46:34.181253 env[1311]: time="2025-05-08T00:46:34.181181521Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/tigera/operator:v1.36.7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:34.226884 env[1311]: time="2025-05-08T00:46:34.226827557Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:34.254634 env[1311]: time="2025-05-08T00:46:34.254566128Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/tigera/operator:v1.36.7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:34.300862 env[1311]: time="2025-05-08T00:46:34.300785471Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:34.301366 env[1311]: time="2025-05-08T00:46:34.301303095Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\" returns image reference \"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\"" May 8 00:46:34.304286 env[1311]: time="2025-05-08T00:46:34.304230349Z" level=info msg="CreateContainer within sandbox \"56ee6c12bb09607ada2c780138ec646ea88bb60199002b37d130f4ff09189391\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" May 8 00:46:34.976806 env[1311]: time="2025-05-08T00:46:34.976610569Z" level=info msg="CreateContainer within sandbox \"56ee6c12bb09607ada2c780138ec646ea88bb60199002b37d130f4ff09189391\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"ae658473d1aa76a95956743bed366905505004986902a579f18eaecdd8f30f31\"" May 8 00:46:34.977563 env[1311]: time="2025-05-08T00:46:34.977481837Z" level=info msg="StartContainer for \"ae658473d1aa76a95956743bed366905505004986902a579f18eaecdd8f30f31\"" May 8 00:46:35.096875 env[1311]: time="2025-05-08T00:46:35.096788579Z" level=info msg="StartContainer for \"ae658473d1aa76a95956743bed366905505004986902a579f18eaecdd8f30f31\" returns successfully" May 8 00:46:35.232100 kubelet[2246]: I0508 00:46:35.231699 2246 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-797db67f8-wvrpq" podStartSLOduration=1.851585677 podStartE2EDuration="5.231669165s" podCreationTimestamp="2025-05-08 00:46:30 +0000 UTC" firstStartedPulling="2025-05-08 00:46:30.922517557 +0000 UTC m=+14.863332018" lastFinishedPulling="2025-05-08 00:46:34.302601035 +0000 UTC m=+18.243415506" observedRunningTime="2025-05-08 00:46:35.231640301 +0000 UTC m=+19.172454772" watchObservedRunningTime="2025-05-08 00:46:35.231669165 +0000 UTC m=+19.172483717" May 8 00:46:37.724000 audit[2612]: NETFILTER_CFG table=filter:89 family=2 entries=15 op=nft_register_rule pid=2612 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:37.726290 kernel: kauditd_printk_skb: 143 callbacks suppressed May 8 00:46:37.726378 kernel: audit: type=1325 audit(1746665197.724:281): table=filter:89 family=2 entries=15 op=nft_register_rule pid=2612 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:37.724000 audit[2612]: SYSCALL arch=c000003e syscall=46 success=yes exit=5908 a0=3 a1=7ffd428b7990 a2=0 a3=7ffd428b797c items=0 ppid=2433 pid=2612 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:37.734415 kernel: audit: type=1300 audit(1746665197.724:281): arch=c000003e syscall=46 success=yes exit=5908 a0=3 a1=7ffd428b7990 a2=0 a3=7ffd428b797c items=0 ppid=2433 pid=2612 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:37.734491 kernel: audit: type=1327 audit(1746665197.724:281): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:37.724000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:37.736000 audit[2612]: NETFILTER_CFG table=nat:90 family=2 entries=12 op=nft_register_rule pid=2612 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:37.736000 audit[2612]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd428b7990 a2=0 a3=0 items=0 ppid=2433 pid=2612 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:37.745331 kernel: audit: type=1325 audit(1746665197.736:282): table=nat:90 family=2 entries=12 op=nft_register_rule pid=2612 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:37.745393 kernel: audit: type=1300 audit(1746665197.736:282): arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffd428b7990 a2=0 a3=0 items=0 ppid=2433 pid=2612 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:37.745426 kernel: audit: type=1327 audit(1746665197.736:282): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:37.736000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:37.751000 audit[2614]: NETFILTER_CFG table=filter:91 family=2 entries=16 op=nft_register_rule pid=2614 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:37.751000 audit[2614]: SYSCALL arch=c000003e syscall=46 success=yes exit=5908 a0=3 a1=7ffee8f60390 a2=0 a3=7ffee8f6037c items=0 ppid=2433 pid=2614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:37.761009 kernel: audit: type=1325 audit(1746665197.751:283): table=filter:91 family=2 entries=16 op=nft_register_rule pid=2614 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:37.761100 kernel: audit: type=1300 audit(1746665197.751:283): arch=c000003e syscall=46 success=yes exit=5908 a0=3 a1=7ffee8f60390 a2=0 a3=7ffee8f6037c items=0 ppid=2433 pid=2614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:37.761139 kernel: audit: type=1327 audit(1746665197.751:283): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:37.751000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:37.766000 audit[2614]: NETFILTER_CFG table=nat:92 family=2 entries=12 op=nft_register_rule pid=2614 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:37.766000 audit[2614]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffee8f60390 a2=0 a3=0 items=0 ppid=2433 pid=2614 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:37.766000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:37.771130 kernel: audit: type=1325 audit(1746665197.766:284): table=nat:92 family=2 entries=12 op=nft_register_rule pid=2614 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:38.404845 kubelet[2246]: I0508 00:46:38.404774 2246 topology_manager.go:215] "Topology Admit Handler" podUID="95f4b4a0-bb2c-4370-8974-f04adce868c6" podNamespace="calico-system" podName="calico-typha-5fd7dd7db7-5ddd9" May 8 00:46:38.500958 kubelet[2246]: I0508 00:46:38.500887 2246 topology_manager.go:215] "Topology Admit Handler" podUID="ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" podNamespace="calico-system" podName="calico-node-k68b7" May 8 00:46:38.506241 kubelet[2246]: W0508 00:46:38.506165 2246 reflector.go:547] object-"calico-system"/"node-certs": failed to list *v1.Secret: secrets "node-certs" is forbidden: User "system:node:localhost" cannot list resource "secrets" in API group "" in the namespace "calico-system": no relationship found between node 'localhost' and this object May 8 00:46:38.506241 kubelet[2246]: E0508 00:46:38.506246 2246 reflector.go:150] object-"calico-system"/"node-certs": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets "node-certs" is forbidden: User "system:node:localhost" cannot list resource "secrets" in API group "" in the namespace "calico-system": no relationship found between node 'localhost' and this object May 8 00:46:38.507099 kubelet[2246]: W0508 00:46:38.507048 2246 reflector.go:547] object-"calico-system"/"cni-config": failed to list *v1.ConfigMap: configmaps "cni-config" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "calico-system": no relationship found between node 'localhost' and this object May 8 00:46:38.507198 kubelet[2246]: E0508 00:46:38.507120 2246 reflector.go:150] object-"calico-system"/"cni-config": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "cni-config" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "calico-system": no relationship found between node 'localhost' and this object May 8 00:46:38.599177 kubelet[2246]: I0508 00:46:38.599125 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/95f4b4a0-bb2c-4370-8974-f04adce868c6-typha-certs\") pod \"calico-typha-5fd7dd7db7-5ddd9\" (UID: \"95f4b4a0-bb2c-4370-8974-f04adce868c6\") " pod="calico-system/calico-typha-5fd7dd7db7-5ddd9" May 8 00:46:38.599360 kubelet[2246]: I0508 00:46:38.599198 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgd2w\" (UniqueName: \"kubernetes.io/projected/95f4b4a0-bb2c-4370-8974-f04adce868c6-kube-api-access-rgd2w\") pod \"calico-typha-5fd7dd7db7-5ddd9\" (UID: \"95f4b4a0-bb2c-4370-8974-f04adce868c6\") " pod="calico-system/calico-typha-5fd7dd7db7-5ddd9" May 8 00:46:38.599360 kubelet[2246]: I0508 00:46:38.599226 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/95f4b4a0-bb2c-4370-8974-f04adce868c6-tigera-ca-bundle\") pod \"calico-typha-5fd7dd7db7-5ddd9\" (UID: \"95f4b4a0-bb2c-4370-8974-f04adce868c6\") " pod="calico-system/calico-typha-5fd7dd7db7-5ddd9" May 8 00:46:38.674947 kubelet[2246]: I0508 00:46:38.674818 2246 topology_manager.go:215] "Topology Admit Handler" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" podNamespace="calico-system" podName="csi-node-driver-2s6tt" May 8 00:46:38.675469 kubelet[2246]: E0508 00:46:38.675443 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2s6tt" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" May 8 00:46:38.699389 kubelet[2246]: I0508 00:46:38.699332 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-lib-modules\") pod \"calico-node-k68b7\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " pod="calico-system/calico-node-k68b7" May 8 00:46:38.699580 kubelet[2246]: I0508 00:46:38.699383 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-xtables-lock\") pod \"calico-node-k68b7\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " pod="calico-system/calico-node-k68b7" May 8 00:46:38.699580 kubelet[2246]: I0508 00:46:38.699477 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-policysync\") pod \"calico-node-k68b7\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " pod="calico-system/calico-node-k68b7" May 8 00:46:38.699580 kubelet[2246]: I0508 00:46:38.699512 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/863a2a82-a51b-4afc-a7be-990ee722b050-socket-dir\") pod \"csi-node-driver-2s6tt\" (UID: \"863a2a82-a51b-4afc-a7be-990ee722b050\") " pod="calico-system/csi-node-driver-2s6tt" May 8 00:46:38.699580 kubelet[2246]: I0508 00:46:38.699560 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-var-run-calico\") pod \"calico-node-k68b7\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " pod="calico-system/calico-node-k68b7" May 8 00:46:38.699688 kubelet[2246]: I0508 00:46:38.699579 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-cni-bin-dir\") pod \"calico-node-k68b7\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " pod="calico-system/calico-node-k68b7" May 8 00:46:38.699688 kubelet[2246]: I0508 00:46:38.699602 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2hmk\" (UniqueName: \"kubernetes.io/projected/863a2a82-a51b-4afc-a7be-990ee722b050-kube-api-access-l2hmk\") pod \"csi-node-driver-2s6tt\" (UID: \"863a2a82-a51b-4afc-a7be-990ee722b050\") " pod="calico-system/csi-node-driver-2s6tt" May 8 00:46:38.699688 kubelet[2246]: I0508 00:46:38.699625 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-cni-net-dir\") pod \"calico-node-k68b7\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " pod="calico-system/calico-node-k68b7" May 8 00:46:38.699688 kubelet[2246]: I0508 00:46:38.699646 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-flexvol-driver-host\") pod \"calico-node-k68b7\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " pod="calico-system/calico-node-k68b7" May 8 00:46:38.699688 kubelet[2246]: I0508 00:46:38.699661 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-node-certs\") pod \"calico-node-k68b7\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " pod="calico-system/calico-node-k68b7" May 8 00:46:38.699805 kubelet[2246]: I0508 00:46:38.699673 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-cni-log-dir\") pod \"calico-node-k68b7\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " pod="calico-system/calico-node-k68b7" May 8 00:46:38.699805 kubelet[2246]: I0508 00:46:38.699688 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/863a2a82-a51b-4afc-a7be-990ee722b050-registration-dir\") pod \"csi-node-driver-2s6tt\" (UID: \"863a2a82-a51b-4afc-a7be-990ee722b050\") " pod="calico-system/csi-node-driver-2s6tt" May 8 00:46:38.699805 kubelet[2246]: I0508 00:46:38.699704 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-var-lib-calico\") pod \"calico-node-k68b7\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " pod="calico-system/calico-node-k68b7" May 8 00:46:38.699805 kubelet[2246]: I0508 00:46:38.699727 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-tigera-ca-bundle\") pod \"calico-node-k68b7\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " pod="calico-system/calico-node-k68b7" May 8 00:46:38.699805 kubelet[2246]: I0508 00:46:38.699742 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbkxf\" (UniqueName: \"kubernetes.io/projected/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-kube-api-access-qbkxf\") pod \"calico-node-k68b7\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " pod="calico-system/calico-node-k68b7" May 8 00:46:38.699919 kubelet[2246]: I0508 00:46:38.699757 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/863a2a82-a51b-4afc-a7be-990ee722b050-varrun\") pod \"csi-node-driver-2s6tt\" (UID: \"863a2a82-a51b-4afc-a7be-990ee722b050\") " pod="calico-system/csi-node-driver-2s6tt" May 8 00:46:38.699919 kubelet[2246]: I0508 00:46:38.699770 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/863a2a82-a51b-4afc-a7be-990ee722b050-kubelet-dir\") pod \"csi-node-driver-2s6tt\" (UID: \"863a2a82-a51b-4afc-a7be-990ee722b050\") " pod="calico-system/csi-node-driver-2s6tt" May 8 00:46:38.779000 audit[2618]: NETFILTER_CFG table=filter:93 family=2 entries=17 op=nft_register_rule pid=2618 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:38.779000 audit[2618]: SYSCALL arch=c000003e syscall=46 success=yes exit=6652 a0=3 a1=7ffc30172cb0 a2=0 a3=7ffc30172c9c items=0 ppid=2433 pid=2618 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:38.779000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:38.784000 audit[2618]: NETFILTER_CFG table=nat:94 family=2 entries=12 op=nft_register_rule pid=2618 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:38.784000 audit[2618]: SYSCALL arch=c000003e syscall=46 success=yes exit=2700 a0=3 a1=7ffc30172cb0 a2=0 a3=0 items=0 ppid=2433 pid=2618 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:38.784000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:38.801371 kubelet[2246]: E0508 00:46:38.801312 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:38.801371 kubelet[2246]: W0508 00:46:38.801344 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:38.801371 kubelet[2246]: E0508 00:46:38.801383 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:38.801673 kubelet[2246]: E0508 00:46:38.801653 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:38.801673 kubelet[2246]: W0508 00:46:38.801669 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:38.801777 kubelet[2246]: E0508 00:46:38.801680 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:38.812914 kubelet[2246]: E0508 00:46:38.812850 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:38.812914 kubelet[2246]: W0508 00:46:38.812874 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:38.812914 kubelet[2246]: E0508 00:46:38.812902 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:38.813212 kubelet[2246]: E0508 00:46:38.813092 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:38.813212 kubelet[2246]: W0508 00:46:38.813100 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:38.813212 kubelet[2246]: E0508 00:46:38.813136 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:38.901704 kubelet[2246]: E0508 00:46:38.901650 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:38.901704 kubelet[2246]: W0508 00:46:38.901683 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:38.901704 kubelet[2246]: E0508 00:46:38.901712 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:39.002959 kubelet[2246]: E0508 00:46:39.002925 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:39.002959 kubelet[2246]: W0508 00:46:39.002947 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:39.003188 kubelet[2246]: E0508 00:46:39.002969 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:39.013190 kubelet[2246]: E0508 00:46:39.013137 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:39.014064 env[1311]: time="2025-05-08T00:46:39.014013535Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5fd7dd7db7-5ddd9,Uid:95f4b4a0-bb2c-4370-8974-f04adce868c6,Namespace:calico-system,Attempt:0,}" May 8 00:46:39.103730 kubelet[2246]: E0508 00:46:39.103684 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:39.103730 kubelet[2246]: W0508 00:46:39.103705 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:39.103730 kubelet[2246]: E0508 00:46:39.103726 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:39.204333 kubelet[2246]: E0508 00:46:39.204286 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:39.204333 kubelet[2246]: W0508 00:46:39.204309 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:39.204333 kubelet[2246]: E0508 00:46:39.204332 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:39.224408 env[1311]: time="2025-05-08T00:46:39.224304783Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 8 00:46:39.224408 env[1311]: time="2025-05-08T00:46:39.224364856Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 8 00:46:39.224408 env[1311]: time="2025-05-08T00:46:39.224374685Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 8 00:46:39.224667 env[1311]: time="2025-05-08T00:46:39.224558900Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/d208322a247871826dd30efedd9ee93ab4548c5237bacd79a8ae19d6a2b118c4 pid=2635 runtime=io.containerd.runc.v2 May 8 00:46:39.275167 env[1311]: time="2025-05-08T00:46:39.274999356Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5fd7dd7db7-5ddd9,Uid:95f4b4a0-bb2c-4370-8974-f04adce868c6,Namespace:calico-system,Attempt:0,} returns sandbox id \"d208322a247871826dd30efedd9ee93ab4548c5237bacd79a8ae19d6a2b118c4\"" May 8 00:46:39.278338 kubelet[2246]: E0508 00:46:39.278090 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:39.281229 env[1311]: time="2025-05-08T00:46:39.281193966Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\"" May 8 00:46:39.305078 kubelet[2246]: E0508 00:46:39.305030 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:39.305078 kubelet[2246]: W0508 00:46:39.305064 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:39.305303 kubelet[2246]: E0508 00:46:39.305094 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:39.323520 kubelet[2246]: E0508 00:46:39.323482 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:39.323520 kubelet[2246]: W0508 00:46:39.323507 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:39.323520 kubelet[2246]: E0508 00:46:39.323528 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:39.408191 kubelet[2246]: E0508 00:46:39.408038 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:39.408676 env[1311]: time="2025-05-08T00:46:39.408628544Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-k68b7,Uid:ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5,Namespace:calico-system,Attempt:0,}" May 8 00:46:39.427589 env[1311]: time="2025-05-08T00:46:39.427509815Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 8 00:46:39.427589 env[1311]: time="2025-05-08T00:46:39.427556933Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 8 00:46:39.427589 env[1311]: time="2025-05-08T00:46:39.427579706Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 8 00:46:39.427832 env[1311]: time="2025-05-08T00:46:39.427757481Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808 pid=2678 runtime=io.containerd.runc.v2 May 8 00:46:39.464327 env[1311]: time="2025-05-08T00:46:39.464249243Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-k68b7,Uid:ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5,Namespace:calico-system,Attempt:0,} returns sandbox id \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\"" May 8 00:46:39.470668 kubelet[2246]: E0508 00:46:39.470643 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:40.165504 kubelet[2246]: E0508 00:46:40.165431 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2s6tt" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" May 8 00:46:41.614407 env[1311]: time="2025-05-08T00:46:41.614301085Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/typha:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:41.616598 env[1311]: time="2025-05-08T00:46:41.616547105Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:41.618661 env[1311]: time="2025-05-08T00:46:41.618617234Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/typha:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:41.620537 env[1311]: time="2025-05-08T00:46:41.620471348Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:41.620907 env[1311]: time="2025-05-08T00:46:41.620866510Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\" returns image reference \"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\"" May 8 00:46:41.622512 env[1311]: time="2025-05-08T00:46:41.622467708Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\"" May 8 00:46:41.634983 env[1311]: time="2025-05-08T00:46:41.634923068Z" level=info msg="CreateContainer within sandbox \"d208322a247871826dd30efedd9ee93ab4548c5237bacd79a8ae19d6a2b118c4\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" May 8 00:46:41.885703 env[1311]: time="2025-05-08T00:46:41.885515324Z" level=info msg="CreateContainer within sandbox \"d208322a247871826dd30efedd9ee93ab4548c5237bacd79a8ae19d6a2b118c4\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"d19a08ab43ef76bf1c3f5a5ded406187b07ad30710250e19211caf8fc00d087a\"" May 8 00:46:41.886417 env[1311]: time="2025-05-08T00:46:41.886381210Z" level=info msg="StartContainer for \"d19a08ab43ef76bf1c3f5a5ded406187b07ad30710250e19211caf8fc00d087a\"" May 8 00:46:42.038581 env[1311]: time="2025-05-08T00:46:42.038494994Z" level=info msg="StartContainer for \"d19a08ab43ef76bf1c3f5a5ded406187b07ad30710250e19211caf8fc00d087a\" returns successfully" May 8 00:46:42.165370 kubelet[2246]: E0508 00:46:42.165236 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2s6tt" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" May 8 00:46:42.239351 kubelet[2246]: E0508 00:46:42.239314 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:42.325625 kubelet[2246]: E0508 00:46:42.325586 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.325625 kubelet[2246]: W0508 00:46:42.325614 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.325845 kubelet[2246]: E0508 00:46:42.325645 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.325910 kubelet[2246]: E0508 00:46:42.325898 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.325951 kubelet[2246]: W0508 00:46:42.325910 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.325951 kubelet[2246]: E0508 00:46:42.325925 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.326172 kubelet[2246]: E0508 00:46:42.326155 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.326172 kubelet[2246]: W0508 00:46:42.326170 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.326274 kubelet[2246]: E0508 00:46:42.326186 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.326456 kubelet[2246]: E0508 00:46:42.326436 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.326456 kubelet[2246]: W0508 00:46:42.326449 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.326562 kubelet[2246]: E0508 00:46:42.326468 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.326752 kubelet[2246]: E0508 00:46:42.326735 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.326752 kubelet[2246]: W0508 00:46:42.326749 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.326853 kubelet[2246]: E0508 00:46:42.326764 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.326945 kubelet[2246]: E0508 00:46:42.326929 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.326945 kubelet[2246]: W0508 00:46:42.326941 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.327043 kubelet[2246]: E0508 00:46:42.326951 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.327163 kubelet[2246]: E0508 00:46:42.327145 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.327163 kubelet[2246]: W0508 00:46:42.327158 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.327266 kubelet[2246]: E0508 00:46:42.327175 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.327381 kubelet[2246]: E0508 00:46:42.327362 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.327381 kubelet[2246]: W0508 00:46:42.327375 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.327477 kubelet[2246]: E0508 00:46:42.327392 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.327599 kubelet[2246]: E0508 00:46:42.327580 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.327599 kubelet[2246]: W0508 00:46:42.327592 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.327697 kubelet[2246]: E0508 00:46:42.327608 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.327800 kubelet[2246]: E0508 00:46:42.327781 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.327800 kubelet[2246]: W0508 00:46:42.327795 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.327920 kubelet[2246]: E0508 00:46:42.327810 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.328008 kubelet[2246]: E0508 00:46:42.327991 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.328008 kubelet[2246]: W0508 00:46:42.328003 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.328098 kubelet[2246]: E0508 00:46:42.328013 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.328187 kubelet[2246]: E0508 00:46:42.328172 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.328223 kubelet[2246]: W0508 00:46:42.328196 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.328223 kubelet[2246]: E0508 00:46:42.328209 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.328413 kubelet[2246]: E0508 00:46:42.328397 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.328413 kubelet[2246]: W0508 00:46:42.328409 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.328503 kubelet[2246]: E0508 00:46:42.328418 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.328576 kubelet[2246]: E0508 00:46:42.328564 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.328614 kubelet[2246]: W0508 00:46:42.328576 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.328614 kubelet[2246]: E0508 00:46:42.328584 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.328756 kubelet[2246]: E0508 00:46:42.328743 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.328756 kubelet[2246]: W0508 00:46:42.328754 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.328846 kubelet[2246]: E0508 00:46:42.328763 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.425988 kubelet[2246]: E0508 00:46:42.425864 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.425988 kubelet[2246]: W0508 00:46:42.425887 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.425988 kubelet[2246]: E0508 00:46:42.425908 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.426227 kubelet[2246]: E0508 00:46:42.426194 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.426274 kubelet[2246]: W0508 00:46:42.426224 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.426274 kubelet[2246]: E0508 00:46:42.426253 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.426670 kubelet[2246]: E0508 00:46:42.426618 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.426670 kubelet[2246]: W0508 00:46:42.426669 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.426771 kubelet[2246]: E0508 00:46:42.426695 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.427076 kubelet[2246]: E0508 00:46:42.427055 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.427182 kubelet[2246]: W0508 00:46:42.427068 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.427182 kubelet[2246]: E0508 00:46:42.427099 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.427457 kubelet[2246]: E0508 00:46:42.427419 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.427457 kubelet[2246]: W0508 00:46:42.427438 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.427540 kubelet[2246]: E0508 00:46:42.427477 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.427673 kubelet[2246]: E0508 00:46:42.427644 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.427673 kubelet[2246]: W0508 00:46:42.427661 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.427769 kubelet[2246]: E0508 00:46:42.427700 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.427862 kubelet[2246]: E0508 00:46:42.427847 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.427862 kubelet[2246]: W0508 00:46:42.427858 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.427946 kubelet[2246]: E0508 00:46:42.427885 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.428044 kubelet[2246]: E0508 00:46:42.428031 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.428044 kubelet[2246]: W0508 00:46:42.428040 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.428147 kubelet[2246]: E0508 00:46:42.428052 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.428243 kubelet[2246]: E0508 00:46:42.428232 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.428243 kubelet[2246]: W0508 00:46:42.428240 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.428316 kubelet[2246]: E0508 00:46:42.428250 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.428566 kubelet[2246]: E0508 00:46:42.428542 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.428566 kubelet[2246]: W0508 00:46:42.428554 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.428566 kubelet[2246]: E0508 00:46:42.428566 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.428727 kubelet[2246]: E0508 00:46:42.428714 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.428727 kubelet[2246]: W0508 00:46:42.428724 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.428789 kubelet[2246]: E0508 00:46:42.428738 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.429003 kubelet[2246]: E0508 00:46:42.428986 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.429003 kubelet[2246]: W0508 00:46:42.429001 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.429181 kubelet[2246]: E0508 00:46:42.429032 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.429243 kubelet[2246]: E0508 00:46:42.429221 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.429243 kubelet[2246]: W0508 00:46:42.429234 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.429328 kubelet[2246]: E0508 00:46:42.429282 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.429431 kubelet[2246]: E0508 00:46:42.429414 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.429431 kubelet[2246]: W0508 00:46:42.429425 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.429524 kubelet[2246]: E0508 00:46:42.429441 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.429641 kubelet[2246]: E0508 00:46:42.429628 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.429670 kubelet[2246]: W0508 00:46:42.429640 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.429670 kubelet[2246]: E0508 00:46:42.429653 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.429825 kubelet[2246]: E0508 00:46:42.429811 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.429825 kubelet[2246]: W0508 00:46:42.429824 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.429916 kubelet[2246]: E0508 00:46:42.429833 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.430176 kubelet[2246]: E0508 00:46:42.430042 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.430176 kubelet[2246]: W0508 00:46:42.430065 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.430176 kubelet[2246]: E0508 00:46:42.430078 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.430847 kubelet[2246]: E0508 00:46:42.430821 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:42.430847 kubelet[2246]: W0508 00:46:42.430838 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:42.430908 kubelet[2246]: E0508 00:46:42.430849 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:42.462827 kubelet[2246]: I0508 00:46:42.462720 2246 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-5fd7dd7db7-5ddd9" podStartSLOduration=2.121593543 podStartE2EDuration="4.46269463s" podCreationTimestamp="2025-05-08 00:46:38 +0000 UTC" firstStartedPulling="2025-05-08 00:46:39.280921915 +0000 UTC m=+23.221736386" lastFinishedPulling="2025-05-08 00:46:41.622022992 +0000 UTC m=+25.562837473" observedRunningTime="2025-05-08 00:46:42.46086291 +0000 UTC m=+26.401677401" watchObservedRunningTime="2025-05-08 00:46:42.46269463 +0000 UTC m=+26.403509101" May 8 00:46:43.241094 kubelet[2246]: I0508 00:46:43.241055 2246 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 8 00:46:43.241811 kubelet[2246]: E0508 00:46:43.241779 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:43.333159 kubelet[2246]: E0508 00:46:43.333069 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.333159 kubelet[2246]: W0508 00:46:43.333148 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.333483 kubelet[2246]: E0508 00:46:43.333175 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.333483 kubelet[2246]: E0508 00:46:43.333409 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.333483 kubelet[2246]: W0508 00:46:43.333418 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.333483 kubelet[2246]: E0508 00:46:43.333428 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.334035 kubelet[2246]: E0508 00:46:43.334008 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.334035 kubelet[2246]: W0508 00:46:43.334023 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.334035 kubelet[2246]: E0508 00:46:43.334033 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.334292 kubelet[2246]: E0508 00:46:43.334273 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.334292 kubelet[2246]: W0508 00:46:43.334289 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.334389 kubelet[2246]: E0508 00:46:43.334300 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.334511 kubelet[2246]: E0508 00:46:43.334488 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.334511 kubelet[2246]: W0508 00:46:43.334498 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.334511 kubelet[2246]: E0508 00:46:43.334505 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.334756 kubelet[2246]: E0508 00:46:43.334664 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.334756 kubelet[2246]: W0508 00:46:43.334671 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.334756 kubelet[2246]: E0508 00:46:43.334678 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.334870 kubelet[2246]: E0508 00:46:43.334854 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.334870 kubelet[2246]: W0508 00:46:43.334869 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.334969 kubelet[2246]: E0508 00:46:43.334880 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.335128 kubelet[2246]: E0508 00:46:43.335088 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.335206 kubelet[2246]: W0508 00:46:43.335102 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.335206 kubelet[2246]: E0508 00:46:43.335147 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.335361 kubelet[2246]: E0508 00:46:43.335343 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.335361 kubelet[2246]: W0508 00:46:43.335356 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.335361 kubelet[2246]: E0508 00:46:43.335366 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.335600 kubelet[2246]: E0508 00:46:43.335581 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.335600 kubelet[2246]: W0508 00:46:43.335594 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.335704 kubelet[2246]: E0508 00:46:43.335610 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.335834 kubelet[2246]: E0508 00:46:43.335818 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.335834 kubelet[2246]: W0508 00:46:43.335830 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.335956 kubelet[2246]: E0508 00:46:43.335841 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.336132 kubelet[2246]: E0508 00:46:43.336082 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.336132 kubelet[2246]: W0508 00:46:43.336118 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.336232 kubelet[2246]: E0508 00:46:43.336148 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.336393 kubelet[2246]: E0508 00:46:43.336367 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.336393 kubelet[2246]: W0508 00:46:43.336379 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.336393 kubelet[2246]: E0508 00:46:43.336387 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.336568 kubelet[2246]: E0508 00:46:43.336534 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.336568 kubelet[2246]: W0508 00:46:43.336545 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.336568 kubelet[2246]: E0508 00:46:43.336553 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.336721 kubelet[2246]: E0508 00:46:43.336706 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.336721 kubelet[2246]: W0508 00:46:43.336716 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.336721 kubelet[2246]: E0508 00:46:43.336723 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.433910 kubelet[2246]: E0508 00:46:43.433859 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.433910 kubelet[2246]: W0508 00:46:43.433890 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.433910 kubelet[2246]: E0508 00:46:43.433912 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.434245 kubelet[2246]: E0508 00:46:43.434095 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.434245 kubelet[2246]: W0508 00:46:43.434127 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.434245 kubelet[2246]: E0508 00:46:43.434148 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.434540 kubelet[2246]: E0508 00:46:43.434493 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.434540 kubelet[2246]: W0508 00:46:43.434529 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.434644 kubelet[2246]: E0508 00:46:43.434564 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.434800 kubelet[2246]: E0508 00:46:43.434772 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.434800 kubelet[2246]: W0508 00:46:43.434787 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.434900 kubelet[2246]: E0508 00:46:43.434809 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.435034 kubelet[2246]: E0508 00:46:43.435015 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.435034 kubelet[2246]: W0508 00:46:43.435028 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.435162 kubelet[2246]: E0508 00:46:43.435054 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.435266 kubelet[2246]: E0508 00:46:43.435250 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.435322 kubelet[2246]: W0508 00:46:43.435267 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.435322 kubelet[2246]: E0508 00:46:43.435288 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.435464 kubelet[2246]: E0508 00:46:43.435445 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.435464 kubelet[2246]: W0508 00:46:43.435458 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.435562 kubelet[2246]: E0508 00:46:43.435491 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.435637 kubelet[2246]: E0508 00:46:43.435622 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.435637 kubelet[2246]: W0508 00:46:43.435633 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.435727 kubelet[2246]: E0508 00:46:43.435661 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.435798 kubelet[2246]: E0508 00:46:43.435783 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.435798 kubelet[2246]: W0508 00:46:43.435794 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.435877 kubelet[2246]: E0508 00:46:43.435819 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.435998 kubelet[2246]: E0508 00:46:43.435981 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.435998 kubelet[2246]: W0508 00:46:43.435993 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.436089 kubelet[2246]: E0508 00:46:43.436010 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.436310 kubelet[2246]: E0508 00:46:43.436287 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.436310 kubelet[2246]: W0508 00:46:43.436302 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.436405 kubelet[2246]: E0508 00:46:43.436317 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.436498 kubelet[2246]: E0508 00:46:43.436480 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.436498 kubelet[2246]: W0508 00:46:43.436491 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.436600 kubelet[2246]: E0508 00:46:43.436504 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.436729 kubelet[2246]: E0508 00:46:43.436712 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.436729 kubelet[2246]: W0508 00:46:43.436722 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.436807 kubelet[2246]: E0508 00:46:43.436734 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.436992 kubelet[2246]: E0508 00:46:43.436975 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.436992 kubelet[2246]: W0508 00:46:43.436987 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.437083 kubelet[2246]: E0508 00:46:43.437000 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.437200 kubelet[2246]: E0508 00:46:43.437187 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.437200 kubelet[2246]: W0508 00:46:43.437196 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.437306 kubelet[2246]: E0508 00:46:43.437208 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.437489 kubelet[2246]: E0508 00:46:43.437467 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.437489 kubelet[2246]: W0508 00:46:43.437483 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.437581 kubelet[2246]: E0508 00:46:43.437498 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.437673 kubelet[2246]: E0508 00:46:43.437657 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.437673 kubelet[2246]: W0508 00:46:43.437669 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.437741 kubelet[2246]: E0508 00:46:43.437682 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:43.437852 kubelet[2246]: E0508 00:46:43.437836 2246 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 8 00:46:43.437852 kubelet[2246]: W0508 00:46:43.437845 2246 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 8 00:46:43.437852 kubelet[2246]: E0508 00:46:43.437852 2246 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 8 00:46:44.095952 env[1311]: time="2025-05-08T00:46:44.095880618Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:44.136716 env[1311]: time="2025-05-08T00:46:44.136644040Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:44.145178 env[1311]: time="2025-05-08T00:46:44.145104461Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:44.148459 env[1311]: time="2025-05-08T00:46:44.148428233Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:44.148898 env[1311]: time="2025-05-08T00:46:44.148871966Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" returns image reference \"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\"" May 8 00:46:44.150989 env[1311]: time="2025-05-08T00:46:44.150911968Z" level=info msg="CreateContainer within sandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" May 8 00:46:44.165986 kubelet[2246]: E0508 00:46:44.165909 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2s6tt" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" May 8 00:46:44.192553 env[1311]: time="2025-05-08T00:46:44.192474852Z" level=info msg="CreateContainer within sandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"a3368ce3ed5ee729b44408f60e027497ad2d3647a905935456d9b4022a7ca29c\"" May 8 00:46:44.193256 env[1311]: time="2025-05-08T00:46:44.193157483Z" level=info msg="StartContainer for \"a3368ce3ed5ee729b44408f60e027497ad2d3647a905935456d9b4022a7ca29c\"" May 8 00:46:44.317364 env[1311]: time="2025-05-08T00:46:44.317302783Z" level=info msg="StartContainer for \"a3368ce3ed5ee729b44408f60e027497ad2d3647a905935456d9b4022a7ca29c\" returns successfully" May 8 00:46:44.361371 env[1311]: time="2025-05-08T00:46:44.361178541Z" level=info msg="shim disconnected" id=a3368ce3ed5ee729b44408f60e027497ad2d3647a905935456d9b4022a7ca29c May 8 00:46:44.361371 env[1311]: time="2025-05-08T00:46:44.361243843Z" level=warning msg="cleaning up after shim disconnected" id=a3368ce3ed5ee729b44408f60e027497ad2d3647a905935456d9b4022a7ca29c namespace=k8s.io May 8 00:46:44.361371 env[1311]: time="2025-05-08T00:46:44.361256167Z" level=info msg="cleaning up dead shim" May 8 00:46:44.368551 env[1311]: time="2025-05-08T00:46:44.368475656Z" level=warning msg="cleanup warnings time=\"2025-05-08T00:46:44Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2868 runtime=io.containerd.runc.v2\n" May 8 00:46:45.039759 kubelet[2246]: I0508 00:46:45.039703 2246 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 8 00:46:45.040504 kubelet[2246]: E0508 00:46:45.040471 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:45.070000 audit[2891]: NETFILTER_CFG table=filter:95 family=2 entries=17 op=nft_register_rule pid=2891 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:45.073030 kernel: kauditd_printk_skb: 8 callbacks suppressed May 8 00:46:45.073138 kernel: audit: type=1325 audit(1746665205.070:287): table=filter:95 family=2 entries=17 op=nft_register_rule pid=2891 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:45.070000 audit[2891]: SYSCALL arch=c000003e syscall=46 success=yes exit=5908 a0=3 a1=7ffea4f16b40 a2=0 a3=7ffea4f16b2c items=0 ppid=2433 pid=2891 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:45.081144 kernel: audit: type=1300 audit(1746665205.070:287): arch=c000003e syscall=46 success=yes exit=5908 a0=3 a1=7ffea4f16b40 a2=0 a3=7ffea4f16b2c items=0 ppid=2433 pid=2891 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:45.081221 kernel: audit: type=1327 audit(1746665205.070:287): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:45.070000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:45.086000 audit[2891]: NETFILTER_CFG table=nat:96 family=2 entries=19 op=nft_register_chain pid=2891 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:45.086000 audit[2891]: SYSCALL arch=c000003e syscall=46 success=yes exit=6276 a0=3 a1=7ffea4f16b40 a2=0 a3=7ffea4f16b2c items=0 ppid=2433 pid=2891 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:45.095678 kernel: audit: type=1325 audit(1746665205.086:288): table=nat:96 family=2 entries=19 op=nft_register_chain pid=2891 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:46:45.095778 kernel: audit: type=1300 audit(1746665205.086:288): arch=c000003e syscall=46 success=yes exit=6276 a0=3 a1=7ffea4f16b40 a2=0 a3=7ffea4f16b2c items=0 ppid=2433 pid=2891 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:45.095805 kernel: audit: type=1327 audit(1746665205.086:288): proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:45.086000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:46:45.173849 systemd[1]: run-containerd-runc-k8s.io-a3368ce3ed5ee729b44408f60e027497ad2d3647a905935456d9b4022a7ca29c-runc.gIVmPs.mount: Deactivated successfully. May 8 00:46:45.174062 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a3368ce3ed5ee729b44408f60e027497ad2d3647a905935456d9b4022a7ca29c-rootfs.mount: Deactivated successfully. May 8 00:46:45.247801 kubelet[2246]: E0508 00:46:45.247757 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:45.248151 kubelet[2246]: E0508 00:46:45.248130 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:45.248827 env[1311]: time="2025-05-08T00:46:45.248791476Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\"" May 8 00:46:46.165522 kubelet[2246]: E0508 00:46:46.165466 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2s6tt" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" May 8 00:46:48.165343 kubelet[2246]: E0508 00:46:48.165265 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2s6tt" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" May 8 00:46:50.165805 kubelet[2246]: E0508 00:46:50.165722 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2s6tt" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" May 8 00:46:51.859820 env[1311]: time="2025-05-08T00:46:51.859677639Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/cni:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:51.862067 env[1311]: time="2025-05-08T00:46:51.862011430Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:51.864571 env[1311]: time="2025-05-08T00:46:51.864535098Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/cni:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:51.866434 env[1311]: time="2025-05-08T00:46:51.866401953Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:46:51.867006 env[1311]: time="2025-05-08T00:46:51.866961203Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\" returns image reference \"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\"" May 8 00:46:51.869371 env[1311]: time="2025-05-08T00:46:51.869338084Z" level=info msg="CreateContainer within sandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" May 8 00:46:51.887229 env[1311]: time="2025-05-08T00:46:51.887172075Z" level=info msg="CreateContainer within sandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"0031612f63728aa09c3b77481dc5f2be8245c926a8e17c06641844ae7d393b8e\"" May 8 00:46:51.887674 env[1311]: time="2025-05-08T00:46:51.887643749Z" level=info msg="StartContainer for \"0031612f63728aa09c3b77481dc5f2be8245c926a8e17c06641844ae7d393b8e\"" May 8 00:46:51.938303 env[1311]: time="2025-05-08T00:46:51.938229423Z" level=info msg="StartContainer for \"0031612f63728aa09c3b77481dc5f2be8245c926a8e17c06641844ae7d393b8e\" returns successfully" May 8 00:46:52.165590 kubelet[2246]: E0508 00:46:52.165401 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2s6tt" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" May 8 00:46:52.264539 kubelet[2246]: E0508 00:46:52.264490 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:53.266494 kubelet[2246]: E0508 00:46:53.266437 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:54.166216 kubelet[2246]: E0508 00:46:54.166136 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-2s6tt" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" May 8 00:46:54.483398 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0031612f63728aa09c3b77481dc5f2be8245c926a8e17c06641844ae7d393b8e-rootfs.mount: Deactivated successfully. May 8 00:46:54.489710 env[1311]: time="2025-05-08T00:46:54.489596969Z" level=info msg="shim disconnected" id=0031612f63728aa09c3b77481dc5f2be8245c926a8e17c06641844ae7d393b8e May 8 00:46:54.489710 env[1311]: time="2025-05-08T00:46:54.489659276Z" level=warning msg="cleaning up after shim disconnected" id=0031612f63728aa09c3b77481dc5f2be8245c926a8e17c06641844ae7d393b8e namespace=k8s.io May 8 00:46:54.489710 env[1311]: time="2025-05-08T00:46:54.489679313Z" level=info msg="cleaning up dead shim" May 8 00:46:54.499279 env[1311]: time="2025-05-08T00:46:54.499196226Z" level=warning msg="cleanup warnings time=\"2025-05-08T00:46:54Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2940 runtime=io.containerd.runc.v2\n" May 8 00:46:54.542338 kubelet[2246]: I0508 00:46:54.542292 2246 kubelet_node_status.go:497] "Fast updating node status as it just became ready" May 8 00:46:54.574094 kubelet[2246]: I0508 00:46:54.574001 2246 topology_manager.go:215] "Topology Admit Handler" podUID="435e64db-0f2e-4f05-8fcf-eefdac14188d" podNamespace="kube-system" podName="coredns-7db6d8ff4d-8xjk8" May 8 00:46:54.583721 kubelet[2246]: I0508 00:46:54.583485 2246 topology_manager.go:215] "Topology Admit Handler" podUID="9c26b328-0da1-4b8c-b690-cd29958643ac" podNamespace="kube-system" podName="coredns-7db6d8ff4d-c2qjt" May 8 00:46:54.584063 kubelet[2246]: I0508 00:46:54.583986 2246 topology_manager.go:215] "Topology Admit Handler" podUID="c5c7ed4b-9783-4fe2-b85a-8452a359cb74" podNamespace="calico-apiserver" podName="calico-apiserver-fd97795bb-tzgmm" May 8 00:46:54.597583 kubelet[2246]: I0508 00:46:54.595908 2246 topology_manager.go:215] "Topology Admit Handler" podUID="191fecaf-17dc-4c3f-a824-7b4430fd6b4f" podNamespace="calico-system" podName="calico-kube-controllers-7f6ff4b949-8b2r4" May 8 00:46:54.597583 kubelet[2246]: I0508 00:46:54.596126 2246 topology_manager.go:215] "Topology Admit Handler" podUID="ca6c7950-e2c3-4481-994b-8bd117426f80" podNamespace="calico-apiserver" podName="calico-apiserver-fd97795bb-zmwqb" May 8 00:46:54.613841 kubelet[2246]: I0508 00:46:54.613754 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/191fecaf-17dc-4c3f-a824-7b4430fd6b4f-tigera-ca-bundle\") pod \"calico-kube-controllers-7f6ff4b949-8b2r4\" (UID: \"191fecaf-17dc-4c3f-a824-7b4430fd6b4f\") " pod="calico-system/calico-kube-controllers-7f6ff4b949-8b2r4" May 8 00:46:54.613841 kubelet[2246]: I0508 00:46:54.613823 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/ca6c7950-e2c3-4481-994b-8bd117426f80-calico-apiserver-certs\") pod \"calico-apiserver-fd97795bb-zmwqb\" (UID: \"ca6c7950-e2c3-4481-994b-8bd117426f80\") " pod="calico-apiserver/calico-apiserver-fd97795bb-zmwqb" May 8 00:46:54.613841 kubelet[2246]: I0508 00:46:54.613856 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swtxg\" (UniqueName: \"kubernetes.io/projected/435e64db-0f2e-4f05-8fcf-eefdac14188d-kube-api-access-swtxg\") pod \"coredns-7db6d8ff4d-8xjk8\" (UID: \"435e64db-0f2e-4f05-8fcf-eefdac14188d\") " pod="kube-system/coredns-7db6d8ff4d-8xjk8" May 8 00:46:54.614345 kubelet[2246]: I0508 00:46:54.613913 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c26b328-0da1-4b8c-b690-cd29958643ac-config-volume\") pod \"coredns-7db6d8ff4d-c2qjt\" (UID: \"9c26b328-0da1-4b8c-b690-cd29958643ac\") " pod="kube-system/coredns-7db6d8ff4d-c2qjt" May 8 00:46:54.614345 kubelet[2246]: I0508 00:46:54.614008 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cdqr\" (UniqueName: \"kubernetes.io/projected/191fecaf-17dc-4c3f-a824-7b4430fd6b4f-kube-api-access-4cdqr\") pod \"calico-kube-controllers-7f6ff4b949-8b2r4\" (UID: \"191fecaf-17dc-4c3f-a824-7b4430fd6b4f\") " pod="calico-system/calico-kube-controllers-7f6ff4b949-8b2r4" May 8 00:46:54.614345 kubelet[2246]: I0508 00:46:54.614132 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqmn9\" (UniqueName: \"kubernetes.io/projected/9c26b328-0da1-4b8c-b690-cd29958643ac-kube-api-access-pqmn9\") pod \"coredns-7db6d8ff4d-c2qjt\" (UID: \"9c26b328-0da1-4b8c-b690-cd29958643ac\") " pod="kube-system/coredns-7db6d8ff4d-c2qjt" May 8 00:46:54.614345 kubelet[2246]: I0508 00:46:54.614193 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss7c7\" (UniqueName: \"kubernetes.io/projected/c5c7ed4b-9783-4fe2-b85a-8452a359cb74-kube-api-access-ss7c7\") pod \"calico-apiserver-fd97795bb-tzgmm\" (UID: \"c5c7ed4b-9783-4fe2-b85a-8452a359cb74\") " pod="calico-apiserver/calico-apiserver-fd97795bb-tzgmm" May 8 00:46:54.614345 kubelet[2246]: I0508 00:46:54.614213 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws22h\" (UniqueName: \"kubernetes.io/projected/ca6c7950-e2c3-4481-994b-8bd117426f80-kube-api-access-ws22h\") pod \"calico-apiserver-fd97795bb-zmwqb\" (UID: \"ca6c7950-e2c3-4481-994b-8bd117426f80\") " pod="calico-apiserver/calico-apiserver-fd97795bb-zmwqb" May 8 00:46:54.614555 kubelet[2246]: I0508 00:46:54.614232 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/435e64db-0f2e-4f05-8fcf-eefdac14188d-config-volume\") pod \"coredns-7db6d8ff4d-8xjk8\" (UID: \"435e64db-0f2e-4f05-8fcf-eefdac14188d\") " pod="kube-system/coredns-7db6d8ff4d-8xjk8" May 8 00:46:54.614555 kubelet[2246]: I0508 00:46:54.614251 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/c5c7ed4b-9783-4fe2-b85a-8452a359cb74-calico-apiserver-certs\") pod \"calico-apiserver-fd97795bb-tzgmm\" (UID: \"c5c7ed4b-9783-4fe2-b85a-8452a359cb74\") " pod="calico-apiserver/calico-apiserver-fd97795bb-tzgmm" May 8 00:46:54.884151 kubelet[2246]: E0508 00:46:54.884036 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:54.884895 env[1311]: time="2025-05-08T00:46:54.884806344Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-8xjk8,Uid:435e64db-0f2e-4f05-8fcf-eefdac14188d,Namespace:kube-system,Attempt:0,}" May 8 00:46:54.890240 env[1311]: time="2025-05-08T00:46:54.890177646Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fd97795bb-tzgmm,Uid:c5c7ed4b-9783-4fe2-b85a-8452a359cb74,Namespace:calico-apiserver,Attempt:0,}" May 8 00:46:54.897629 kubelet[2246]: E0508 00:46:54.897570 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:54.898367 env[1311]: time="2025-05-08T00:46:54.898311583Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-c2qjt,Uid:9c26b328-0da1-4b8c-b690-cd29958643ac,Namespace:kube-system,Attempt:0,}" May 8 00:46:54.903386 env[1311]: time="2025-05-08T00:46:54.903304896Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fd97795bb-zmwqb,Uid:ca6c7950-e2c3-4481-994b-8bd117426f80,Namespace:calico-apiserver,Attempt:0,}" May 8 00:46:54.904866 env[1311]: time="2025-05-08T00:46:54.904807706Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7f6ff4b949-8b2r4,Uid:191fecaf-17dc-4c3f-a824-7b4430fd6b4f,Namespace:calico-system,Attempt:0,}" May 8 00:46:55.272699 kubelet[2246]: E0508 00:46:55.272659 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:46:55.274564 env[1311]: time="2025-05-08T00:46:55.274517967Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\"" May 8 00:46:55.735401 env[1311]: time="2025-05-08T00:46:55.735215908Z" level=error msg="Failed to destroy network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.736432 env[1311]: time="2025-05-08T00:46:55.736389049Z" level=error msg="encountered an error cleaning up failed sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.736613 env[1311]: time="2025-05-08T00:46:55.736553659Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fd97795bb-zmwqb,Uid:ca6c7950-e2c3-4481-994b-8bd117426f80,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.738409 kubelet[2246]: E0508 00:46:55.737143 2246 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.738409 kubelet[2246]: E0508 00:46:55.737262 2246 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-fd97795bb-zmwqb" May 8 00:46:55.738409 kubelet[2246]: E0508 00:46:55.737300 2246 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-fd97795bb-zmwqb" May 8 00:46:55.738742 kubelet[2246]: E0508 00:46:55.737363 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-fd97795bb-zmwqb_calico-apiserver(ca6c7950-e2c3-4481-994b-8bd117426f80)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-fd97795bb-zmwqb_calico-apiserver(ca6c7950-e2c3-4481-994b-8bd117426f80)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-fd97795bb-zmwqb" podUID="ca6c7950-e2c3-4481-994b-8bd117426f80" May 8 00:46:55.760535 env[1311]: time="2025-05-08T00:46:55.760461823Z" level=error msg="Failed to destroy network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.761065 env[1311]: time="2025-05-08T00:46:55.760866602Z" level=error msg="Failed to destroy network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.761524 env[1311]: time="2025-05-08T00:46:55.761494010Z" level=error msg="encountered an error cleaning up failed sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.761686 env[1311]: time="2025-05-08T00:46:55.761632891Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-c2qjt,Uid:9c26b328-0da1-4b8c-b690-cd29958643ac,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.761851 env[1311]: time="2025-05-08T00:46:55.761654441Z" level=error msg="encountered an error cleaning up failed sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.761971 env[1311]: time="2025-05-08T00:46:55.761937373Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7f6ff4b949-8b2r4,Uid:191fecaf-17dc-4c3f-a824-7b4430fd6b4f,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.762172 kubelet[2246]: E0508 00:46:55.762095 2246 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.762248 kubelet[2246]: E0508 00:46:55.762211 2246 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-c2qjt" May 8 00:46:55.762281 kubelet[2246]: E0508 00:46:55.762242 2246 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-c2qjt" May 8 00:46:55.762362 kubelet[2246]: E0508 00:46:55.762316 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-c2qjt_kube-system(9c26b328-0da1-4b8c-b690-cd29958643ac)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-c2qjt_kube-system(9c26b328-0da1-4b8c-b690-cd29958643ac)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-c2qjt" podUID="9c26b328-0da1-4b8c-b690-cd29958643ac" May 8 00:46:55.762580 kubelet[2246]: E0508 00:46:55.762548 2246 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.762652 kubelet[2246]: E0508 00:46:55.762586 2246 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7f6ff4b949-8b2r4" May 8 00:46:55.762652 kubelet[2246]: E0508 00:46:55.762604 2246 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7f6ff4b949-8b2r4" May 8 00:46:55.762708 kubelet[2246]: E0508 00:46:55.762666 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7f6ff4b949-8b2r4_calico-system(191fecaf-17dc-4c3f-a824-7b4430fd6b4f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7f6ff4b949-8b2r4_calico-system(191fecaf-17dc-4c3f-a824-7b4430fd6b4f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7f6ff4b949-8b2r4" podUID="191fecaf-17dc-4c3f-a824-7b4430fd6b4f" May 8 00:46:55.763663 env[1311]: time="2025-05-08T00:46:55.763626493Z" level=error msg="Failed to destroy network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.764093 env[1311]: time="2025-05-08T00:46:55.764040940Z" level=error msg="encountered an error cleaning up failed sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.764171 env[1311]: time="2025-05-08T00:46:55.764143222Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fd97795bb-tzgmm,Uid:c5c7ed4b-9783-4fe2-b85a-8452a359cb74,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.764397 kubelet[2246]: E0508 00:46:55.764367 2246 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.764469 kubelet[2246]: E0508 00:46:55.764410 2246 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-fd97795bb-tzgmm" May 8 00:46:55.764469 kubelet[2246]: E0508 00:46:55.764433 2246 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-fd97795bb-tzgmm" May 8 00:46:55.764530 kubelet[2246]: E0508 00:46:55.764475 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-fd97795bb-tzgmm_calico-apiserver(c5c7ed4b-9783-4fe2-b85a-8452a359cb74)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-fd97795bb-tzgmm_calico-apiserver(c5c7ed4b-9783-4fe2-b85a-8452a359cb74)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-fd97795bb-tzgmm" podUID="c5c7ed4b-9783-4fe2-b85a-8452a359cb74" May 8 00:46:55.768265 env[1311]: time="2025-05-08T00:46:55.768216828Z" level=error msg="Failed to destroy network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.768566 env[1311]: time="2025-05-08T00:46:55.768523174Z" level=error msg="encountered an error cleaning up failed sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.768662 env[1311]: time="2025-05-08T00:46:55.768563800Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-8xjk8,Uid:435e64db-0f2e-4f05-8fcf-eefdac14188d,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.768830 kubelet[2246]: E0508 00:46:55.768777 2246 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:55.768917 kubelet[2246]: E0508 00:46:55.768849 2246 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-8xjk8" May 8 00:46:55.768917 kubelet[2246]: E0508 00:46:55.768870 2246 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-8xjk8" May 8 00:46:55.769008 kubelet[2246]: E0508 00:46:55.768921 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-8xjk8_kube-system(435e64db-0f2e-4f05-8fcf-eefdac14188d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-8xjk8_kube-system(435e64db-0f2e-4f05-8fcf-eefdac14188d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-8xjk8" podUID="435e64db-0f2e-4f05-8fcf-eefdac14188d" May 8 00:46:56.168828 env[1311]: time="2025-05-08T00:46:56.168766943Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-2s6tt,Uid:863a2a82-a51b-4afc-a7be-990ee722b050,Namespace:calico-system,Attempt:0,}" May 8 00:46:56.275035 kubelet[2246]: I0508 00:46:56.274997 2246 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:46:56.276127 kubelet[2246]: I0508 00:46:56.276083 2246 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:46:56.276228 env[1311]: time="2025-05-08T00:46:56.276091596Z" level=info msg="StopPodSandbox for \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\"" May 8 00:46:56.276552 env[1311]: time="2025-05-08T00:46:56.276522494Z" level=info msg="StopPodSandbox for \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\"" May 8 00:46:56.277989 kubelet[2246]: I0508 00:46:56.277950 2246 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:46:56.278552 env[1311]: time="2025-05-08T00:46:56.278519843Z" level=info msg="StopPodSandbox for \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\"" May 8 00:46:56.280198 kubelet[2246]: I0508 00:46:56.280126 2246 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:46:56.281775 env[1311]: time="2025-05-08T00:46:56.281731130Z" level=info msg="StopPodSandbox for \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\"" May 8 00:46:56.283164 kubelet[2246]: I0508 00:46:56.283039 2246 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:46:56.284048 env[1311]: time="2025-05-08T00:46:56.283986914Z" level=info msg="StopPodSandbox for \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\"" May 8 00:46:56.320357 env[1311]: time="2025-05-08T00:46:56.320275332Z" level=error msg="StopPodSandbox for \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\" failed" error="failed to destroy network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:56.321197 kubelet[2246]: E0508 00:46:56.320901 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:46:56.321197 kubelet[2246]: E0508 00:46:56.320982 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2"} May 8 00:46:56.321197 kubelet[2246]: E0508 00:46:56.321074 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"435e64db-0f2e-4f05-8fcf-eefdac14188d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:46:56.321197 kubelet[2246]: E0508 00:46:56.321126 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"435e64db-0f2e-4f05-8fcf-eefdac14188d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-8xjk8" podUID="435e64db-0f2e-4f05-8fcf-eefdac14188d" May 8 00:46:56.323688 env[1311]: time="2025-05-08T00:46:56.323603539Z" level=error msg="StopPodSandbox for \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\" failed" error="failed to destroy network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:56.323924 kubelet[2246]: E0508 00:46:56.323880 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:46:56.324001 kubelet[2246]: E0508 00:46:56.323938 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b"} May 8 00:46:56.324001 kubelet[2246]: E0508 00:46:56.323985 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"ca6c7950-e2c3-4481-994b-8bd117426f80\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:46:56.324138 kubelet[2246]: E0508 00:46:56.324014 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"ca6c7950-e2c3-4481-994b-8bd117426f80\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-fd97795bb-zmwqb" podUID="ca6c7950-e2c3-4481-994b-8bd117426f80" May 8 00:46:56.335159 env[1311]: time="2025-05-08T00:46:56.335057596Z" level=error msg="StopPodSandbox for \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\" failed" error="failed to destroy network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:56.335456 kubelet[2246]: E0508 00:46:56.335404 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:46:56.335542 kubelet[2246]: E0508 00:46:56.335473 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd"} May 8 00:46:56.335542 kubelet[2246]: E0508 00:46:56.335516 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"c5c7ed4b-9783-4fe2-b85a-8452a359cb74\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:46:56.335681 kubelet[2246]: E0508 00:46:56.335550 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"c5c7ed4b-9783-4fe2-b85a-8452a359cb74\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-fd97795bb-tzgmm" podUID="c5c7ed4b-9783-4fe2-b85a-8452a359cb74" May 8 00:46:56.336486 env[1311]: time="2025-05-08T00:46:56.336427598Z" level=error msg="StopPodSandbox for \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\" failed" error="failed to destroy network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:56.336661 kubelet[2246]: E0508 00:46:56.336613 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:46:56.336722 kubelet[2246]: E0508 00:46:56.336661 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4"} May 8 00:46:56.336722 kubelet[2246]: E0508 00:46:56.336691 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"191fecaf-17dc-4c3f-a824-7b4430fd6b4f\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:46:56.336825 kubelet[2246]: E0508 00:46:56.336721 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"191fecaf-17dc-4c3f-a824-7b4430fd6b4f\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7f6ff4b949-8b2r4" podUID="191fecaf-17dc-4c3f-a824-7b4430fd6b4f" May 8 00:46:56.341938 env[1311]: time="2025-05-08T00:46:56.341798027Z" level=error msg="StopPodSandbox for \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\" failed" error="failed to destroy network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:56.342239 kubelet[2246]: E0508 00:46:56.342182 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:46:56.342314 kubelet[2246]: E0508 00:46:56.342251 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380"} May 8 00:46:56.342314 kubelet[2246]: E0508 00:46:56.342301 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"9c26b328-0da1-4b8c-b690-cd29958643ac\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:46:56.342448 kubelet[2246]: E0508 00:46:56.342331 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"9c26b328-0da1-4b8c-b690-cd29958643ac\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-c2qjt" podUID="9c26b328-0da1-4b8c-b690-cd29958643ac" May 8 00:46:56.484520 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2-shm.mount: Deactivated successfully. May 8 00:46:56.484761 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380-shm.mount: Deactivated successfully. May 8 00:46:56.544095 env[1311]: time="2025-05-08T00:46:56.544013046Z" level=error msg="Failed to destroy network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:56.544553 env[1311]: time="2025-05-08T00:46:56.544502606Z" level=error msg="encountered an error cleaning up failed sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:56.544612 env[1311]: time="2025-05-08T00:46:56.544567628Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-2s6tt,Uid:863a2a82-a51b-4afc-a7be-990ee722b050,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:56.544942 kubelet[2246]: E0508 00:46:56.544887 2246 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:56.545034 kubelet[2246]: E0508 00:46:56.544979 2246 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-2s6tt" May 8 00:46:56.545034 kubelet[2246]: E0508 00:46:56.545009 2246 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-2s6tt" May 8 00:46:56.545270 kubelet[2246]: E0508 00:46:56.545074 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-2s6tt_calico-system(863a2a82-a51b-4afc-a7be-990ee722b050)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-2s6tt_calico-system(863a2a82-a51b-4afc-a7be-990ee722b050)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-2s6tt" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" May 8 00:46:56.546799 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce-shm.mount: Deactivated successfully. May 8 00:46:57.286942 kubelet[2246]: I0508 00:46:57.286891 2246 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:46:57.287548 env[1311]: time="2025-05-08T00:46:57.287516929Z" level=info msg="StopPodSandbox for \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\"" May 8 00:46:57.324942 env[1311]: time="2025-05-08T00:46:57.324857980Z" level=error msg="StopPodSandbox for \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\" failed" error="failed to destroy network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:46:57.325256 kubelet[2246]: E0508 00:46:57.325175 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:46:57.325339 kubelet[2246]: E0508 00:46:57.325254 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce"} May 8 00:46:57.325339 kubelet[2246]: E0508 00:46:57.325292 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"863a2a82-a51b-4afc-a7be-990ee722b050\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:46:57.325339 kubelet[2246]: E0508 00:46:57.325319 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"863a2a82-a51b-4afc-a7be-990ee722b050\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-2s6tt" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" May 8 00:46:59.132000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.82:22-10.0.0.1:57056 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:59.133478 systemd[1]: Started sshd@7-10.0.0.82:22-10.0.0.1:57056.service. May 8 00:46:59.149158 kernel: audit: type=1130 audit(1746665219.132:289): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.82:22-10.0.0.1:57056 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:59.270000 audit[3333]: USER_ACCT pid=3333 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:59.271976 sshd[3333]: Accepted publickey for core from 10.0.0.1 port 57056 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:46:59.272000 audit[3333]: CRED_ACQ pid=3333 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:59.280907 kernel: audit: type=1101 audit(1746665219.270:290): pid=3333 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:59.281038 kernel: audit: type=1103 audit(1746665219.272:291): pid=3333 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:59.281071 kernel: audit: type=1006 audit(1746665219.273:292): pid=3333 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=8 res=1 May 8 00:46:59.281460 sshd[3333]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:46:59.283597 kernel: audit: type=1300 audit(1746665219.273:292): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffc4febbf80 a2=3 a3=0 items=0 ppid=1 pid=3333 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=8 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:59.273000 audit[3333]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffc4febbf80 a2=3 a3=0 items=0 ppid=1 pid=3333 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=8 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:46:59.287915 kernel: audit: type=1327 audit(1746665219.273:292): proctitle=737368643A20636F7265205B707269765D May 8 00:46:59.273000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:46:59.306205 systemd-logind[1291]: New session 8 of user core. May 8 00:46:59.307323 systemd[1]: Started session-8.scope. May 8 00:46:59.310000 audit[3333]: USER_START pid=3333 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:59.312000 audit[3336]: CRED_ACQ pid=3336 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:59.327644 kernel: audit: type=1105 audit(1746665219.310:293): pid=3333 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:59.327823 kernel: audit: type=1103 audit(1746665219.312:294): pid=3336 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:59.759419 sshd[3333]: pam_unix(sshd:session): session closed for user core May 8 00:46:59.760000 audit[3333]: USER_END pid=3333 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:59.767307 kernel: audit: type=1106 audit(1746665219.760:295): pid=3333 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:59.766000 audit[3333]: CRED_DISP pid=3333 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:59.769272 systemd[1]: sshd@7-10.0.0.82:22-10.0.0.1:57056.service: Deactivated successfully. May 8 00:46:59.770196 systemd[1]: session-8.scope: Deactivated successfully. May 8 00:46:59.772623 kernel: audit: type=1104 audit(1746665219.766:296): pid=3333 uid=0 auid=500 ses=8 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:46:59.768000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@7-10.0.0.82:22-10.0.0.1:57056 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:46:59.772229 systemd-logind[1291]: Session 8 logged out. Waiting for processes to exit. May 8 00:46:59.773391 systemd-logind[1291]: Removed session 8. May 8 00:47:03.371781 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2683631742.mount: Deactivated successfully. May 8 00:47:04.764059 systemd[1]: Started sshd@8-10.0.0.82:22-10.0.0.1:57058.service. May 8 00:47:04.763000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.82:22-10.0.0.1:57058 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:04.765360 kernel: kauditd_printk_skb: 1 callbacks suppressed May 8 00:47:04.765445 kernel: audit: type=1130 audit(1746665224.763:298): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.82:22-10.0.0.1:57058 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:05.130000 audit[3351]: USER_ACCT pid=3351 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:05.131552 sshd[3351]: Accepted publickey for core from 10.0.0.1 port 57058 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:05.134000 audit[3351]: CRED_ACQ pid=3351 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:05.136597 sshd[3351]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:05.140495 kernel: audit: type=1101 audit(1746665225.130:299): pid=3351 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:05.140623 kernel: audit: type=1103 audit(1746665225.134:300): pid=3351 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:05.140706 kernel: audit: type=1006 audit(1746665225.134:301): pid=3351 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=9 res=1 May 8 00:47:05.148074 kernel: audit: type=1300 audit(1746665225.134:301): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffd9f7390d0 a2=3 a3=0 items=0 ppid=1 pid=3351 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=9 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:05.148261 kernel: audit: type=1327 audit(1746665225.134:301): proctitle=737368643A20636F7265205B707269765D May 8 00:47:05.134000 audit[3351]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffd9f7390d0 a2=3 a3=0 items=0 ppid=1 pid=3351 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=9 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:05.134000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:05.142898 systemd[1]: Started session-9.scope. May 8 00:47:05.143490 systemd-logind[1291]: New session 9 of user core. May 8 00:47:05.148000 audit[3351]: USER_START pid=3351 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:05.149000 audit[3354]: CRED_ACQ pid=3354 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:05.198456 kernel: audit: type=1105 audit(1746665225.148:302): pid=3351 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:05.198617 kernel: audit: type=1103 audit(1746665225.149:303): pid=3354 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:05.456431 sshd[3351]: pam_unix(sshd:session): session closed for user core May 8 00:47:05.458000 audit[3351]: USER_END pid=3351 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:05.461438 systemd[1]: sshd@8-10.0.0.82:22-10.0.0.1:57058.service: Deactivated successfully. May 8 00:47:05.463462 systemd-logind[1291]: Session 9 logged out. Waiting for processes to exit. May 8 00:47:05.463484 systemd[1]: session-9.scope: Deactivated successfully. May 8 00:47:05.464724 systemd-logind[1291]: Removed session 9. May 8 00:47:05.458000 audit[3351]: CRED_DISP pid=3351 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:05.469866 kernel: audit: type=1106 audit(1746665225.458:304): pid=3351 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:05.470325 kernel: audit: type=1104 audit(1746665225.458:305): pid=3351 uid=0 auid=500 ses=9 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:05.458000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@8-10.0.0.82:22-10.0.0.1:57058 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:07.166461 env[1311]: time="2025-05-08T00:47:07.166402703Z" level=info msg="StopPodSandbox for \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\"" May 8 00:47:07.404349 env[1311]: time="2025-05-08T00:47:07.404268714Z" level=error msg="StopPodSandbox for \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\" failed" error="failed to destroy network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:07.404690 kubelet[2246]: E0508 00:47:07.404571 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:47:07.404690 kubelet[2246]: E0508 00:47:07.404687 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380"} May 8 00:47:07.405238 kubelet[2246]: E0508 00:47:07.404728 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"9c26b328-0da1-4b8c-b690-cd29958643ac\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:07.405238 kubelet[2246]: E0508 00:47:07.404766 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"9c26b328-0da1-4b8c-b690-cd29958643ac\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-c2qjt" podUID="9c26b328-0da1-4b8c-b690-cd29958643ac" May 8 00:47:07.447743 env[1311]: time="2025-05-08T00:47:07.447599647Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/node:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:07.454681 env[1311]: time="2025-05-08T00:47:07.454612090Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:07.459733 env[1311]: time="2025-05-08T00:47:07.458817173Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/node:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:07.462415 env[1311]: time="2025-05-08T00:47:07.462340141Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:07.463174 env[1311]: time="2025-05-08T00:47:07.463096669Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\" returns image reference \"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\"" May 8 00:47:07.478518 env[1311]: time="2025-05-08T00:47:07.478450073Z" level=info msg="CreateContainer within sandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" May 8 00:47:07.858504 env[1311]: time="2025-05-08T00:47:07.858406657Z" level=info msg="CreateContainer within sandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"c9d2a5c264f81e8d3436644c6c4ad2f413a5d9d47f1cb3a2a8166d92570e8226\"" May 8 00:47:07.859288 env[1311]: time="2025-05-08T00:47:07.859208653Z" level=info msg="StartContainer for \"c9d2a5c264f81e8d3436644c6c4ad2f413a5d9d47f1cb3a2a8166d92570e8226\"" May 8 00:47:07.991146 env[1311]: time="2025-05-08T00:47:07.991050877Z" level=info msg="StartContainer for \"c9d2a5c264f81e8d3436644c6c4ad2f413a5d9d47f1cb3a2a8166d92570e8226\" returns successfully" May 8 00:47:08.032937 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. May 8 00:47:08.033167 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. May 8 00:47:08.083641 env[1311]: time="2025-05-08T00:47:08.083571511Z" level=info msg="shim disconnected" id=c9d2a5c264f81e8d3436644c6c4ad2f413a5d9d47f1cb3a2a8166d92570e8226 May 8 00:47:08.083641 env[1311]: time="2025-05-08T00:47:08.083635414Z" level=warning msg="cleaning up after shim disconnected" id=c9d2a5c264f81e8d3436644c6c4ad2f413a5d9d47f1cb3a2a8166d92570e8226 namespace=k8s.io May 8 00:47:08.083641 env[1311]: time="2025-05-08T00:47:08.083647787Z" level=info msg="cleaning up dead shim" May 8 00:47:08.091995 env[1311]: time="2025-05-08T00:47:08.091929294Z" level=warning msg="cleanup warnings time=\"2025-05-08T00:47:08Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3450 runtime=io.containerd.runc.v2\n" May 8 00:47:08.167252 env[1311]: time="2025-05-08T00:47:08.165957626Z" level=info msg="StopPodSandbox for \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\"" May 8 00:47:08.192157 env[1311]: time="2025-05-08T00:47:08.192060567Z" level=error msg="StopPodSandbox for \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\" failed" error="failed to destroy network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:08.192473 kubelet[2246]: E0508 00:47:08.192411 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:47:08.192549 kubelet[2246]: E0508 00:47:08.192486 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2"} May 8 00:47:08.192549 kubelet[2246]: E0508 00:47:08.192532 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"435e64db-0f2e-4f05-8fcf-eefdac14188d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:08.192650 kubelet[2246]: E0508 00:47:08.192565 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"435e64db-0f2e-4f05-8fcf-eefdac14188d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-8xjk8" podUID="435e64db-0f2e-4f05-8fcf-eefdac14188d" May 8 00:47:08.313636 kubelet[2246]: I0508 00:47:08.313587 2246 scope.go:117] "RemoveContainer" containerID="c9d2a5c264f81e8d3436644c6c4ad2f413a5d9d47f1cb3a2a8166d92570e8226" May 8 00:47:08.313901 kubelet[2246]: E0508 00:47:08.313722 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:08.317122 env[1311]: time="2025-05-08T00:47:08.317061116Z" level=info msg="CreateContainer within sandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" for container &ContainerMetadata{Name:calico-node,Attempt:1,}" May 8 00:47:08.342848 env[1311]: time="2025-05-08T00:47:08.342269405Z" level=info msg="CreateContainer within sandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" for &ContainerMetadata{Name:calico-node,Attempt:1,} returns container id \"af0424ae4b2bc3f499e117c178382d0b6c3efb4eae77dd7bb7145114f6fccfae\"" May 8 00:47:08.344065 env[1311]: time="2025-05-08T00:47:08.344015898Z" level=info msg="StartContainer for \"af0424ae4b2bc3f499e117c178382d0b6c3efb4eae77dd7bb7145114f6fccfae\"" May 8 00:47:08.403046 env[1311]: time="2025-05-08T00:47:08.402963341Z" level=info msg="StartContainer for \"af0424ae4b2bc3f499e117c178382d0b6c3efb4eae77dd7bb7145114f6fccfae\" returns successfully" May 8 00:47:08.473772 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c9d2a5c264f81e8d3436644c6c4ad2f413a5d9d47f1cb3a2a8166d92570e8226-rootfs.mount: Deactivated successfully. May 8 00:47:08.487397 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-af0424ae4b2bc3f499e117c178382d0b6c3efb4eae77dd7bb7145114f6fccfae-rootfs.mount: Deactivated successfully. May 8 00:47:08.494097 env[1311]: time="2025-05-08T00:47:08.494032034Z" level=info msg="shim disconnected" id=af0424ae4b2bc3f499e117c178382d0b6c3efb4eae77dd7bb7145114f6fccfae May 8 00:47:08.494097 env[1311]: time="2025-05-08T00:47:08.494095828Z" level=warning msg="cleaning up after shim disconnected" id=af0424ae4b2bc3f499e117c178382d0b6c3efb4eae77dd7bb7145114f6fccfae namespace=k8s.io May 8 00:47:08.494097 env[1311]: time="2025-05-08T00:47:08.494128069Z" level=info msg="cleaning up dead shim" May 8 00:47:08.501420 env[1311]: time="2025-05-08T00:47:08.501347831Z" level=warning msg="cleanup warnings time=\"2025-05-08T00:47:08Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3538 runtime=io.containerd.runc.v2\n" May 8 00:47:09.317719 kubelet[2246]: I0508 00:47:09.317675 2246 scope.go:117] "RemoveContainer" containerID="c9d2a5c264f81e8d3436644c6c4ad2f413a5d9d47f1cb3a2a8166d92570e8226" May 8 00:47:09.318254 kubelet[2246]: I0508 00:47:09.318120 2246 scope.go:117] "RemoveContainer" containerID="af0424ae4b2bc3f499e117c178382d0b6c3efb4eae77dd7bb7145114f6fccfae" May 8 00:47:09.318254 kubelet[2246]: E0508 00:47:09.318215 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:09.318710 kubelet[2246]: E0508 00:47:09.318680 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-node\" with CrashLoopBackOff: \"back-off 10s restarting failed container=calico-node pod=calico-node-k68b7_calico-system(ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5)\"" pod="calico-system/calico-node-k68b7" podUID="ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" May 8 00:47:09.319262 env[1311]: time="2025-05-08T00:47:09.319228419Z" level=info msg="RemoveContainer for \"c9d2a5c264f81e8d3436644c6c4ad2f413a5d9d47f1cb3a2a8166d92570e8226\"" May 8 00:47:09.326550 env[1311]: time="2025-05-08T00:47:09.326484943Z" level=info msg="RemoveContainer for \"c9d2a5c264f81e8d3436644c6c4ad2f413a5d9d47f1cb3a2a8166d92570e8226\" returns successfully" May 8 00:47:10.166580 env[1311]: time="2025-05-08T00:47:10.166508863Z" level=info msg="StopPodSandbox for \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\"" May 8 00:47:10.191360 env[1311]: time="2025-05-08T00:47:10.191276208Z" level=error msg="StopPodSandbox for \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\" failed" error="failed to destroy network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:10.191655 kubelet[2246]: E0508 00:47:10.191583 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:47:10.191748 kubelet[2246]: E0508 00:47:10.191658 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b"} May 8 00:47:10.191748 kubelet[2246]: E0508 00:47:10.191697 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"ca6c7950-e2c3-4481-994b-8bd117426f80\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:10.191748 kubelet[2246]: E0508 00:47:10.191721 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"ca6c7950-e2c3-4481-994b-8bd117426f80\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-fd97795bb-zmwqb" podUID="ca6c7950-e2c3-4481-994b-8bd117426f80" May 8 00:47:10.460000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.82:22-10.0.0.1:58584 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:10.460558 systemd[1]: Started sshd@9-10.0.0.82:22-10.0.0.1:58584.service. May 8 00:47:10.462020 kernel: kauditd_printk_skb: 1 callbacks suppressed May 8 00:47:10.462241 kernel: audit: type=1130 audit(1746665230.460:307): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.82:22-10.0.0.1:58584 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:10.510000 audit[3574]: USER_ACCT pid=3574 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:10.511148 sshd[3574]: Accepted publickey for core from 10.0.0.1 port 58584 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:10.514794 sshd[3574]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:10.514000 audit[3574]: CRED_ACQ pid=3574 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:10.519090 kernel: audit: type=1101 audit(1746665230.510:308): pid=3574 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:10.519247 kernel: audit: type=1103 audit(1746665230.514:309): pid=3574 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:10.519278 kernel: audit: type=1006 audit(1746665230.514:310): pid=3574 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=10 res=1 May 8 00:47:10.519576 systemd-logind[1291]: New session 10 of user core. May 8 00:47:10.520860 systemd[1]: Started session-10.scope. May 8 00:47:10.514000 audit[3574]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7fff82885690 a2=3 a3=0 items=0 ppid=1 pid=3574 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:10.525323 kernel: audit: type=1300 audit(1746665230.514:310): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7fff82885690 a2=3 a3=0 items=0 ppid=1 pid=3574 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=10 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:10.525410 kernel: audit: type=1327 audit(1746665230.514:310): proctitle=737368643A20636F7265205B707269765D May 8 00:47:10.514000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:10.527000 audit[3574]: USER_START pid=3574 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:10.529000 audit[3577]: CRED_ACQ pid=3577 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:10.535351 kernel: audit: type=1105 audit(1746665230.527:311): pid=3574 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:10.535413 kernel: audit: type=1103 audit(1746665230.529:312): pid=3577 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:10.682177 sshd[3574]: pam_unix(sshd:session): session closed for user core May 8 00:47:10.683000 audit[3574]: USER_END pid=3574 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:10.685247 systemd[1]: sshd@9-10.0.0.82:22-10.0.0.1:58584.service: Deactivated successfully. May 8 00:47:10.686387 systemd-logind[1291]: Session 10 logged out. Waiting for processes to exit. May 8 00:47:10.686405 systemd[1]: session-10.scope: Deactivated successfully. May 8 00:47:10.687226 systemd-logind[1291]: Removed session 10. May 8 00:47:10.683000 audit[3574]: CRED_DISP pid=3574 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:10.691154 kernel: audit: type=1106 audit(1746665230.683:313): pid=3574 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:10.691247 kernel: audit: type=1104 audit(1746665230.683:314): pid=3574 uid=0 auid=500 ses=10 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:10.683000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@9-10.0.0.82:22-10.0.0.1:58584 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:11.166529 env[1311]: time="2025-05-08T00:47:11.166457212Z" level=info msg="StopPodSandbox for \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\"" May 8 00:47:11.166529 env[1311]: time="2025-05-08T00:47:11.166482021Z" level=info msg="StopPodSandbox for \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\"" May 8 00:47:11.167133 env[1311]: time="2025-05-08T00:47:11.166457282Z" level=info msg="StopPodSandbox for \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\"" May 8 00:47:11.199650 env[1311]: time="2025-05-08T00:47:11.199573619Z" level=error msg="StopPodSandbox for \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\" failed" error="failed to destroy network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:11.200422 kubelet[2246]: E0508 00:47:11.200198 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:47:11.200422 kubelet[2246]: E0508 00:47:11.200273 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4"} May 8 00:47:11.200422 kubelet[2246]: E0508 00:47:11.200318 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"191fecaf-17dc-4c3f-a824-7b4430fd6b4f\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:11.200422 kubelet[2246]: E0508 00:47:11.200349 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"191fecaf-17dc-4c3f-a824-7b4430fd6b4f\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7f6ff4b949-8b2r4" podUID="191fecaf-17dc-4c3f-a824-7b4430fd6b4f" May 8 00:47:11.202386 env[1311]: time="2025-05-08T00:47:11.202306962Z" level=error msg="StopPodSandbox for \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\" failed" error="failed to destroy network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:11.202576 kubelet[2246]: E0508 00:47:11.202547 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:47:11.202638 kubelet[2246]: E0508 00:47:11.202586 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce"} May 8 00:47:11.202638 kubelet[2246]: E0508 00:47:11.202606 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"863a2a82-a51b-4afc-a7be-990ee722b050\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:11.202638 kubelet[2246]: E0508 00:47:11.202623 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"863a2a82-a51b-4afc-a7be-990ee722b050\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-2s6tt" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" May 8 00:47:11.210141 env[1311]: time="2025-05-08T00:47:11.210053107Z" level=error msg="StopPodSandbox for \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\" failed" error="failed to destroy network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:11.210335 kubelet[2246]: E0508 00:47:11.210287 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:47:11.210411 kubelet[2246]: E0508 00:47:11.210341 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd"} May 8 00:47:11.210411 kubelet[2246]: E0508 00:47:11.210364 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"c5c7ed4b-9783-4fe2-b85a-8452a359cb74\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:11.210411 kubelet[2246]: E0508 00:47:11.210380 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"c5c7ed4b-9783-4fe2-b85a-8452a359cb74\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-fd97795bb-tzgmm" podUID="c5c7ed4b-9783-4fe2-b85a-8452a359cb74" May 8 00:47:15.685676 systemd[1]: Started sshd@10-10.0.0.82:22-10.0.0.1:50822.service. May 8 00:47:15.684000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.82:22-10.0.0.1:50822 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:15.688221 kernel: kauditd_printk_skb: 1 callbacks suppressed May 8 00:47:15.688306 kernel: audit: type=1130 audit(1746665235.684:316): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.82:22-10.0.0.1:50822 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:15.741000 audit[3664]: USER_ACCT pid=3664 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:15.743319 sshd[3664]: Accepted publickey for core from 10.0.0.1 port 50822 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:15.745869 sshd[3664]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:15.744000 audit[3664]: CRED_ACQ pid=3664 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:15.751024 systemd-logind[1291]: New session 11 of user core. May 8 00:47:15.751765 kernel: audit: type=1101 audit(1746665235.741:317): pid=3664 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:15.751820 kernel: audit: type=1103 audit(1746665235.744:318): pid=3664 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:15.751841 kernel: audit: type=1006 audit(1746665235.744:319): pid=3664 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=11 res=1 May 8 00:47:15.752348 systemd[1]: Started session-11.scope. May 8 00:47:15.753789 kernel: audit: type=1300 audit(1746665235.744:319): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffedfd22660 a2=3 a3=0 items=0 ppid=1 pid=3664 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=11 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:15.744000 audit[3664]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffedfd22660 a2=3 a3=0 items=0 ppid=1 pid=3664 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=11 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:15.757843 kernel: audit: type=1327 audit(1746665235.744:319): proctitle=737368643A20636F7265205B707269765D May 8 00:47:15.744000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:15.759042 kernel: audit: type=1105 audit(1746665235.757:320): pid=3664 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:15.757000 audit[3664]: USER_START pid=3664 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:15.776574 kernel: audit: type=1103 audit(1746665235.758:321): pid=3667 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:15.758000 audit[3667]: CRED_ACQ pid=3667 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:15.897815 sshd[3664]: pam_unix(sshd:session): session closed for user core May 8 00:47:15.897000 audit[3664]: USER_END pid=3664 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:15.901072 systemd[1]: sshd@10-10.0.0.82:22-10.0.0.1:50822.service: Deactivated successfully. May 8 00:47:15.902190 systemd[1]: session-11.scope: Deactivated successfully. May 8 00:47:15.902717 systemd-logind[1291]: Session 11 logged out. Waiting for processes to exit. May 8 00:47:15.903534 systemd-logind[1291]: Removed session 11. May 8 00:47:15.897000 audit[3664]: CRED_DISP pid=3664 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:15.910426 kernel: audit: type=1106 audit(1746665235.897:322): pid=3664 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:15.910480 kernel: audit: type=1104 audit(1746665235.897:323): pid=3664 uid=0 auid=500 ses=11 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:15.897000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@10-10.0.0.82:22-10.0.0.1:50822 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:20.166933 kubelet[2246]: I0508 00:47:20.166820 2246 scope.go:117] "RemoveContainer" containerID="af0424ae4b2bc3f499e117c178382d0b6c3efb4eae77dd7bb7145114f6fccfae" May 8 00:47:20.168016 kubelet[2246]: E0508 00:47:20.167044 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:20.168159 env[1311]: time="2025-05-08T00:47:20.167586455Z" level=info msg="StopPodSandbox for \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\"" May 8 00:47:20.170896 env[1311]: time="2025-05-08T00:47:20.170849485Z" level=info msg="CreateContainer within sandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" for container &ContainerMetadata{Name:calico-node,Attempt:2,}" May 8 00:47:20.260741 env[1311]: time="2025-05-08T00:47:20.260641319Z" level=error msg="StopPodSandbox for \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\" failed" error="failed to destroy network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:20.261068 kubelet[2246]: E0508 00:47:20.260997 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:47:20.261169 kubelet[2246]: E0508 00:47:20.261096 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2"} May 8 00:47:20.261204 kubelet[2246]: E0508 00:47:20.261180 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"435e64db-0f2e-4f05-8fcf-eefdac14188d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:20.261298 kubelet[2246]: E0508 00:47:20.261215 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"435e64db-0f2e-4f05-8fcf-eefdac14188d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-8xjk8" podUID="435e64db-0f2e-4f05-8fcf-eefdac14188d" May 8 00:47:20.405872 env[1311]: time="2025-05-08T00:47:20.405804329Z" level=info msg="CreateContainer within sandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" for &ContainerMetadata{Name:calico-node,Attempt:2,} returns container id \"68edd6e95fddc102379a29cffc4ea28aed018582e663a0d5f501d88325e45eaf\"" May 8 00:47:20.406403 env[1311]: time="2025-05-08T00:47:20.406367216Z" level=info msg="StartContainer for \"68edd6e95fddc102379a29cffc4ea28aed018582e663a0d5f501d88325e45eaf\"" May 8 00:47:20.531166 env[1311]: time="2025-05-08T00:47:20.531058095Z" level=info msg="StartContainer for \"68edd6e95fddc102379a29cffc4ea28aed018582e663a0d5f501d88325e45eaf\" returns successfully" May 8 00:47:20.642780 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-68edd6e95fddc102379a29cffc4ea28aed018582e663a0d5f501d88325e45eaf-rootfs.mount: Deactivated successfully. May 8 00:47:20.651491 env[1311]: time="2025-05-08T00:47:20.651422752Z" level=info msg="shim disconnected" id=68edd6e95fddc102379a29cffc4ea28aed018582e663a0d5f501d88325e45eaf May 8 00:47:20.651491 env[1311]: time="2025-05-08T00:47:20.651485713Z" level=warning msg="cleaning up after shim disconnected" id=68edd6e95fddc102379a29cffc4ea28aed018582e663a0d5f501d88325e45eaf namespace=k8s.io May 8 00:47:20.651491 env[1311]: time="2025-05-08T00:47:20.651501553Z" level=info msg="cleaning up dead shim" May 8 00:47:20.660321 env[1311]: time="2025-05-08T00:47:20.660248356Z" level=warning msg="cleanup warnings time=\"2025-05-08T00:47:20Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3756 runtime=io.containerd.runc.v2\n" May 8 00:47:20.902694 systemd[1]: Started sshd@11-10.0.0.82:22-10.0.0.1:50836.service. May 8 00:47:20.901000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.82:22-10.0.0.1:50836 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:20.904314 kernel: kauditd_printk_skb: 1 callbacks suppressed May 8 00:47:20.904423 kernel: audit: type=1130 audit(1746665240.901:325): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.82:22-10.0.0.1:50836 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:20.953000 audit[3769]: USER_ACCT pid=3769 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:20.954761 sshd[3769]: Accepted publickey for core from 10.0.0.1 port 50836 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:20.963150 kernel: audit: type=1101 audit(1746665240.953:326): pid=3769 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:20.962000 audit[3769]: CRED_ACQ pid=3769 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:20.964136 sshd[3769]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:20.969513 kernel: audit: type=1103 audit(1746665240.962:327): pid=3769 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:20.969633 kernel: audit: type=1006 audit(1746665240.962:328): pid=3769 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=12 res=1 May 8 00:47:20.969662 kernel: audit: type=1300 audit(1746665240.962:328): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffc2940a6a0 a2=3 a3=0 items=0 ppid=1 pid=3769 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=12 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:20.962000 audit[3769]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffc2940a6a0 a2=3 a3=0 items=0 ppid=1 pid=3769 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=12 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:20.970059 systemd-logind[1291]: New session 12 of user core. May 8 00:47:20.970782 systemd[1]: Started session-12.scope. May 8 00:47:20.962000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:20.974843 kernel: audit: type=1327 audit(1746665240.962:328): proctitle=737368643A20636F7265205B707269765D May 8 00:47:20.975000 audit[3769]: USER_START pid=3769 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:20.977000 audit[3772]: CRED_ACQ pid=3772 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:20.986696 kernel: audit: type=1105 audit(1746665240.975:329): pid=3769 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:20.986794 kernel: audit: type=1103 audit(1746665240.977:330): pid=3772 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.093842 sshd[3769]: pam_unix(sshd:session): session closed for user core May 8 00:47:21.095000 audit[3769]: USER_END pid=3769 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.096542 systemd[1]: Started sshd@12-10.0.0.82:22-10.0.0.1:50840.service. May 8 00:47:21.097719 systemd[1]: sshd@11-10.0.0.82:22-10.0.0.1:50836.service: Deactivated successfully. May 8 00:47:21.098840 systemd[1]: session-12.scope: Deactivated successfully. May 8 00:47:21.100310 systemd-logind[1291]: Session 12 logged out. Waiting for processes to exit. May 8 00:47:21.101638 systemd-logind[1291]: Removed session 12. May 8 00:47:21.095000 audit[3769]: CRED_DISP pid=3769 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.114153 kernel: audit: type=1106 audit(1746665241.095:331): pid=3769 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.114211 kernel: audit: type=1104 audit(1746665241.095:332): pid=3769 uid=0 auid=500 ses=12 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.095000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.82:22-10.0.0.1:50840 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:21.095000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@11-10.0.0.82:22-10.0.0.1:50836 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:21.141000 audit[3782]: USER_ACCT pid=3782 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.143158 sshd[3782]: Accepted publickey for core from 10.0.0.1 port 50840 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:21.142000 audit[3782]: CRED_ACQ pid=3782 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.142000 audit[3782]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffc3b22be30 a2=3 a3=0 items=0 ppid=1 pid=3782 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=13 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:21.142000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:21.144433 sshd[3782]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:21.148757 systemd-logind[1291]: New session 13 of user core. May 8 00:47:21.149875 systemd[1]: Started session-13.scope. May 8 00:47:21.157000 audit[3782]: USER_START pid=3782 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.158000 audit[3787]: CRED_ACQ pid=3787 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.353887 kubelet[2246]: I0508 00:47:21.353838 2246 scope.go:117] "RemoveContainer" containerID="af0424ae4b2bc3f499e117c178382d0b6c3efb4eae77dd7bb7145114f6fccfae" May 8 00:47:21.354430 kubelet[2246]: I0508 00:47:21.354298 2246 scope.go:117] "RemoveContainer" containerID="68edd6e95fddc102379a29cffc4ea28aed018582e663a0d5f501d88325e45eaf" May 8 00:47:21.354430 kubelet[2246]: E0508 00:47:21.354398 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:21.354982 kubelet[2246]: E0508 00:47:21.354812 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-node\" with CrashLoopBackOff: \"back-off 20s restarting failed container=calico-node pod=calico-node-k68b7_calico-system(ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5)\"" pod="calico-system/calico-node-k68b7" podUID="ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" May 8 00:47:21.355145 env[1311]: time="2025-05-08T00:47:21.355062622Z" level=info msg="RemoveContainer for \"af0424ae4b2bc3f499e117c178382d0b6c3efb4eae77dd7bb7145114f6fccfae\"" May 8 00:47:21.371441 sshd[3782]: pam_unix(sshd:session): session closed for user core May 8 00:47:21.372877 systemd[1]: Started sshd@13-10.0.0.82:22-10.0.0.1:50848.service. May 8 00:47:21.371000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.82:22-10.0.0.1:50848 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:21.372000 audit[3782]: USER_END pid=3782 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.372000 audit[3782]: CRED_DISP pid=3782 uid=0 auid=500 ses=13 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.375753 systemd[1]: sshd@12-10.0.0.82:22-10.0.0.1:50840.service: Deactivated successfully. May 8 00:47:21.374000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@12-10.0.0.82:22-10.0.0.1:50840 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:21.377080 systemd[1]: session-13.scope: Deactivated successfully. May 8 00:47:21.377191 systemd-logind[1291]: Session 13 logged out. Waiting for processes to exit. May 8 00:47:21.378336 systemd-logind[1291]: Removed session 13. May 8 00:47:21.391052 env[1311]: time="2025-05-08T00:47:21.390989107Z" level=info msg="RemoveContainer for \"af0424ae4b2bc3f499e117c178382d0b6c3efb4eae77dd7bb7145114f6fccfae\" returns successfully" May 8 00:47:21.417000 audit[3794]: USER_ACCT pid=3794 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.421335 sshd[3794]: Accepted publickey for core from 10.0.0.1 port 50848 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:21.420000 audit[3794]: CRED_ACQ pid=3794 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.421000 audit[3794]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffe6e3c8190 a2=3 a3=0 items=0 ppid=1 pid=3794 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=14 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:21.421000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:21.422745 sshd[3794]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:21.427858 systemd[1]: Started session-14.scope. May 8 00:47:21.432369 systemd-logind[1291]: New session 14 of user core. May 8 00:47:21.437000 audit[3794]: USER_START pid=3794 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.439000 audit[3799]: CRED_ACQ pid=3799 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.544757 sshd[3794]: pam_unix(sshd:session): session closed for user core May 8 00:47:21.544000 audit[3794]: USER_END pid=3794 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.544000 audit[3794]: CRED_DISP pid=3794 uid=0 auid=500 ses=14 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:21.547450 systemd[1]: sshd@13-10.0.0.82:22-10.0.0.1:50848.service: Deactivated successfully. May 8 00:47:21.546000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@13-10.0.0.82:22-10.0.0.1:50848 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:21.548706 systemd-logind[1291]: Session 14 logged out. Waiting for processes to exit. May 8 00:47:21.548867 systemd[1]: session-14.scope: Deactivated successfully. May 8 00:47:21.549875 systemd-logind[1291]: Removed session 14. May 8 00:47:22.166207 env[1311]: time="2025-05-08T00:47:22.166161188Z" level=info msg="StopPodSandbox for \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\"" May 8 00:47:22.167226 env[1311]: time="2025-05-08T00:47:22.167159645Z" level=info msg="StopPodSandbox for \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\"" May 8 00:47:22.197031 env[1311]: time="2025-05-08T00:47:22.196954227Z" level=error msg="StopPodSandbox for \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\" failed" error="failed to destroy network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:22.197306 kubelet[2246]: E0508 00:47:22.197249 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:47:22.197395 kubelet[2246]: E0508 00:47:22.197324 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b"} May 8 00:47:22.197395 kubelet[2246]: E0508 00:47:22.197365 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"ca6c7950-e2c3-4481-994b-8bd117426f80\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:22.197493 kubelet[2246]: E0508 00:47:22.197394 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"ca6c7950-e2c3-4481-994b-8bd117426f80\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-fd97795bb-zmwqb" podUID="ca6c7950-e2c3-4481-994b-8bd117426f80" May 8 00:47:22.203486 env[1311]: time="2025-05-08T00:47:22.203403866Z" level=error msg="StopPodSandbox for \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\" failed" error="failed to destroy network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:22.203673 kubelet[2246]: E0508 00:47:22.203639 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:47:22.203728 kubelet[2246]: E0508 00:47:22.203677 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd"} May 8 00:47:22.203728 kubelet[2246]: E0508 00:47:22.203704 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"c5c7ed4b-9783-4fe2-b85a-8452a359cb74\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:22.203805 kubelet[2246]: E0508 00:47:22.203723 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"c5c7ed4b-9783-4fe2-b85a-8452a359cb74\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-fd97795bb-tzgmm" podUID="c5c7ed4b-9783-4fe2-b85a-8452a359cb74" May 8 00:47:23.166478 env[1311]: time="2025-05-08T00:47:23.166428895Z" level=info msg="StopPodSandbox for \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\"" May 8 00:47:23.194138 env[1311]: time="2025-05-08T00:47:23.194046044Z" level=error msg="StopPodSandbox for \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\" failed" error="failed to destroy network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:23.194433 kubelet[2246]: E0508 00:47:23.194365 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:47:23.194744 kubelet[2246]: E0508 00:47:23.194442 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380"} May 8 00:47:23.194744 kubelet[2246]: E0508 00:47:23.194478 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"9c26b328-0da1-4b8c-b690-cd29958643ac\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:23.194744 kubelet[2246]: E0508 00:47:23.194509 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"9c26b328-0da1-4b8c-b690-cd29958643ac\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-c2qjt" podUID="9c26b328-0da1-4b8c-b690-cd29958643ac" May 8 00:47:24.166508 env[1311]: time="2025-05-08T00:47:24.166441181Z" level=info msg="StopPodSandbox for \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\"" May 8 00:47:24.169549 env[1311]: time="2025-05-08T00:47:24.169473748Z" level=info msg="StopPodSandbox for \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\"" May 8 00:47:24.196639 env[1311]: time="2025-05-08T00:47:24.196543488Z" level=error msg="StopPodSandbox for \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\" failed" error="failed to destroy network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:24.196980 kubelet[2246]: E0508 00:47:24.196905 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:47:24.197312 kubelet[2246]: E0508 00:47:24.197004 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce"} May 8 00:47:24.197312 kubelet[2246]: E0508 00:47:24.197054 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"863a2a82-a51b-4afc-a7be-990ee722b050\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:24.197312 kubelet[2246]: E0508 00:47:24.197087 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"863a2a82-a51b-4afc-a7be-990ee722b050\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-2s6tt" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" May 8 00:47:24.202864 env[1311]: time="2025-05-08T00:47:24.202792955Z" level=error msg="StopPodSandbox for \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\" failed" error="failed to destroy network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:24.203093 kubelet[2246]: E0508 00:47:24.203050 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:47:24.203178 kubelet[2246]: E0508 00:47:24.203105 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4"} May 8 00:47:24.203178 kubelet[2246]: E0508 00:47:24.203152 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"191fecaf-17dc-4c3f-a824-7b4430fd6b4f\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:24.203310 kubelet[2246]: E0508 00:47:24.203178 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"191fecaf-17dc-4c3f-a824-7b4430fd6b4f\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7f6ff4b949-8b2r4" podUID="191fecaf-17dc-4c3f-a824-7b4430fd6b4f" May 8 00:47:26.550622 systemd[1]: Started sshd@14-10.0.0.82:22-10.0.0.1:46502.service. May 8 00:47:26.583242 kernel: kauditd_printk_skb: 23 callbacks suppressed May 8 00:47:26.583384 kernel: audit: type=1130 audit(1746665246.549:352): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.82:22-10.0.0.1:46502 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:26.549000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.82:22-10.0.0.1:46502 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:26.622000 audit[3931]: USER_ACCT pid=3931 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:26.624206 sshd[3931]: Accepted publickey for core from 10.0.0.1 port 46502 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:26.625964 sshd[3931]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:26.629941 systemd-logind[1291]: New session 15 of user core. May 8 00:47:26.630230 systemd[1]: Started session-15.scope. May 8 00:47:26.624000 audit[3931]: CRED_ACQ pid=3931 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:26.697385 kernel: audit: type=1101 audit(1746665246.622:353): pid=3931 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:26.697456 kernel: audit: type=1103 audit(1746665246.624:354): pid=3931 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:26.698130 kernel: audit: type=1006 audit(1746665246.624:355): pid=3931 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=15 res=1 May 8 00:47:26.624000 audit[3931]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffe5a4990d0 a2=3 a3=0 items=0 ppid=1 pid=3931 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:26.704376 kernel: audit: type=1300 audit(1746665246.624:355): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffe5a4990d0 a2=3 a3=0 items=0 ppid=1 pid=3931 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=15 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:26.704434 kernel: audit: type=1327 audit(1746665246.624:355): proctitle=737368643A20636F7265205B707269765D May 8 00:47:26.624000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:26.705721 kernel: audit: type=1105 audit(1746665246.634:356): pid=3931 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:26.634000 audit[3931]: USER_START pid=3931 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:26.709913 kernel: audit: type=1103 audit(1746665246.636:357): pid=3934 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:26.636000 audit[3934]: CRED_ACQ pid=3934 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:26.882979 sshd[3931]: pam_unix(sshd:session): session closed for user core May 8 00:47:26.882000 audit[3931]: USER_END pid=3931 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:26.885664 systemd[1]: sshd@14-10.0.0.82:22-10.0.0.1:46502.service: Deactivated successfully. May 8 00:47:26.886493 systemd[1]: session-15.scope: Deactivated successfully. May 8 00:47:26.890909 systemd-logind[1291]: Session 15 logged out. Waiting for processes to exit. May 8 00:47:26.891870 systemd-logind[1291]: Removed session 15. May 8 00:47:26.883000 audit[3931]: CRED_DISP pid=3931 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:26.917647 kernel: audit: type=1106 audit(1746665246.882:358): pid=3931 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:26.917745 kernel: audit: type=1104 audit(1746665246.883:359): pid=3931 uid=0 auid=500 ses=15 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:26.884000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@14-10.0.0.82:22-10.0.0.1:46502 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:29.165684 kubelet[2246]: E0508 00:47:29.165638 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:31.886052 systemd[1]: Started sshd@15-10.0.0.82:22-10.0.0.1:46508.service. May 8 00:47:31.885000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.82:22-10.0.0.1:46508 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:31.896311 kernel: kauditd_printk_skb: 1 callbacks suppressed May 8 00:47:31.900603 kernel: audit: type=1130 audit(1746665251.885:361): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.82:22-10.0.0.1:46508 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:32.015762 sshd[3948]: Accepted publickey for core from 10.0.0.1 port 46508 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:32.014000 audit[3948]: USER_ACCT pid=3948 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:32.018589 sshd[3948]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:32.022666 systemd-logind[1291]: New session 16 of user core. May 8 00:47:32.023100 systemd[1]: Started session-16.scope. May 8 00:47:32.017000 audit[3948]: CRED_ACQ pid=3948 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:32.044840 kernel: audit: type=1101 audit(1746665252.014:362): pid=3948 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:32.044956 kernel: audit: type=1103 audit(1746665252.017:363): pid=3948 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:32.045000 kernel: audit: type=1006 audit(1746665252.017:364): pid=3948 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=16 res=1 May 8 00:47:32.017000 audit[3948]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffcabda3460 a2=3 a3=0 items=0 ppid=1 pid=3948 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:32.074970 kernel: audit: type=1300 audit(1746665252.017:364): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffcabda3460 a2=3 a3=0 items=0 ppid=1 pid=3948 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=16 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:32.075040 kernel: audit: type=1327 audit(1746665252.017:364): proctitle=737368643A20636F7265205B707269765D May 8 00:47:32.017000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:32.027000 audit[3948]: USER_START pid=3948 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:32.080806 kernel: audit: type=1105 audit(1746665252.027:365): pid=3948 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:32.080939 kernel: audit: type=1103 audit(1746665252.028:366): pid=3951 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:32.028000 audit[3951]: CRED_ACQ pid=3951 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:32.147005 sshd[3948]: pam_unix(sshd:session): session closed for user core May 8 00:47:32.147000 audit[3948]: USER_END pid=3948 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:32.150616 systemd[1]: sshd@15-10.0.0.82:22-10.0.0.1:46508.service: Deactivated successfully. May 8 00:47:32.151585 systemd[1]: session-16.scope: Deactivated successfully. May 8 00:47:32.147000 audit[3948]: CRED_DISP pid=3948 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:32.153785 systemd-logind[1291]: Session 16 logged out. Waiting for processes to exit. May 8 00:47:32.155202 systemd-logind[1291]: Removed session 16. May 8 00:47:32.157191 kernel: audit: type=1106 audit(1746665252.147:367): pid=3948 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:32.157252 kernel: audit: type=1104 audit(1746665252.147:368): pid=3948 uid=0 auid=500 ses=16 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:32.149000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@15-10.0.0.82:22-10.0.0.1:46508 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:33.165899 kubelet[2246]: E0508 00:47:33.165852 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:34.166235 env[1311]: time="2025-05-08T00:47:34.166173752Z" level=info msg="StopPodSandbox for \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\"" May 8 00:47:34.167272 env[1311]: time="2025-05-08T00:47:34.167239709Z" level=info msg="StopPodSandbox for \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\"" May 8 00:47:34.193574 env[1311]: time="2025-05-08T00:47:34.193452069Z" level=error msg="StopPodSandbox for \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\" failed" error="failed to destroy network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:34.193914 kubelet[2246]: E0508 00:47:34.193805 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:47:34.193914 kubelet[2246]: E0508 00:47:34.193881 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd"} May 8 00:47:34.194270 kubelet[2246]: E0508 00:47:34.193941 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"c5c7ed4b-9783-4fe2-b85a-8452a359cb74\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:34.194270 kubelet[2246]: E0508 00:47:34.193975 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"c5c7ed4b-9783-4fe2-b85a-8452a359cb74\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-fd97795bb-tzgmm" podUID="c5c7ed4b-9783-4fe2-b85a-8452a359cb74" May 8 00:47:34.194399 env[1311]: time="2025-05-08T00:47:34.194282297Z" level=error msg="StopPodSandbox for \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\" failed" error="failed to destroy network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:34.194527 kubelet[2246]: E0508 00:47:34.194488 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:47:34.194527 kubelet[2246]: E0508 00:47:34.194517 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2"} May 8 00:47:34.194610 kubelet[2246]: E0508 00:47:34.194545 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"435e64db-0f2e-4f05-8fcf-eefdac14188d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:34.194610 kubelet[2246]: E0508 00:47:34.194562 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"435e64db-0f2e-4f05-8fcf-eefdac14188d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-8xjk8" podUID="435e64db-0f2e-4f05-8fcf-eefdac14188d" May 8 00:47:34.888133 kubelet[2246]: I0508 00:47:34.888068 2246 scope.go:117] "RemoveContainer" containerID="68edd6e95fddc102379a29cffc4ea28aed018582e663a0d5f501d88325e45eaf" May 8 00:47:34.888373 kubelet[2246]: E0508 00:47:34.888200 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:34.888671 kubelet[2246]: E0508 00:47:34.888636 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"calico-node\" with CrashLoopBackOff: \"back-off 20s restarting failed container=calico-node pod=calico-node-k68b7_calico-system(ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5)\"" pod="calico-system/calico-node-k68b7" podUID="ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" May 8 00:47:36.166969 env[1311]: time="2025-05-08T00:47:36.166904280Z" level=info msg="StopPodSandbox for \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\"" May 8 00:47:36.170294 env[1311]: time="2025-05-08T00:47:36.170246060Z" level=info msg="StopPodSandbox for \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\"" May 8 00:47:36.198588 env[1311]: time="2025-05-08T00:47:36.198524776Z" level=error msg="StopPodSandbox for \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\" failed" error="failed to destroy network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:36.199049 kubelet[2246]: E0508 00:47:36.198990 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:47:36.199384 kubelet[2246]: E0508 00:47:36.199068 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380"} May 8 00:47:36.199384 kubelet[2246]: E0508 00:47:36.199129 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"9c26b328-0da1-4b8c-b690-cd29958643ac\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:36.199384 kubelet[2246]: E0508 00:47:36.199160 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"9c26b328-0da1-4b8c-b690-cd29958643ac\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-c2qjt" podUID="9c26b328-0da1-4b8c-b690-cd29958643ac" May 8 00:47:36.199845 env[1311]: time="2025-05-08T00:47:36.199811792Z" level=error msg="StopPodSandbox for \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\" failed" error="failed to destroy network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:36.199987 kubelet[2246]: E0508 00:47:36.199947 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:47:36.200059 kubelet[2246]: E0508 00:47:36.199991 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce"} May 8 00:47:36.200059 kubelet[2246]: E0508 00:47:36.200022 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"863a2a82-a51b-4afc-a7be-990ee722b050\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:36.200059 kubelet[2246]: E0508 00:47:36.200041 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"863a2a82-a51b-4afc-a7be-990ee722b050\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-2s6tt" podUID="863a2a82-a51b-4afc-a7be-990ee722b050" May 8 00:47:37.150577 systemd[1]: Started sshd@16-10.0.0.82:22-10.0.0.1:37166.service. May 8 00:47:37.149000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.82:22-10.0.0.1:37166 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:37.151810 kernel: kauditd_printk_skb: 1 callbacks suppressed May 8 00:47:37.151874 kernel: audit: type=1130 audit(1746665257.149:370): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.82:22-10.0.0.1:37166 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:37.166353 env[1311]: time="2025-05-08T00:47:37.166272842Z" level=info msg="StopPodSandbox for \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\"" May 8 00:47:37.196279 env[1311]: time="2025-05-08T00:47:37.196191759Z" level=error msg="StopPodSandbox for \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\" failed" error="failed to destroy network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:37.196772 kubelet[2246]: E0508 00:47:37.196442 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:47:37.196772 kubelet[2246]: E0508 00:47:37.196490 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b"} May 8 00:47:37.196772 kubelet[2246]: E0508 00:47:37.196540 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"ca6c7950-e2c3-4481-994b-8bd117426f80\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:37.196772 kubelet[2246]: E0508 00:47:37.196565 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"ca6c7950-e2c3-4481-994b-8bd117426f80\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-fd97795bb-zmwqb" podUID="ca6c7950-e2c3-4481-994b-8bd117426f80" May 8 00:47:37.197000 audit[4057]: USER_ACCT pid=4057 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:37.198867 sshd[4057]: Accepted publickey for core from 10.0.0.1 port 37166 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:37.202572 sshd[4057]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:37.201000 audit[4057]: CRED_ACQ pid=4057 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:37.206791 systemd-logind[1291]: New session 17 of user core. May 8 00:47:37.207644 systemd[1]: Started session-17.scope. May 8 00:47:37.207934 kernel: audit: type=1101 audit(1746665257.197:371): pid=4057 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:37.207985 kernel: audit: type=1103 audit(1746665257.201:372): pid=4057 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:37.210991 kernel: audit: type=1006 audit(1746665257.201:373): pid=4057 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=17 res=1 May 8 00:47:37.211167 kernel: audit: type=1300 audit(1746665257.201:373): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffef01bb240 a2=3 a3=0 items=0 ppid=1 pid=4057 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=17 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:37.201000 audit[4057]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffef01bb240 a2=3 a3=0 items=0 ppid=1 pid=4057 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=17 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:37.216402 kernel: audit: type=1327 audit(1746665257.201:373): proctitle=737368643A20636F7265205B707269765D May 8 00:47:37.201000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:37.218192 kernel: audit: type=1105 audit(1746665257.214:374): pid=4057 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:37.214000 audit[4057]: USER_START pid=4057 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:37.216000 audit[4083]: CRED_ACQ pid=4083 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:37.227697 kernel: audit: type=1103 audit(1746665257.216:375): pid=4083 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:37.326600 sshd[4057]: pam_unix(sshd:session): session closed for user core May 8 00:47:37.326000 audit[4057]: USER_END pid=4057 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:37.328789 systemd[1]: sshd@16-10.0.0.82:22-10.0.0.1:37166.service: Deactivated successfully. May 8 00:47:37.329937 systemd-logind[1291]: Session 17 logged out. Waiting for processes to exit. May 8 00:47:37.329994 systemd[1]: session-17.scope: Deactivated successfully. May 8 00:47:37.331415 systemd-logind[1291]: Removed session 17. May 8 00:47:37.326000 audit[4057]: CRED_DISP pid=4057 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:37.336516 kernel: audit: type=1106 audit(1746665257.326:376): pid=4057 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:37.336710 kernel: audit: type=1104 audit(1746665257.326:377): pid=4057 uid=0 auid=500 ses=17 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:37.328000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@16-10.0.0.82:22-10.0.0.1:37166 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:38.165817 env[1311]: time="2025-05-08T00:47:38.165763278Z" level=info msg="StopPodSandbox for \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\"" May 8 00:47:38.192851 env[1311]: time="2025-05-08T00:47:38.192759678Z" level=error msg="StopPodSandbox for \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\" failed" error="failed to destroy network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 8 00:47:38.193082 kubelet[2246]: E0508 00:47:38.193031 2246 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:47:38.193414 kubelet[2246]: E0508 00:47:38.193099 2246 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4"} May 8 00:47:38.193414 kubelet[2246]: E0508 00:47:38.193149 2246 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"191fecaf-17dc-4c3f-a824-7b4430fd6b4f\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" May 8 00:47:38.193414 kubelet[2246]: E0508 00:47:38.193173 2246 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"191fecaf-17dc-4c3f-a824-7b4430fd6b4f\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7f6ff4b949-8b2r4" podUID="191fecaf-17dc-4c3f-a824-7b4430fd6b4f" May 8 00:47:39.281267 env[1311]: time="2025-05-08T00:47:39.281206047Z" level=info msg="StopPodSandbox for \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\"" May 8 00:47:39.281856 env[1311]: time="2025-05-08T00:47:39.281297040Z" level=info msg="Container to stop \"a3368ce3ed5ee729b44408f60e027497ad2d3647a905935456d9b4022a7ca29c\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 00:47:39.281856 env[1311]: time="2025-05-08T00:47:39.281314944Z" level=info msg="Container to stop \"0031612f63728aa09c3b77481dc5f2be8245c926a8e17c06641844ae7d393b8e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 00:47:39.281856 env[1311]: time="2025-05-08T00:47:39.281325514Z" level=info msg="Container to stop \"68edd6e95fddc102379a29cffc4ea28aed018582e663a0d5f501d88325e45eaf\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 8 00:47:39.284543 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808-shm.mount: Deactivated successfully. May 8 00:47:39.312810 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808-rootfs.mount: Deactivated successfully. May 8 00:47:39.325886 env[1311]: time="2025-05-08T00:47:39.325775228Z" level=info msg="shim disconnected" id=c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808 May 8 00:47:39.325886 env[1311]: time="2025-05-08T00:47:39.325829171Z" level=warning msg="cleaning up after shim disconnected" id=c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808 namespace=k8s.io May 8 00:47:39.325886 env[1311]: time="2025-05-08T00:47:39.325838899Z" level=info msg="cleaning up dead shim" May 8 00:47:39.339447 env[1311]: time="2025-05-08T00:47:39.339389477Z" level=warning msg="cleanup warnings time=\"2025-05-08T00:47:39Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4136 runtime=io.containerd.runc.v2\n" May 8 00:47:39.339731 env[1311]: time="2025-05-08T00:47:39.339702972Z" level=info msg="TearDown network for sandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" successfully" May 8 00:47:39.339731 env[1311]: time="2025-05-08T00:47:39.339728330Z" level=info msg="StopPodSandbox for \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" returns successfully" May 8 00:47:39.387791 kubelet[2246]: I0508 00:47:39.387717 2246 topology_manager.go:215] "Topology Admit Handler" podUID="43189189-d4b3-46f2-9f87-3c934dc5751e" podNamespace="calico-system" podName="calico-node-c9gns" May 8 00:47:39.388320 kubelet[2246]: E0508 00:47:39.387847 2246 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" containerName="flexvol-driver" May 8 00:47:39.388320 kubelet[2246]: E0508 00:47:39.387863 2246 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" containerName="calico-node" May 8 00:47:39.388320 kubelet[2246]: E0508 00:47:39.387871 2246 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" containerName="install-cni" May 8 00:47:39.388320 kubelet[2246]: E0508 00:47:39.387878 2246 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" containerName="calico-node" May 8 00:47:39.388320 kubelet[2246]: I0508 00:47:39.387922 2246 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" containerName="calico-node" May 8 00:47:39.388320 kubelet[2246]: I0508 00:47:39.387935 2246 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" containerName="calico-node" May 8 00:47:39.388320 kubelet[2246]: E0508 00:47:39.387959 2246 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" containerName="calico-node" May 8 00:47:39.388320 kubelet[2246]: I0508 00:47:39.387990 2246 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" containerName="calico-node" May 8 00:47:39.396243 kubelet[2246]: I0508 00:47:39.396207 2246 scope.go:117] "RemoveContainer" containerID="68edd6e95fddc102379a29cffc4ea28aed018582e663a0d5f501d88325e45eaf" May 8 00:47:39.397490 env[1311]: time="2025-05-08T00:47:39.397435075Z" level=info msg="RemoveContainer for \"68edd6e95fddc102379a29cffc4ea28aed018582e663a0d5f501d88325e45eaf\"" May 8 00:47:39.405335 env[1311]: time="2025-05-08T00:47:39.405271754Z" level=info msg="RemoveContainer for \"68edd6e95fddc102379a29cffc4ea28aed018582e663a0d5f501d88325e45eaf\" returns successfully" May 8 00:47:39.405846 kubelet[2246]: I0508 00:47:39.405672 2246 scope.go:117] "RemoveContainer" containerID="0031612f63728aa09c3b77481dc5f2be8245c926a8e17c06641844ae7d393b8e" May 8 00:47:39.407318 kubelet[2246]: I0508 00:47:39.407279 2246 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-policysync\") pod \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " May 8 00:47:39.407393 kubelet[2246]: I0508 00:47:39.407322 2246 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-node-certs\") pod \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " May 8 00:47:39.407393 kubelet[2246]: I0508 00:47:39.407346 2246 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-xtables-lock\") pod \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " May 8 00:47:39.407393 kubelet[2246]: I0508 00:47:39.407366 2246 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-var-run-calico\") pod \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " May 8 00:47:39.407393 kubelet[2246]: I0508 00:47:39.407387 2246 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-flexvol-driver-host\") pod \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " May 8 00:47:39.407551 kubelet[2246]: I0508 00:47:39.407407 2246 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-cni-log-dir\") pod \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " May 8 00:47:39.407551 kubelet[2246]: I0508 00:47:39.407430 2246 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-tigera-ca-bundle\") pod \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " May 8 00:47:39.407551 kubelet[2246]: I0508 00:47:39.407447 2246 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-lib-modules\") pod \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " May 8 00:47:39.407551 kubelet[2246]: I0508 00:47:39.407465 2246 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-cni-net-dir\") pod \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " May 8 00:47:39.407551 kubelet[2246]: I0508 00:47:39.407485 2246 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-var-lib-calico\") pod \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " May 8 00:47:39.407551 kubelet[2246]: I0508 00:47:39.407510 2246 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-cni-bin-dir\") pod \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " May 8 00:47:39.407766 kubelet[2246]: I0508 00:47:39.407532 2246 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbkxf\" (UniqueName: \"kubernetes.io/projected/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-kube-api-access-qbkxf\") pod \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\" (UID: \"ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5\") " May 8 00:47:39.407766 kubelet[2246]: I0508 00:47:39.407585 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/43189189-d4b3-46f2-9f87-3c934dc5751e-var-run-calico\") pod \"calico-node-c9gns\" (UID: \"43189189-d4b3-46f2-9f87-3c934dc5751e\") " pod="calico-system/calico-node-c9gns" May 8 00:47:39.407766 kubelet[2246]: I0508 00:47:39.407611 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/43189189-d4b3-46f2-9f87-3c934dc5751e-cni-bin-dir\") pod \"calico-node-c9gns\" (UID: \"43189189-d4b3-46f2-9f87-3c934dc5751e\") " pod="calico-system/calico-node-c9gns" May 8 00:47:39.407766 kubelet[2246]: I0508 00:47:39.407634 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/43189189-d4b3-46f2-9f87-3c934dc5751e-node-certs\") pod \"calico-node-c9gns\" (UID: \"43189189-d4b3-46f2-9f87-3c934dc5751e\") " pod="calico-system/calico-node-c9gns" May 8 00:47:39.407766 kubelet[2246]: I0508 00:47:39.407661 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dj46\" (UniqueName: \"kubernetes.io/projected/43189189-d4b3-46f2-9f87-3c934dc5751e-kube-api-access-4dj46\") pod \"calico-node-c9gns\" (UID: \"43189189-d4b3-46f2-9f87-3c934dc5751e\") " pod="calico-system/calico-node-c9gns" May 8 00:47:39.407926 kubelet[2246]: I0508 00:47:39.407682 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/43189189-d4b3-46f2-9f87-3c934dc5751e-policysync\") pod \"calico-node-c9gns\" (UID: \"43189189-d4b3-46f2-9f87-3c934dc5751e\") " pod="calico-system/calico-node-c9gns" May 8 00:47:39.407926 kubelet[2246]: I0508 00:47:39.407706 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/43189189-d4b3-46f2-9f87-3c934dc5751e-cni-log-dir\") pod \"calico-node-c9gns\" (UID: \"43189189-d4b3-46f2-9f87-3c934dc5751e\") " pod="calico-system/calico-node-c9gns" May 8 00:47:39.407926 kubelet[2246]: I0508 00:47:39.407726 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/43189189-d4b3-46f2-9f87-3c934dc5751e-xtables-lock\") pod \"calico-node-c9gns\" (UID: \"43189189-d4b3-46f2-9f87-3c934dc5751e\") " pod="calico-system/calico-node-c9gns" May 8 00:47:39.407926 kubelet[2246]: I0508 00:47:39.407751 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43189189-d4b3-46f2-9f87-3c934dc5751e-tigera-ca-bundle\") pod \"calico-node-c9gns\" (UID: \"43189189-d4b3-46f2-9f87-3c934dc5751e\") " pod="calico-system/calico-node-c9gns" May 8 00:47:39.407926 kubelet[2246]: I0508 00:47:39.407774 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/43189189-d4b3-46f2-9f87-3c934dc5751e-flexvol-driver-host\") pod \"calico-node-c9gns\" (UID: \"43189189-d4b3-46f2-9f87-3c934dc5751e\") " pod="calico-system/calico-node-c9gns" May 8 00:47:39.408105 kubelet[2246]: I0508 00:47:39.407842 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/43189189-d4b3-46f2-9f87-3c934dc5751e-lib-modules\") pod \"calico-node-c9gns\" (UID: \"43189189-d4b3-46f2-9f87-3c934dc5751e\") " pod="calico-system/calico-node-c9gns" May 8 00:47:39.408105 kubelet[2246]: I0508 00:47:39.407864 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/43189189-d4b3-46f2-9f87-3c934dc5751e-var-lib-calico\") pod \"calico-node-c9gns\" (UID: \"43189189-d4b3-46f2-9f87-3c934dc5751e\") " pod="calico-system/calico-node-c9gns" May 8 00:47:39.408105 kubelet[2246]: I0508 00:47:39.407890 2246 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/43189189-d4b3-46f2-9f87-3c934dc5751e-cni-net-dir\") pod \"calico-node-c9gns\" (UID: \"43189189-d4b3-46f2-9f87-3c934dc5751e\") " pod="calico-system/calico-node-c9gns" May 8 00:47:39.408105 kubelet[2246]: I0508 00:47:39.408020 2246 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-policysync" (OuterVolumeSpecName: "policysync") pod "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" (UID: "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5"). InnerVolumeSpecName "policysync". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 00:47:39.410354 kubelet[2246]: I0508 00:47:39.410307 2246 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" (UID: "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 00:47:39.410448 kubelet[2246]: I0508 00:47:39.410382 2246 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-var-run-calico" (OuterVolumeSpecName: "var-run-calico") pod "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" (UID: "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5"). InnerVolumeSpecName "var-run-calico". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 00:47:39.410448 kubelet[2246]: I0508 00:47:39.410413 2246 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-flexvol-driver-host" (OuterVolumeSpecName: "flexvol-driver-host") pod "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" (UID: "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5"). InnerVolumeSpecName "flexvol-driver-host". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 00:47:39.410448 kubelet[2246]: I0508 00:47:39.410437 2246 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-cni-log-dir" (OuterVolumeSpecName: "cni-log-dir") pod "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" (UID: "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5"). InnerVolumeSpecName "cni-log-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 00:47:39.410799 kubelet[2246]: I0508 00:47:39.410762 2246 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-var-lib-calico" (OuterVolumeSpecName: "var-lib-calico") pod "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" (UID: "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5"). InnerVolumeSpecName "var-lib-calico". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 00:47:39.410799 kubelet[2246]: I0508 00:47:39.410797 2246 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" (UID: "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 00:47:39.410900 kubelet[2246]: I0508 00:47:39.410832 2246 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-cni-net-dir" (OuterVolumeSpecName: "cni-net-dir") pod "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" (UID: "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5"). InnerVolumeSpecName "cni-net-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 00:47:39.410900 kubelet[2246]: I0508 00:47:39.410837 2246 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-cni-bin-dir" (OuterVolumeSpecName: "cni-bin-dir") pod "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" (UID: "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5"). InnerVolumeSpecName "cni-bin-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 8 00:47:39.417095 systemd[1]: var-lib-kubelet-pods-ab7a4e1f\x2dcf30\x2d4bf5\x2d85d0\x2dcb86b9a282c5-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dqbkxf.mount: Deactivated successfully. May 8 00:47:39.420025 systemd[1]: var-lib-kubelet-pods-ab7a4e1f\x2dcf30\x2d4bf5\x2d85d0\x2dcb86b9a282c5-volume\x2dsubpaths-tigera\x2dca\x2dbundle-calico\x2dnode-1.mount: Deactivated successfully. May 8 00:47:39.421421 kubelet[2246]: I0508 00:47:39.421367 2246 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-tigera-ca-bundle" (OuterVolumeSpecName: "tigera-ca-bundle") pod "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" (UID: "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5"). InnerVolumeSpecName "tigera-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 8 00:47:39.421493 kubelet[2246]: I0508 00:47:39.421425 2246 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-kube-api-access-qbkxf" (OuterVolumeSpecName: "kube-api-access-qbkxf") pod "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" (UID: "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5"). InnerVolumeSpecName "kube-api-access-qbkxf". PluginName "kubernetes.io/projected", VolumeGidValue "" May 8 00:47:39.421830 env[1311]: time="2025-05-08T00:47:39.421779978Z" level=info msg="RemoveContainer for \"0031612f63728aa09c3b77481dc5f2be8245c926a8e17c06641844ae7d393b8e\"" May 8 00:47:39.424593 systemd[1]: var-lib-kubelet-pods-ab7a4e1f\x2dcf30\x2d4bf5\x2d85d0\x2dcb86b9a282c5-volumes-kubernetes.io\x7esecret-node\x2dcerts.mount: Deactivated successfully. May 8 00:47:39.425267 kubelet[2246]: I0508 00:47:39.425235 2246 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-node-certs" (OuterVolumeSpecName: "node-certs") pod "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" (UID: "ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5"). InnerVolumeSpecName "node-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" May 8 00:47:39.427805 env[1311]: time="2025-05-08T00:47:39.427760122Z" level=info msg="RemoveContainer for \"0031612f63728aa09c3b77481dc5f2be8245c926a8e17c06641844ae7d393b8e\" returns successfully" May 8 00:47:39.427998 kubelet[2246]: I0508 00:47:39.427971 2246 scope.go:117] "RemoveContainer" containerID="a3368ce3ed5ee729b44408f60e027497ad2d3647a905935456d9b4022a7ca29c" May 8 00:47:39.429390 env[1311]: time="2025-05-08T00:47:39.429355922Z" level=info msg="RemoveContainer for \"a3368ce3ed5ee729b44408f60e027497ad2d3647a905935456d9b4022a7ca29c\"" May 8 00:47:39.432837 env[1311]: time="2025-05-08T00:47:39.432807335Z" level=info msg="RemoveContainer for \"a3368ce3ed5ee729b44408f60e027497ad2d3647a905935456d9b4022a7ca29c\" returns successfully" May 8 00:47:39.508734 kubelet[2246]: I0508 00:47:39.508674 2246 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-xtables-lock\") on node \"localhost\" DevicePath \"\"" May 8 00:47:39.509063 kubelet[2246]: I0508 00:47:39.509041 2246 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-lib-modules\") on node \"localhost\" DevicePath \"\"" May 8 00:47:39.509252 kubelet[2246]: I0508 00:47:39.509232 2246 reconciler_common.go:289] "Volume detached for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-cni-net-dir\") on node \"localhost\" DevicePath \"\"" May 8 00:47:39.509409 kubelet[2246]: I0508 00:47:39.509368 2246 reconciler_common.go:289] "Volume detached for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-var-lib-calico\") on node \"localhost\" DevicePath \"\"" May 8 00:47:39.509409 kubelet[2246]: I0508 00:47:39.509393 2246 reconciler_common.go:289] "Volume detached for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-var-run-calico\") on node \"localhost\" DevicePath \"\"" May 8 00:47:39.509409 kubelet[2246]: I0508 00:47:39.509407 2246 reconciler_common.go:289] "Volume detached for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-flexvol-driver-host\") on node \"localhost\" DevicePath \"\"" May 8 00:47:39.509562 kubelet[2246]: I0508 00:47:39.509420 2246 reconciler_common.go:289] "Volume detached for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-cni-log-dir\") on node \"localhost\" DevicePath \"\"" May 8 00:47:39.509562 kubelet[2246]: I0508 00:47:39.509430 2246 reconciler_common.go:289] "Volume detached for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-tigera-ca-bundle\") on node \"localhost\" DevicePath \"\"" May 8 00:47:39.509562 kubelet[2246]: I0508 00:47:39.509441 2246 reconciler_common.go:289] "Volume detached for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-cni-bin-dir\") on node \"localhost\" DevicePath \"\"" May 8 00:47:39.509562 kubelet[2246]: I0508 00:47:39.509452 2246 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-qbkxf\" (UniqueName: \"kubernetes.io/projected/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-kube-api-access-qbkxf\") on node \"localhost\" DevicePath \"\"" May 8 00:47:39.509562 kubelet[2246]: I0508 00:47:39.509462 2246 reconciler_common.go:289] "Volume detached for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-policysync\") on node \"localhost\" DevicePath \"\"" May 8 00:47:39.509562 kubelet[2246]: I0508 00:47:39.509472 2246 reconciler_common.go:289] "Volume detached for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5-node-certs\") on node \"localhost\" DevicePath \"\"" May 8 00:47:39.692299 kubelet[2246]: E0508 00:47:39.692150 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:39.692832 env[1311]: time="2025-05-08T00:47:39.692786885Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-c9gns,Uid:43189189-d4b3-46f2-9f87-3c934dc5751e,Namespace:calico-system,Attempt:0,}" May 8 00:47:39.713430 env[1311]: time="2025-05-08T00:47:39.713321333Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 8 00:47:39.713622 env[1311]: time="2025-05-08T00:47:39.713382420Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 8 00:47:39.713622 env[1311]: time="2025-05-08T00:47:39.713405112Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 8 00:47:39.713858 env[1311]: time="2025-05-08T00:47:39.713828396Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/c43cd87b0239100f2802359fe24cf205cc28976eb60480577bbdf6d0683539fc pid=4159 runtime=io.containerd.runc.v2 May 8 00:47:39.758227 env[1311]: time="2025-05-08T00:47:39.758173081Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-c9gns,Uid:43189189-d4b3-46f2-9f87-3c934dc5751e,Namespace:calico-system,Attempt:0,} returns sandbox id \"c43cd87b0239100f2802359fe24cf205cc28976eb60480577bbdf6d0683539fc\"" May 8 00:47:39.759526 kubelet[2246]: E0508 00:47:39.758954 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:39.761538 env[1311]: time="2025-05-08T00:47:39.761500619Z" level=info msg="CreateContainer within sandbox \"c43cd87b0239100f2802359fe24cf205cc28976eb60480577bbdf6d0683539fc\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" May 8 00:47:40.166387 kubelet[2246]: E0508 00:47:40.166336 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:40.168215 kubelet[2246]: I0508 00:47:40.168174 2246 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5" path="/var/lib/kubelet/pods/ab7a4e1f-cf30-4bf5-85d0-cb86b9a282c5/volumes" May 8 00:47:40.277924 env[1311]: time="2025-05-08T00:47:40.277835330Z" level=info msg="CreateContainer within sandbox \"c43cd87b0239100f2802359fe24cf205cc28976eb60480577bbdf6d0683539fc\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"449db13ee6ae812ba1f83b330ab46f1b868496aa92f84d31b599fea3160ebffb\"" May 8 00:47:40.278989 env[1311]: time="2025-05-08T00:47:40.278946018Z" level=info msg="StartContainer for \"449db13ee6ae812ba1f83b330ab46f1b868496aa92f84d31b599fea3160ebffb\"" May 8 00:47:40.357922 env[1311]: time="2025-05-08T00:47:40.357838317Z" level=info msg="StartContainer for \"449db13ee6ae812ba1f83b330ab46f1b868496aa92f84d31b599fea3160ebffb\" returns successfully" May 8 00:47:40.400706 kubelet[2246]: E0508 00:47:40.400652 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:40.415323 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-449db13ee6ae812ba1f83b330ab46f1b868496aa92f84d31b599fea3160ebffb-rootfs.mount: Deactivated successfully. May 8 00:47:40.436611 env[1311]: time="2025-05-08T00:47:40.436464561Z" level=info msg="shim disconnected" id=449db13ee6ae812ba1f83b330ab46f1b868496aa92f84d31b599fea3160ebffb May 8 00:47:40.436611 env[1311]: time="2025-05-08T00:47:40.436513084Z" level=warning msg="cleaning up after shim disconnected" id=449db13ee6ae812ba1f83b330ab46f1b868496aa92f84d31b599fea3160ebffb namespace=k8s.io May 8 00:47:40.436611 env[1311]: time="2025-05-08T00:47:40.436522863Z" level=info msg="cleaning up dead shim" May 8 00:47:40.444149 env[1311]: time="2025-05-08T00:47:40.444064307Z" level=warning msg="cleanup warnings time=\"2025-05-08T00:47:40Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4248 runtime=io.containerd.runc.v2\n" May 8 00:47:41.404709 kubelet[2246]: E0508 00:47:41.404668 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:41.407154 env[1311]: time="2025-05-08T00:47:41.407072625Z" level=info msg="CreateContainer within sandbox \"c43cd87b0239100f2802359fe24cf205cc28976eb60480577bbdf6d0683539fc\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" May 8 00:47:41.715766 env[1311]: time="2025-05-08T00:47:41.715554931Z" level=info msg="CreateContainer within sandbox \"c43cd87b0239100f2802359fe24cf205cc28976eb60480577bbdf6d0683539fc\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"f93f1175e106ac49b2db1e44df5bf4074f450527733224280a56c85786a4f372\"" May 8 00:47:41.716461 env[1311]: time="2025-05-08T00:47:41.716401187Z" level=info msg="StartContainer for \"f93f1175e106ac49b2db1e44df5bf4074f450527733224280a56c85786a4f372\"" May 8 00:47:41.791178 env[1311]: time="2025-05-08T00:47:41.791087717Z" level=info msg="StartContainer for \"f93f1175e106ac49b2db1e44df5bf4074f450527733224280a56c85786a4f372\" returns successfully" May 8 00:47:42.330978 systemd[1]: Started sshd@17-10.0.0.82:22-10.0.0.1:37176.service. May 8 00:47:42.331000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-10.0.0.82:22-10.0.0.1:37176 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:42.332501 kernel: kauditd_printk_skb: 1 callbacks suppressed May 8 00:47:42.332595 kernel: audit: type=1130 audit(1746665262.331:379): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-10.0.0.82:22-10.0.0.1:37176 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:42.375000 audit[4297]: USER_ACCT pid=4297 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:42.375673 sshd[4297]: Accepted publickey for core from 10.0.0.1 port 37176 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:42.377000 audit[4297]: CRED_ACQ pid=4297 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:42.383787 kernel: audit: type=1101 audit(1746665262.375:380): pid=4297 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:42.383842 kernel: audit: type=1103 audit(1746665262.377:381): pid=4297 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:42.383874 kernel: audit: type=1006 audit(1746665262.377:382): pid=4297 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=18 res=1 May 8 00:47:42.386119 kernel: audit: type=1300 audit(1746665262.377:382): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffd3b61d840 a2=3 a3=0 items=0 ppid=1 pid=4297 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=18 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:42.377000 audit[4297]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffd3b61d840 a2=3 a3=0 items=0 ppid=1 pid=4297 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=18 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:42.387297 sshd[4297]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:42.390321 kernel: audit: type=1327 audit(1746665262.377:382): proctitle=737368643A20636F7265205B707269765D May 8 00:47:42.377000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:42.391040 systemd-logind[1291]: New session 18 of user core. May 8 00:47:42.392142 systemd[1]: Started session-18.scope. May 8 00:47:42.396000 audit[4297]: USER_START pid=4297 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:42.398000 audit[4300]: CRED_ACQ pid=4300 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:42.405042 kernel: audit: type=1105 audit(1746665262.396:383): pid=4297 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:42.405098 kernel: audit: type=1103 audit(1746665262.398:384): pid=4300 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:42.408301 kubelet[2246]: E0508 00:47:42.408275 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:42.515487 sshd[4297]: pam_unix(sshd:session): session closed for user core May 8 00:47:42.516000 audit[4297]: USER_END pid=4297 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:42.520626 systemd[1]: sshd@17-10.0.0.82:22-10.0.0.1:37176.service: Deactivated successfully. May 8 00:47:42.521830 systemd[1]: session-18.scope: Deactivated successfully. May 8 00:47:42.516000 audit[4297]: CRED_DISP pid=4297 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:42.522322 systemd-logind[1291]: Session 18 logged out. Waiting for processes to exit. May 8 00:47:42.523318 systemd-logind[1291]: Removed session 18. May 8 00:47:42.526458 kernel: audit: type=1106 audit(1746665262.516:385): pid=4297 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:42.526563 kernel: audit: type=1104 audit(1746665262.516:386): pid=4297 uid=0 auid=500 ses=18 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:42.520000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@17-10.0.0.82:22-10.0.0.1:37176 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:43.180493 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f93f1175e106ac49b2db1e44df5bf4074f450527733224280a56c85786a4f372-rootfs.mount: Deactivated successfully. May 8 00:47:43.182341 env[1311]: time="2025-05-08T00:47:43.182293601Z" level=info msg="shim disconnected" id=f93f1175e106ac49b2db1e44df5bf4074f450527733224280a56c85786a4f372 May 8 00:47:43.182634 env[1311]: time="2025-05-08T00:47:43.182345269Z" level=warning msg="cleaning up after shim disconnected" id=f93f1175e106ac49b2db1e44df5bf4074f450527733224280a56c85786a4f372 namespace=k8s.io May 8 00:47:43.182634 env[1311]: time="2025-05-08T00:47:43.182355628Z" level=info msg="cleaning up dead shim" May 8 00:47:43.189541 env[1311]: time="2025-05-08T00:47:43.189495541Z" level=warning msg="cleanup warnings time=\"2025-05-08T00:47:43Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4324 runtime=io.containerd.runc.v2\n" May 8 00:47:43.411837 kubelet[2246]: E0508 00:47:43.411804 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:43.428180 env[1311]: time="2025-05-08T00:47:43.428123065Z" level=info msg="CreateContainer within sandbox \"c43cd87b0239100f2802359fe24cf205cc28976eb60480577bbdf6d0683539fc\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" May 8 00:47:43.447057 env[1311]: time="2025-05-08T00:47:43.446904689Z" level=info msg="CreateContainer within sandbox \"c43cd87b0239100f2802359fe24cf205cc28976eb60480577bbdf6d0683539fc\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"4d1318708da5e2a4b1ad3316a4f2763d39d0e1832a54721c31612b0afdee5ae3\"" May 8 00:47:43.448257 env[1311]: time="2025-05-08T00:47:43.448224603Z" level=info msg="StartContainer for \"4d1318708da5e2a4b1ad3316a4f2763d39d0e1832a54721c31612b0afdee5ae3\"" May 8 00:47:43.505821 env[1311]: time="2025-05-08T00:47:43.505755605Z" level=info msg="StartContainer for \"4d1318708da5e2a4b1ad3316a4f2763d39d0e1832a54721c31612b0afdee5ae3\" returns successfully" May 8 00:47:44.416986 kubelet[2246]: E0508 00:47:44.416924 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:44.434614 kubelet[2246]: I0508 00:47:44.434262 2246 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-c9gns" podStartSLOduration=5.434229596 podStartE2EDuration="5.434229596s" podCreationTimestamp="2025-05-08 00:47:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 00:47:44.43383608 +0000 UTC m=+88.374650571" watchObservedRunningTime="2025-05-08 00:47:44.434229596 +0000 UTC m=+88.375044067" May 8 00:47:45.045000 audit[4463]: AVC avc: denied { write } for pid=4463 comm="tee" name="fd" dev="proc" ino=26913 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 May 8 00:47:45.045000 audit[4463]: SYSCALL arch=c000003e syscall=257 success=yes exit=3 a0=ffffff9c a1=7fff35072a2b a2=241 a3=1b6 items=1 ppid=4432 pid=4463 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.045000 audit: CWD cwd="/etc/service/enabled/bird/log" May 8 00:47:45.045000 audit: PATH item=0 name="/dev/fd/63" inode=26901 dev=00:0c mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:47:45.045000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 May 8 00:47:45.047000 audit[4470]: AVC avc: denied { write } for pid=4470 comm="tee" name="fd" dev="proc" ino=26918 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 May 8 00:47:45.050000 audit[4492]: AVC avc: denied { write } for pid=4492 comm="tee" name="fd" dev="proc" ino=28830 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 May 8 00:47:45.050000 audit[4492]: SYSCALL arch=c000003e syscall=257 success=yes exit=3 a0=ffffff9c a1=7ffe96b2aa2a a2=241 a3=1b6 items=1 ppid=4455 pid=4492 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.050000 audit: CWD cwd="/etc/service/enabled/felix/log" May 8 00:47:45.050000 audit: PATH item=0 name="/dev/fd/63" inode=26919 dev=00:0c mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:47:45.050000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 May 8 00:47:45.047000 audit[4470]: SYSCALL arch=c000003e syscall=257 success=yes exit=3 a0=ffffff9c a1=7ffc953eea1a a2=241 a3=1b6 items=1 ppid=4431 pid=4470 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.047000 audit: CWD cwd="/etc/service/enabled/allocate-tunnel-addrs/log" May 8 00:47:45.047000 audit: PATH item=0 name="/dev/fd/63" inode=26906 dev=00:0c mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:47:45.047000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 May 8 00:47:45.060000 audit[4487]: AVC avc: denied { write } for pid=4487 comm="tee" name="fd" dev="proc" ino=26516 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 May 8 00:47:45.060000 audit[4487]: SYSCALL arch=c000003e syscall=257 success=yes exit=3 a0=ffffff9c a1=7ffdb4047a2a a2=241 a3=1b6 items=1 ppid=4453 pid=4487 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.060000 audit: CWD cwd="/etc/service/enabled/confd/log" May 8 00:47:45.060000 audit: PATH item=0 name="/dev/fd/63" inode=26915 dev=00:0c mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:47:45.060000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 May 8 00:47:45.060000 audit[4506]: AVC avc: denied { write } for pid=4506 comm="tee" name="fd" dev="proc" ino=26518 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 May 8 00:47:45.060000 audit[4506]: SYSCALL arch=c000003e syscall=257 success=yes exit=3 a0=ffffff9c a1=7ffcea21aa1b a2=241 a3=1b6 items=1 ppid=4449 pid=4506 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.060000 audit: CWD cwd="/etc/service/enabled/node-status-reporter/log" May 8 00:47:45.060000 audit: PATH item=0 name="/dev/fd/63" inode=27847 dev=00:0c mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:47:45.060000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 May 8 00:47:45.063000 audit[4472]: AVC avc: denied { write } for pid=4472 comm="tee" name="fd" dev="proc" ino=27850 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 May 8 00:47:45.063000 audit[4472]: SYSCALL arch=c000003e syscall=257 success=yes exit=3 a0=ffffff9c a1=7ffe1b2c8a2a a2=241 a3=1b6 items=1 ppid=4436 pid=4472 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.063000 audit: CWD cwd="/etc/service/enabled/bird6/log" May 8 00:47:45.063000 audit: PATH item=0 name="/dev/fd/63" inode=27835 dev=00:0c mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:47:45.063000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 May 8 00:47:45.104000 audit[4514]: AVC avc: denied { write } for pid=4514 comm="tee" name="fd" dev="proc" ino=26523 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=dir permissive=0 May 8 00:47:45.104000 audit[4514]: SYSCALL arch=c000003e syscall=257 success=yes exit=3 a0=ffffff9c a1=7ffd5c173a2c a2=241 a3=1b6 items=1 ppid=4452 pid=4514 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="tee" exe="/usr/bin/coreutils" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.104000 audit: CWD cwd="/etc/service/enabled/cni/log" May 8 00:47:45.104000 audit: PATH item=0 name="/dev/fd/63" inode=26924 dev=00:0c mode=010600 ouid=0 ogid=0 rdev=00:00 obj=system_u:system_r:kernel_t:s0 nametype=NORMAL cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 8 00:47:45.104000 audit: PROCTITLE proctitle=2F7573722F62696E2F636F72657574696C73002D2D636F72657574696C732D70726F672D73686562616E673D746565002F7573722F62696E2F746565002F6465762F66642F3633 May 8 00:47:45.165915 env[1311]: time="2025-05-08T00:47:45.165847982Z" level=info msg="StopPodSandbox for \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\"" May 8 00:47:45.317958 env[1311]: 2025-05-08 00:47:45.276 [INFO][4535] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:47:45.317958 env[1311]: 2025-05-08 00:47:45.276 [INFO][4535] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" iface="eth0" netns="/var/run/netns/cni-4c4c6460-84b9-f9ac-b7a6-b5cc45c27356" May 8 00:47:45.317958 env[1311]: 2025-05-08 00:47:45.276 [INFO][4535] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" iface="eth0" netns="/var/run/netns/cni-4c4c6460-84b9-f9ac-b7a6-b5cc45c27356" May 8 00:47:45.317958 env[1311]: 2025-05-08 00:47:45.277 [INFO][4535] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" iface="eth0" netns="/var/run/netns/cni-4c4c6460-84b9-f9ac-b7a6-b5cc45c27356" May 8 00:47:45.317958 env[1311]: 2025-05-08 00:47:45.277 [INFO][4535] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:47:45.317958 env[1311]: 2025-05-08 00:47:45.277 [INFO][4535] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:47:45.317958 env[1311]: 2025-05-08 00:47:45.303 [INFO][4562] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" HandleID="k8s-pod-network.cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" Workload="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:47:45.317958 env[1311]: 2025-05-08 00:47:45.303 [INFO][4562] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:47:45.317958 env[1311]: 2025-05-08 00:47:45.303 [INFO][4562] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:47:45.317958 env[1311]: 2025-05-08 00:47:45.309 [WARNING][4562] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" HandleID="k8s-pod-network.cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" Workload="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:47:45.317958 env[1311]: 2025-05-08 00:47:45.309 [INFO][4562] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" HandleID="k8s-pod-network.cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" Workload="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:47:45.317958 env[1311]: 2025-05-08 00:47:45.312 [INFO][4562] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:47:45.317958 env[1311]: 2025-05-08 00:47:45.314 [INFO][4535] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:47:45.320324 systemd[1]: run-netns-cni\x2d4c4c6460\x2d84b9\x2df9ac\x2db7a6\x2db5cc45c27356.mount: Deactivated successfully. May 8 00:47:45.321017 env[1311]: time="2025-05-08T00:47:45.320639226Z" level=info msg="TearDown network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\" successfully" May 8 00:47:45.321017 env[1311]: time="2025-05-08T00:47:45.320682228Z" level=info msg="StopPodSandbox for \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\" returns successfully" May 8 00:47:45.321532 env[1311]: time="2025-05-08T00:47:45.321502283Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fd97795bb-tzgmm,Uid:c5c7ed4b-9783-4fe2-b85a-8452a359cb74,Namespace:calico-apiserver,Attempt:1,}" May 8 00:47:45.407000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.407000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.407000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.407000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.407000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.407000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.407000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.407000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.407000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.407000 audit: BPF prog-id=10 op=LOAD May 8 00:47:45.407000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7fff97b420d0 a2=98 a3=3 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.407000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.407000 audit: BPF prog-id=10 op=UNLOAD May 8 00:47:45.409000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit: BPF prog-id=11 op=LOAD May 8 00:47:45.409000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fff97b41eb0 a2=74 a3=540051 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.409000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.409000 audit: BPF prog-id=11 op=UNLOAD May 8 00:47:45.409000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.409000 audit: BPF prog-id=12 op=LOAD May 8 00:47:45.409000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fff97b41ee0 a2=94 a3=2 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.409000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.409000 audit: BPF prog-id=12 op=UNLOAD May 8 00:47:45.419882 kubelet[2246]: E0508 00:47:45.419490 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:45.546000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.546000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.546000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.546000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.546000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.546000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.546000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.546000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.546000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.546000 audit: BPF prog-id=13 op=LOAD May 8 00:47:45.546000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fff97b41da0 a2=40 a3=1 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.546000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.546000 audit: BPF prog-id=13 op=UNLOAD May 8 00:47:45.546000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.546000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=0 a1=7fff97b41e70 a2=50 a3=7fff97b41f50 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.546000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.556000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.556000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=12 a1=7fff97b41db0 a2=28 a3=0 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.556000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.556000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.556000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7fff97b41de0 a2=28 a3=0 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.556000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.556000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.556000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7fff97b41cf0 a2=28 a3=0 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.556000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.557000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.557000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=12 a1=7fff97b41e00 a2=28 a3=0 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.557000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.557000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.557000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=12 a1=7fff97b41de0 a2=28 a3=0 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.557000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.557925 systemd-networkd[1072]: calidaae6f6c0e9: Link UP May 8 00:47:45.558000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.558000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=12 a1=7fff97b41dd0 a2=28 a3=0 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.558000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.558000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.558000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=12 a1=7fff97b41e00 a2=28 a3=0 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.558000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.558000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.558000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7fff97b41de0 a2=28 a3=0 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.558000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.559000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.559000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7fff97b41e00 a2=28 a3=0 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.559000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.559000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.559000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7fff97b41dd0 a2=28 a3=0 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.559000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.559000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.559000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=12 a1=7fff97b41e40 a2=28 a3=0 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.559000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.560000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.560000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=0 a1=7fff97b41bf0 a2=50 a3=1 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.560652 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready May 8 00:47:45.560704 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): calidaae6f6c0e9: link becomes ready May 8 00:47:45.560675 systemd-networkd[1072]: calidaae6f6c0e9: Gained carrier May 8 00:47:45.560000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.560000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.560000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.560000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.560000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.560000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.560000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.560000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.560000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.560000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.560000 audit: BPF prog-id=14 op=LOAD May 8 00:47:45.560000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fff97b41bf0 a2=94 a3=5 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.560000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.561000 audit: BPF prog-id=14 op=UNLOAD May 8 00:47:45.561000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.561000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=0 a1=7fff97b41ca0 a2=50 a3=1 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.561000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.562000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.562000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=yes exit=0 a0=16 a1=7fff97b41dc0 a2=4 a3=38 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.562000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.562000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.562000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.562000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.562000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.562000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.562000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.562000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.562000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.562000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.562000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.562000 audit[4581]: AVC avc: denied { confidentiality } for pid=4581 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 May 8 00:47:45.562000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=5 a1=7fff97b41e10 a2=94 a3=6 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.562000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.563000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.563000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.563000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.563000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.563000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.563000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.563000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.563000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.563000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.563000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.563000 audit[4581]: AVC avc: denied { confidentiality } for pid=4581 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 May 8 00:47:45.563000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=5 a1=7fff97b415c0 a2=94 a3=83 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.563000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.564000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.564000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.564000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.564000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.564000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.564000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.564000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.564000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.564000 audit[4581]: AVC avc: denied { perfmon } for pid=4581 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.564000 audit[4581]: AVC avc: denied { bpf } for pid=4581 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.564000 audit[4581]: AVC avc: denied { confidentiality } for pid=4581 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 May 8 00:47:45.564000 audit[4581]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=5 a1=7fff97b415c0 a2=94 a3=83 items=0 ppid=4459 pid=4581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.564000 audit: PROCTITLE proctitle=627066746F6F6C006D6170006C697374002D2D6A736F6E May 8 00:47:45.572000 audit[4631]: AVC avc: denied { bpf } for pid=4631 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.572000 audit[4631]: AVC avc: denied { bpf } for pid=4631 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.572000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.572000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.572000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.572000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.572000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.572000 audit[4631]: AVC avc: denied { bpf } for pid=4631 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.572000 audit[4631]: AVC avc: denied { bpf } for pid=4631 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.572000 audit: BPF prog-id=15 op=LOAD May 8 00:47:45.572000 audit[4631]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe7e967a60 a2=98 a3=1999999999999999 items=0 ppid=4459 pid=4631 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.572000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F May 8 00:47:45.576000 audit: BPF prog-id=15 op=UNLOAD May 8 00:47:45.576000 audit[4631]: AVC avc: denied { bpf } for pid=4631 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.576000 audit[4631]: AVC avc: denied { bpf } for pid=4631 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.576000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.576000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.576000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.576000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.576000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.576000 audit[4631]: AVC avc: denied { bpf } for pid=4631 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.479 [INFO][4580] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0 calico-apiserver-fd97795bb- calico-apiserver c5c7ed4b-9783-4fe2-b85a-8452a359cb74 1153 0 2025-05-08 00:46:38 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:fd97795bb projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-fd97795bb-tzgmm eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calidaae6f6c0e9 [] []}} ContainerID="3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-tzgmm" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-" May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.479 [INFO][4580] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-tzgmm" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.512 [INFO][4616] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" HandleID="k8s-pod-network.3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" Workload="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.522 [INFO][4616] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" HandleID="k8s-pod-network.3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" Workload="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000265a90), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-fd97795bb-tzgmm", "timestamp":"2025-05-08 00:47:45.512933452 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.522 [INFO][4616] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.522 [INFO][4616] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.522 [INFO][4616] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.525 [INFO][4616] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" host="localhost" May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.528 [INFO][4616] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.532 [INFO][4616] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.533 [INFO][4616] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.535 [INFO][4616] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.535 [INFO][4616] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" host="localhost" May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.537 [INFO][4616] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.541 [INFO][4616] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" host="localhost" May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.545 [INFO][4616] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" host="localhost" May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.545 [INFO][4616] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" host="localhost" May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.545 [INFO][4616] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:47:45.577249 env[1311]: 2025-05-08 00:47:45.545 [INFO][4616] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" HandleID="k8s-pod-network.3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" Workload="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:47:45.577817 env[1311]: 2025-05-08 00:47:45.548 [INFO][4580] cni-plugin/k8s.go 386: Populated endpoint ContainerID="3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-tzgmm" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0", GenerateName:"calico-apiserver-fd97795bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"c5c7ed4b-9783-4fe2-b85a-8452a359cb74", ResourceVersion:"1153", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"fd97795bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-fd97795bb-tzgmm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calidaae6f6c0e9", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:47:45.577817 env[1311]: 2025-05-08 00:47:45.548 [INFO][4580] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.129/32] ContainerID="3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-tzgmm" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:47:45.577817 env[1311]: 2025-05-08 00:47:45.548 [INFO][4580] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calidaae6f6c0e9 ContainerID="3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-tzgmm" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:47:45.577817 env[1311]: 2025-05-08 00:47:45.561 [INFO][4580] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-tzgmm" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:47:45.577817 env[1311]: 2025-05-08 00:47:45.563 [INFO][4580] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-tzgmm" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0", GenerateName:"calico-apiserver-fd97795bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"c5c7ed4b-9783-4fe2-b85a-8452a359cb74", ResourceVersion:"1153", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"fd97795bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f", Pod:"calico-apiserver-fd97795bb-tzgmm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calidaae6f6c0e9", MAC:"3a:73:0f:08:1b:b2", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:47:45.577817 env[1311]: 2025-05-08 00:47:45.574 [INFO][4580] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-tzgmm" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:47:45.576000 audit[4631]: AVC avc: denied { bpf } for pid=4631 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.576000 audit: BPF prog-id=16 op=LOAD May 8 00:47:45.576000 audit[4631]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe7e967940 a2=74 a3=ffff items=0 ppid=4459 pid=4631 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.576000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F May 8 00:47:45.578000 audit: BPF prog-id=16 op=UNLOAD May 8 00:47:45.578000 audit[4631]: AVC avc: denied { bpf } for pid=4631 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.578000 audit[4631]: AVC avc: denied { bpf } for pid=4631 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.578000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.578000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.578000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.578000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.578000 audit[4631]: AVC avc: denied { perfmon } for pid=4631 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.578000 audit[4631]: AVC avc: denied { bpf } for pid=4631 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.578000 audit[4631]: AVC avc: denied { bpf } for pid=4631 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.578000 audit: BPF prog-id=17 op=LOAD May 8 00:47:45.578000 audit[4631]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffe7e967980 a2=40 a3=7ffe7e967b60 items=0 ppid=4459 pid=4631 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.578000 audit: PROCTITLE proctitle=627066746F6F6C006D617000637265617465002F7379732F66732F6270662F63616C69636F2F63616C69636F5F6661696C736166655F706F7274735F763100747970650068617368006B657900340076616C7565003100656E7472696573003635353335006E616D650063616C69636F5F6661696C736166655F706F7274735F May 8 00:47:45.579000 audit: BPF prog-id=17 op=UNLOAD May 8 00:47:45.591218 env[1311]: time="2025-05-08T00:47:45.589061410Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 8 00:47:45.591218 env[1311]: time="2025-05-08T00:47:45.589127786Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 8 00:47:45.591218 env[1311]: time="2025-05-08T00:47:45.589138527Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 8 00:47:45.591218 env[1311]: time="2025-05-08T00:47:45.589362170Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f pid=4658 runtime=io.containerd.runc.v2 May 8 00:47:45.620139 systemd-resolved[1221]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit: BPF prog-id=18 op=LOAD May 8 00:47:45.633000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffee9ad9ca0 a2=98 a3=ffffffff items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.633000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.633000 audit: BPF prog-id=18 op=UNLOAD May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit: BPF prog-id=19 op=LOAD May 8 00:47:45.633000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffee9ad9ab0 a2=74 a3=540051 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.633000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.633000 audit: BPF prog-id=19 op=UNLOAD May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit: BPF prog-id=20 op=LOAD May 8 00:47:45.633000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7ffee9ad9ae0 a2=94 a3=2 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.633000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.633000 audit: BPF prog-id=20 op=UNLOAD May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=12 a1=7ffee9ad99b0 a2=28 a3=0 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.633000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7ffee9ad99e0 a2=28 a3=0 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.633000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7ffee9ad98f0 a2=28 a3=0 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.633000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=12 a1=7ffee9ad9a00 a2=28 a3=0 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.633000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=12 a1=7ffee9ad99e0 a2=28 a3=0 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.633000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.633000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.633000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=12 a1=7ffee9ad99d0 a2=28 a3=0 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.633000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=12 a1=7ffee9ad9a00 a2=28 a3=0 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.639000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7ffee9ad99e0 a2=28 a3=0 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.639000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7ffee9ad9a00 a2=28 a3=0 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.639000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7ffee9ad99d0 a2=28 a3=0 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.639000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=12 a1=7ffee9ad9a40 a2=28 a3=0 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.639000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit: BPF prog-id=21 op=LOAD May 8 00:47:45.639000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffee9ad98b0 a2=40 a3=0 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.639000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.639000 audit: BPF prog-id=21 op=UNLOAD May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=0 a1=7ffee9ad98a0 a2=50 a3=2800 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.639000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=0 a1=7ffee9ad98a0 a2=50 a3=2800 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.639000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit: BPF prog-id=22 op=LOAD May 8 00:47:45.639000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffee9ad90c0 a2=94 a3=2 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.639000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.639000 audit: BPF prog-id=22 op=UNLOAD May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { perfmon } for pid=4701 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.638950 systemd-networkd[1072]: vxlan.calico: Link UP May 8 00:47:45.638955 systemd-networkd[1072]: vxlan.calico: Gained carrier May 8 00:47:45.639000 audit[4701]: AVC avc: denied { bpf } for pid=4701 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.639000 audit: BPF prog-id=23 op=LOAD May 8 00:47:45.639000 audit[4701]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7ffee9ad91c0 a2=94 a3=30 items=0 ppid=4459 pid=4701 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.639000 audit: PROCTITLE proctitle=627066746F6F6C0070726F67006C6F6164002F7573722F6C69622F63616C69636F2F6270662F66696C7465722E6F002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41007479706500786470 May 8 00:47:45.651000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.651000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.651000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.651000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.651000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.651000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.651000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.651000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.651000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.651000 audit: BPF prog-id=24 op=LOAD May 8 00:47:45.651000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=3 a0=5 a1=7fff5e5477e0 a2=98 a3=0 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.651000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.651000 audit: BPF prog-id=24 op=UNLOAD May 8 00:47:45.652000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit: BPF prog-id=25 op=LOAD May 8 00:47:45.652000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fff5e5475c0 a2=74 a3=540051 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.652000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.652000 audit: BPF prog-id=25 op=UNLOAD May 8 00:47:45.652000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.652000 audit: BPF prog-id=26 op=LOAD May 8 00:47:45.652000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fff5e5475f0 a2=94 a3=2 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.652000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.652000 audit: BPF prog-id=26 op=UNLOAD May 8 00:47:45.664910 env[1311]: time="2025-05-08T00:47:45.664857499Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fd97795bb-tzgmm,Uid:c5c7ed4b-9783-4fe2-b85a-8452a359cb74,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f\"" May 8 00:47:45.667168 env[1311]: time="2025-05-08T00:47:45.667141460Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" May 8 00:47:45.770000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.770000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.770000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.770000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.770000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.770000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.770000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.770000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.770000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.770000 audit: BPF prog-id=27 op=LOAD May 8 00:47:45.770000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=5 a1=7fff5e5474b0 a2=40 a3=1 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.770000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.770000 audit: BPF prog-id=27 op=UNLOAD May 8 00:47:45.770000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.770000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=0 a1=7fff5e547580 a2=50 a3=7fff5e547660 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.770000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=12 a1=7fff5e5474c0 a2=28 a3=0 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7fff5e5474f0 a2=28 a3=0 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7fff5e547400 a2=28 a3=0 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=12 a1=7fff5e547510 a2=28 a3=0 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=12 a1=7fff5e5474f0 a2=28 a3=0 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=12 a1=7fff5e5474e0 a2=28 a3=0 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=12 a1=7fff5e547510 a2=28 a3=0 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7fff5e5474f0 a2=28 a3=0 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7fff5e547510 a2=28 a3=0 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=12 a1=7fff5e5474e0 a2=28 a3=0 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=4 a0=12 a1=7fff5e547550 a2=28 a3=0 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=0 a1=7fff5e547300 a2=50 a3=1 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit: BPF prog-id=28 op=LOAD May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=6 a0=5 a1=7fff5e547300 a2=94 a3=5 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit: BPF prog-id=28 op=UNLOAD May 8 00:47:45.778000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=5 a0=0 a1=7fff5e5473b0 a2=50 a3=1 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=0 a0=16 a1=7fff5e5474d0 a2=4 a3=38 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.778000 audit[4712]: AVC avc: denied { confidentiality } for pid=4712 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 May 8 00:47:45.778000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=5 a1=7fff5e547520 a2=94 a3=6 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.778000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { confidentiality } for pid=4712 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 May 8 00:47:45.779000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=5 a1=7fff5e546cd0 a2=94 a3=83 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.779000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { perfmon } for pid=4712 comm="bpftool" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { confidentiality } for pid=4712 comm="bpftool" lockdown_reason="use of bpf to read kernel RAM" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=0 May 8 00:47:45.779000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=no exit=-22 a0=5 a1=7fff5e546cd0 a2=94 a3=83 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.779000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.779000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.779000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=0 a0=f a1=7fff5e548710 a2=10 a3=208 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.779000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.780000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.780000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=0 a0=f a1=7fff5e5485b0 a2=10 a3=3 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.780000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.780000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.780000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=0 a0=f a1=7fff5e548550 a2=10 a3=3 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.780000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.780000 audit[4712]: AVC avc: denied { bpf } for pid=4712 comm="bpftool" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=0 May 8 00:47:45.780000 audit[4712]: SYSCALL arch=c000003e syscall=321 success=yes exit=0 a0=f a1=7fff5e548550 a2=10 a3=7 items=0 ppid=4459 pid=4712 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="bpftool" exe="/usr/bin/bpftool" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.780000 audit: PROCTITLE proctitle=627066746F6F6C002D2D6A736F6E002D2D7072657474790070726F670073686F770070696E6E6564002F7379732F66732F6270662F63616C69636F2F7864702F70726566696C7465725F76315F63616C69636F5F746D705F41 May 8 00:47:45.788000 audit: BPF prog-id=23 op=UNLOAD May 8 00:47:45.837000 audit[4741]: NETFILTER_CFG table=mangle:97 family=2 entries=16 op=nft_register_chain pid=4741 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" May 8 00:47:45.837000 audit[4741]: SYSCALL arch=c000003e syscall=46 success=yes exit=6868 a0=3 a1=7fff8a3be9d0 a2=0 a3=7fff8a3be9bc items=0 ppid=4459 pid=4741 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.837000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 May 8 00:47:45.850000 audit[4742]: NETFILTER_CFG table=nat:98 family=2 entries=15 op=nft_register_chain pid=4742 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" May 8 00:47:45.850000 audit[4742]: SYSCALL arch=c000003e syscall=46 success=yes exit=5084 a0=3 a1=7ffd04c27330 a2=0 a3=7ffd04c2731c items=0 ppid=4459 pid=4742 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.850000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 May 8 00:47:45.850000 audit[4745]: NETFILTER_CFG table=filter:99 family=2 entries=75 op=nft_register_chain pid=4745 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" May 8 00:47:45.850000 audit[4745]: SYSCALL arch=c000003e syscall=46 success=yes exit=40748 a0=3 a1=7ffc425c0d60 a2=0 a3=556d0f389000 items=0 ppid=4459 pid=4745 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.850000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 May 8 00:47:45.854000 audit[4740]: NETFILTER_CFG table=raw:100 family=2 entries=21 op=nft_register_chain pid=4740 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" May 8 00:47:45.854000 audit[4740]: SYSCALL arch=c000003e syscall=46 success=yes exit=8452 a0=3 a1=7ffc89dcf640 a2=0 a3=7ffc89dcf62c items=0 ppid=4459 pid=4740 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:45.854000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 May 8 00:47:46.320438 systemd[1]: run-containerd-runc-k8s.io-4d1318708da5e2a4b1ad3316a4f2763d39d0e1832a54721c31612b0afdee5ae3-runc.HAZN3H.mount: Deactivated successfully. May 8 00:47:46.902347 systemd-networkd[1072]: calidaae6f6c0e9: Gained IPv6LL May 8 00:47:47.030780 systemd-networkd[1072]: vxlan.calico: Gained IPv6LL May 8 00:47:47.519088 systemd[1]: Started sshd@18-10.0.0.82:22-10.0.0.1:33262.service. May 8 00:47:47.518000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.82:22-10.0.0.1:33262 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:47.523435 kernel: kauditd_printk_skb: 516 callbacks suppressed May 8 00:47:47.523504 kernel: audit: type=1130 audit(1746665267.518:490): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.82:22-10.0.0.1:33262 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:47.574000 audit[4754]: USER_ACCT pid=4754 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:47.575642 sshd[4754]: Accepted publickey for core from 10.0.0.1 port 33262 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:47.584766 sshd[4754]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:47.576000 audit[4754]: CRED_ACQ pid=4754 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:47.589611 systemd-logind[1291]: New session 19 of user core. May 8 00:47:47.590133 systemd[1]: Started session-19.scope. May 8 00:47:47.607394 kernel: audit: type=1101 audit(1746665267.574:491): pid=4754 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:47.607541 kernel: audit: type=1103 audit(1746665267.576:492): pid=4754 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:47.607590 kernel: audit: type=1006 audit(1746665267.576:493): pid=4754 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=19 res=1 May 8 00:47:47.576000 audit[4754]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7fffe7b30650 a2=3 a3=0 items=0 ppid=1 pid=4754 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=19 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:47.614845 kernel: audit: type=1300 audit(1746665267.576:493): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7fffe7b30650 a2=3 a3=0 items=0 ppid=1 pid=4754 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=19 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:47.614893 kernel: audit: type=1327 audit(1746665267.576:493): proctitle=737368643A20636F7265205B707269765D May 8 00:47:47.576000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:47.594000 audit[4754]: USER_START pid=4754 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:47.622976 kernel: audit: type=1105 audit(1746665267.594:494): pid=4754 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:47.623059 kernel: audit: type=1103 audit(1746665267.595:495): pid=4757 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:47.595000 audit[4757]: CRED_ACQ pid=4757 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:47.714401 sshd[4754]: pam_unix(sshd:session): session closed for user core May 8 00:47:47.715000 audit[4754]: USER_END pid=4754 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:47.718177 systemd[1]: sshd@18-10.0.0.82:22-10.0.0.1:33262.service: Deactivated successfully. May 8 00:47:47.719215 systemd[1]: session-19.scope: Deactivated successfully. May 8 00:47:47.720331 systemd-logind[1291]: Session 19 logged out. Waiting for processes to exit. May 8 00:47:47.715000 audit[4754]: CRED_DISP pid=4754 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:47.721579 systemd-logind[1291]: Removed session 19. May 8 00:47:47.726138 kernel: audit: type=1106 audit(1746665267.715:496): pid=4754 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:47.726194 kernel: audit: type=1104 audit(1746665267.715:497): pid=4754 uid=0 auid=500 ses=19 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:47.717000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@18-10.0.0.82:22-10.0.0.1:33262 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:48.168660 env[1311]: time="2025-05-08T00:47:48.168607929Z" level=info msg="StopPodSandbox for \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\"" May 8 00:47:48.350627 env[1311]: 2025-05-08 00:47:48.317 [INFO][4785] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:47:48.350627 env[1311]: 2025-05-08 00:47:48.317 [INFO][4785] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" iface="eth0" netns="/var/run/netns/cni-422cfa05-a00d-25b2-dd74-77eff573347f" May 8 00:47:48.350627 env[1311]: 2025-05-08 00:47:48.317 [INFO][4785] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" iface="eth0" netns="/var/run/netns/cni-422cfa05-a00d-25b2-dd74-77eff573347f" May 8 00:47:48.350627 env[1311]: 2025-05-08 00:47:48.318 [INFO][4785] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" iface="eth0" netns="/var/run/netns/cni-422cfa05-a00d-25b2-dd74-77eff573347f" May 8 00:47:48.350627 env[1311]: 2025-05-08 00:47:48.318 [INFO][4785] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:47:48.350627 env[1311]: 2025-05-08 00:47:48.318 [INFO][4785] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:47:48.350627 env[1311]: 2025-05-08 00:47:48.339 [INFO][4794] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" HandleID="k8s-pod-network.85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" Workload="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:47:48.350627 env[1311]: 2025-05-08 00:47:48.339 [INFO][4794] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:47:48.350627 env[1311]: 2025-05-08 00:47:48.339 [INFO][4794] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:47:48.350627 env[1311]: 2025-05-08 00:47:48.346 [WARNING][4794] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" HandleID="k8s-pod-network.85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" Workload="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:47:48.350627 env[1311]: 2025-05-08 00:47:48.346 [INFO][4794] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" HandleID="k8s-pod-network.85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" Workload="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:47:48.350627 env[1311]: 2025-05-08 00:47:48.347 [INFO][4794] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:47:48.350627 env[1311]: 2025-05-08 00:47:48.348 [INFO][4785] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:47:48.351161 env[1311]: time="2025-05-08T00:47:48.350756350Z" level=info msg="TearDown network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\" successfully" May 8 00:47:48.351161 env[1311]: time="2025-05-08T00:47:48.350798460Z" level=info msg="StopPodSandbox for \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\" returns successfully" May 8 00:47:48.351287 kubelet[2246]: E0508 00:47:48.351249 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:48.352082 env[1311]: time="2025-05-08T00:47:48.352027449Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-c2qjt,Uid:9c26b328-0da1-4b8c-b690-cd29958643ac,Namespace:kube-system,Attempt:1,}" May 8 00:47:48.353434 systemd[1]: run-netns-cni\x2d422cfa05\x2da00d\x2d25b2\x2ddd74\x2d77eff573347f.mount: Deactivated successfully. May 8 00:47:48.638267 systemd-networkd[1072]: cali62884932f3c: Link UP May 8 00:47:48.641776 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready May 8 00:47:48.641841 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cali62884932f3c: link becomes ready May 8 00:47:48.641180 systemd-networkd[1072]: cali62884932f3c: Gained carrier May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.572 [INFO][4801] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0 coredns-7db6d8ff4d- kube-system 9c26b328-0da1-4b8c-b690-cd29958643ac 1170 0 2025-05-08 00:46:30 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7db6d8ff4d projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7db6d8ff4d-c2qjt eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali62884932f3c [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" Namespace="kube-system" Pod="coredns-7db6d8ff4d-c2qjt" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--c2qjt-" May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.572 [INFO][4801] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" Namespace="kube-system" Pod="coredns-7db6d8ff4d-c2qjt" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.604 [INFO][4816] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" HandleID="k8s-pod-network.6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" Workload="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.611 [INFO][4816] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" HandleID="k8s-pod-network.6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" Workload="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0004e67a0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7db6d8ff4d-c2qjt", "timestamp":"2025-05-08 00:47:48.604203569 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.612 [INFO][4816] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.612 [INFO][4816] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.612 [INFO][4816] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.613 [INFO][4816] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" host="localhost" May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.616 [INFO][4816] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.620 [INFO][4816] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.621 [INFO][4816] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.623 [INFO][4816] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.623 [INFO][4816] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" host="localhost" May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.624 [INFO][4816] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388 May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.628 [INFO][4816] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" host="localhost" May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.633 [INFO][4816] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" host="localhost" May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.633 [INFO][4816] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" host="localhost" May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.633 [INFO][4816] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:47:48.654640 env[1311]: 2025-05-08 00:47:48.633 [INFO][4816] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" HandleID="k8s-pod-network.6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" Workload="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:47:48.655400 env[1311]: 2025-05-08 00:47:48.636 [INFO][4801] cni-plugin/k8s.go 386: Populated endpoint ContainerID="6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" Namespace="kube-system" Pod="coredns-7db6d8ff4d-c2qjt" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"9c26b328-0da1-4b8c-b690-cd29958643ac", ResourceVersion:"1170", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7db6d8ff4d-c2qjt", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali62884932f3c", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:47:48.655400 env[1311]: 2025-05-08 00:47:48.636 [INFO][4801] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.130/32] ContainerID="6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" Namespace="kube-system" Pod="coredns-7db6d8ff4d-c2qjt" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:47:48.655400 env[1311]: 2025-05-08 00:47:48.636 [INFO][4801] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali62884932f3c ContainerID="6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" Namespace="kube-system" Pod="coredns-7db6d8ff4d-c2qjt" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:47:48.655400 env[1311]: 2025-05-08 00:47:48.641 [INFO][4801] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" Namespace="kube-system" Pod="coredns-7db6d8ff4d-c2qjt" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:47:48.655400 env[1311]: 2025-05-08 00:47:48.642 [INFO][4801] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" Namespace="kube-system" Pod="coredns-7db6d8ff4d-c2qjt" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"9c26b328-0da1-4b8c-b690-cd29958643ac", ResourceVersion:"1170", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388", Pod:"coredns-7db6d8ff4d-c2qjt", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali62884932f3c", MAC:"c2:98:b0:ed:20:3b", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:47:48.655400 env[1311]: 2025-05-08 00:47:48.652 [INFO][4801] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388" Namespace="kube-system" Pod="coredns-7db6d8ff4d-c2qjt" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:47:48.666000 audit[4843]: NETFILTER_CFG table=filter:101 family=2 entries=38 op=nft_register_chain pid=4843 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" May 8 00:47:48.666000 audit[4843]: SYSCALL arch=c000003e syscall=46 success=yes exit=20336 a0=3 a1=7ffc385a4130 a2=0 a3=7ffc385a411c items=0 ppid=4459 pid=4843 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:48.666000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 May 8 00:47:48.668402 env[1311]: time="2025-05-08T00:47:48.668308005Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 8 00:47:48.668402 env[1311]: time="2025-05-08T00:47:48.668371436Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 8 00:47:48.668402 env[1311]: time="2025-05-08T00:47:48.668386705Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 8 00:47:48.668631 env[1311]: time="2025-05-08T00:47:48.668596793Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388 pid=4844 runtime=io.containerd.runc.v2 May 8 00:47:48.703738 systemd-resolved[1221]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 8 00:47:48.727790 env[1311]: time="2025-05-08T00:47:48.727730992Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-c2qjt,Uid:9c26b328-0da1-4b8c-b690-cd29958643ac,Namespace:kube-system,Attempt:1,} returns sandbox id \"6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388\"" May 8 00:47:48.728406 kubelet[2246]: E0508 00:47:48.728377 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:48.730641 env[1311]: time="2025-05-08T00:47:48.730597182Z" level=info msg="CreateContainer within sandbox \"6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 8 00:47:48.753706 env[1311]: time="2025-05-08T00:47:48.753631636Z" level=info msg="CreateContainer within sandbox \"6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"f3c3c3d668d442f7137bc6447f8451d199519800e56d8a2674069fcb85c726c1\"" May 8 00:47:48.754780 env[1311]: time="2025-05-08T00:47:48.754723454Z" level=info msg="StartContainer for \"f3c3c3d668d442f7137bc6447f8451d199519800e56d8a2674069fcb85c726c1\"" May 8 00:47:48.816990 env[1311]: time="2025-05-08T00:47:48.816916378Z" level=info msg="StartContainer for \"f3c3c3d668d442f7137bc6447f8451d199519800e56d8a2674069fcb85c726c1\" returns successfully" May 8 00:47:49.166514 kubelet[2246]: E0508 00:47:49.166145 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:49.166731 env[1311]: time="2025-05-08T00:47:49.166674643Z" level=info msg="StopPodSandbox for \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\"" May 8 00:47:49.431062 kubelet[2246]: E0508 00:47:49.430734 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:49.438075 env[1311]: 2025-05-08 00:47:49.394 [INFO][4932] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:47:49.438075 env[1311]: 2025-05-08 00:47:49.395 [INFO][4932] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" iface="eth0" netns="/var/run/netns/cni-ccddadbb-e9c3-7ed0-b17d-fcb88006574b" May 8 00:47:49.438075 env[1311]: 2025-05-08 00:47:49.395 [INFO][4932] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" iface="eth0" netns="/var/run/netns/cni-ccddadbb-e9c3-7ed0-b17d-fcb88006574b" May 8 00:47:49.438075 env[1311]: 2025-05-08 00:47:49.395 [INFO][4932] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" iface="eth0" netns="/var/run/netns/cni-ccddadbb-e9c3-7ed0-b17d-fcb88006574b" May 8 00:47:49.438075 env[1311]: 2025-05-08 00:47:49.396 [INFO][4932] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:47:49.438075 env[1311]: 2025-05-08 00:47:49.396 [INFO][4932] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:47:49.438075 env[1311]: 2025-05-08 00:47:49.422 [INFO][4941] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" HandleID="k8s-pod-network.df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" Workload="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:47:49.438075 env[1311]: 2025-05-08 00:47:49.422 [INFO][4941] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:47:49.438075 env[1311]: 2025-05-08 00:47:49.422 [INFO][4941] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:47:49.438075 env[1311]: 2025-05-08 00:47:49.430 [WARNING][4941] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" HandleID="k8s-pod-network.df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" Workload="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:47:49.438075 env[1311]: 2025-05-08 00:47:49.430 [INFO][4941] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" HandleID="k8s-pod-network.df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" Workload="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:47:49.438075 env[1311]: 2025-05-08 00:47:49.434 [INFO][4941] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:47:49.438075 env[1311]: 2025-05-08 00:47:49.436 [INFO][4932] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:47:49.442730 env[1311]: time="2025-05-08T00:47:49.438306983Z" level=info msg="TearDown network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\" successfully" May 8 00:47:49.442730 env[1311]: time="2025-05-08T00:47:49.438349704Z" level=info msg="StopPodSandbox for \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\" returns successfully" May 8 00:47:49.442730 env[1311]: time="2025-05-08T00:47:49.439047516Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-8xjk8,Uid:435e64db-0f2e-4f05-8fcf-eefdac14188d,Namespace:kube-system,Attempt:1,}" May 8 00:47:49.440711 systemd[1]: run-netns-cni\x2dccddadbb\x2de9c3\x2d7ed0\x2db17d\x2dfcb88006574b.mount: Deactivated successfully. May 8 00:47:49.442997 kubelet[2246]: E0508 00:47:49.438643 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:49.525863 kubelet[2246]: I0508 00:47:49.523372 2246 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-c2qjt" podStartSLOduration=79.523338127 podStartE2EDuration="1m19.523338127s" podCreationTimestamp="2025-05-08 00:46:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 00:47:49.522395872 +0000 UTC m=+93.463210343" watchObservedRunningTime="2025-05-08 00:47:49.523338127 +0000 UTC m=+93.464152598" May 8 00:47:49.581590 env[1311]: time="2025-05-08T00:47:49.581487739Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/apiserver:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:49.587503 env[1311]: time="2025-05-08T00:47:49.587412182Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:49.588984 env[1311]: time="2025-05-08T00:47:49.588917344Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/apiserver:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:49.592332 env[1311]: time="2025-05-08T00:47:49.592287918Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:49.592554 env[1311]: time="2025-05-08T00:47:49.592523094Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\"" May 8 00:47:49.596791 env[1311]: time="2025-05-08T00:47:49.596748868Z" level=info msg="CreateContainer within sandbox \"3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 8 00:47:49.602000 audit[4962]: NETFILTER_CFG table=filter:102 family=2 entries=16 op=nft_register_rule pid=4962 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:47:49.602000 audit[4962]: SYSCALL arch=c000003e syscall=46 success=yes exit=5908 a0=3 a1=7ffc55797e70 a2=0 a3=7ffc55797e5c items=0 ppid=2433 pid=4962 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:49.602000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:47:49.609000 audit[4962]: NETFILTER_CFG table=nat:103 family=2 entries=14 op=nft_register_rule pid=4962 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:47:49.609000 audit[4962]: SYSCALL arch=c000003e syscall=46 success=yes exit=3468 a0=3 a1=7ffc55797e70 a2=0 a3=0 items=0 ppid=2433 pid=4962 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:49.609000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:47:49.622000 audit[4966]: NETFILTER_CFG table=filter:104 family=2 entries=13 op=nft_register_rule pid=4966 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:47:49.622000 audit[4966]: SYSCALL arch=c000003e syscall=46 success=yes exit=3676 a0=3 a1=7ffc60a36260 a2=0 a3=7ffc60a3624c items=0 ppid=2433 pid=4966 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:49.622000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:47:49.626000 audit[4966]: NETFILTER_CFG table=nat:105 family=2 entries=35 op=nft_register_chain pid=4966 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:47:49.626000 audit[4966]: SYSCALL arch=c000003e syscall=46 success=yes exit=14196 a0=3 a1=7ffc60a36260 a2=0 a3=7ffc60a3624c items=0 ppid=2433 pid=4966 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:49.626000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:47:49.653406 env[1311]: time="2025-05-08T00:47:49.653341051Z" level=info msg="CreateContainer within sandbox \"3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"17e8524f08ea58d8b5178980bf187ffdc08d972eae9aa8592d4bbcd3965e0871\"" May 8 00:47:49.654591 env[1311]: time="2025-05-08T00:47:49.654563566Z" level=info msg="StartContainer for \"17e8524f08ea58d8b5178980bf187ffdc08d972eae9aa8592d4bbcd3965e0871\"" May 8 00:47:49.722975 systemd-networkd[1072]: calica037e43738: Link UP May 8 00:47:49.726283 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready May 8 00:47:49.726426 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): calica037e43738: link becomes ready May 8 00:47:49.726612 systemd-networkd[1072]: calica037e43738: Gained carrier May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.620 [INFO][4948] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0 coredns-7db6d8ff4d- kube-system 435e64db-0f2e-4f05-8fcf-eefdac14188d 1185 0 2025-05-08 00:46:30 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7db6d8ff4d projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7db6d8ff4d-8xjk8 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calica037e43738 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" Namespace="kube-system" Pod="coredns-7db6d8ff4d-8xjk8" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--8xjk8-" May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.620 [INFO][4948] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" Namespace="kube-system" Pod="coredns-7db6d8ff4d-8xjk8" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.674 [INFO][4968] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" HandleID="k8s-pod-network.7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" Workload="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.685 [INFO][4968] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" HandleID="k8s-pod-network.7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" Workload="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003e3950), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7db6d8ff4d-8xjk8", "timestamp":"2025-05-08 00:47:49.674788438 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.685 [INFO][4968] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.685 [INFO][4968] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.685 [INFO][4968] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.687 [INFO][4968] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" host="localhost" May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.691 [INFO][4968] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.695 [INFO][4968] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.697 [INFO][4968] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.700 [INFO][4968] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.700 [INFO][4968] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" host="localhost" May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.702 [INFO][4968] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98 May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.708 [INFO][4968] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" host="localhost" May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.717 [INFO][4968] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" host="localhost" May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.717 [INFO][4968] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" host="localhost" May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.717 [INFO][4968] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:47:49.745133 env[1311]: 2025-05-08 00:47:49.718 [INFO][4968] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" HandleID="k8s-pod-network.7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" Workload="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:47:49.745985 env[1311]: 2025-05-08 00:47:49.720 [INFO][4948] cni-plugin/k8s.go 386: Populated endpoint ContainerID="7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" Namespace="kube-system" Pod="coredns-7db6d8ff4d-8xjk8" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"435e64db-0f2e-4f05-8fcf-eefdac14188d", ResourceVersion:"1185", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7db6d8ff4d-8xjk8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calica037e43738", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:47:49.745985 env[1311]: 2025-05-08 00:47:49.721 [INFO][4948] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.131/32] ContainerID="7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" Namespace="kube-system" Pod="coredns-7db6d8ff4d-8xjk8" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:47:49.745985 env[1311]: 2025-05-08 00:47:49.721 [INFO][4948] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calica037e43738 ContainerID="7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" Namespace="kube-system" Pod="coredns-7db6d8ff4d-8xjk8" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:47:49.745985 env[1311]: 2025-05-08 00:47:49.727 [INFO][4948] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" Namespace="kube-system" Pod="coredns-7db6d8ff4d-8xjk8" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:47:49.745985 env[1311]: 2025-05-08 00:47:49.728 [INFO][4948] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" Namespace="kube-system" Pod="coredns-7db6d8ff4d-8xjk8" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"435e64db-0f2e-4f05-8fcf-eefdac14188d", ResourceVersion:"1185", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98", Pod:"coredns-7db6d8ff4d-8xjk8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calica037e43738", MAC:"32:72:e9:d9:4c:3b", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:47:49.745985 env[1311]: 2025-05-08 00:47:49.740 [INFO][4948] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98" Namespace="kube-system" Pod="coredns-7db6d8ff4d-8xjk8" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:47:49.748546 env[1311]: time="2025-05-08T00:47:49.748486173Z" level=info msg="StartContainer for \"17e8524f08ea58d8b5178980bf187ffdc08d972eae9aa8592d4bbcd3965e0871\" returns successfully" May 8 00:47:49.757000 audit[5023]: NETFILTER_CFG table=filter:106 family=2 entries=34 op=nft_register_chain pid=5023 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" May 8 00:47:49.757000 audit[5023]: SYSCALL arch=c000003e syscall=46 success=yes exit=18220 a0=3 a1=7ffe9ba33d60 a2=0 a3=7ffe9ba33d4c items=0 ppid=4459 pid=5023 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:49.757000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 May 8 00:47:49.762169 env[1311]: time="2025-05-08T00:47:49.762055587Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 8 00:47:49.762283 env[1311]: time="2025-05-08T00:47:49.762172749Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 8 00:47:49.762283 env[1311]: time="2025-05-08T00:47:49.762216572Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 8 00:47:49.762947 env[1311]: time="2025-05-08T00:47:49.762651917Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98 pid=5032 runtime=io.containerd.runc.v2 May 8 00:47:49.787038 systemd-resolved[1221]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 8 00:47:49.812664 env[1311]: time="2025-05-08T00:47:49.812606085Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-8xjk8,Uid:435e64db-0f2e-4f05-8fcf-eefdac14188d,Namespace:kube-system,Attempt:1,} returns sandbox id \"7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98\"" May 8 00:47:49.814678 kubelet[2246]: E0508 00:47:49.813413 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:49.815568 env[1311]: time="2025-05-08T00:47:49.815527158Z" level=info msg="CreateContainer within sandbox \"7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 8 00:47:49.926621 env[1311]: time="2025-05-08T00:47:49.926543185Z" level=info msg="CreateContainer within sandbox \"7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"86d3c8b555b39c5f911a4cb141ddad27d008df7dd74b4fdc25965483d11b990f\"" May 8 00:47:49.927497 env[1311]: time="2025-05-08T00:47:49.927464480Z" level=info msg="StartContainer for \"86d3c8b555b39c5f911a4cb141ddad27d008df7dd74b4fdc25965483d11b990f\"" May 8 00:47:49.987800 systemd-networkd[1072]: cali62884932f3c: Gained IPv6LL May 8 00:47:50.032363 env[1311]: time="2025-05-08T00:47:50.032305308Z" level=info msg="StartContainer for \"86d3c8b555b39c5f911a4cb141ddad27d008df7dd74b4fdc25965483d11b990f\" returns successfully" May 8 00:47:50.166649 env[1311]: time="2025-05-08T00:47:50.166495554Z" level=info msg="StopPodSandbox for \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\"" May 8 00:47:50.271418 env[1311]: 2025-05-08 00:47:50.222 [INFO][5122] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:47:50.271418 env[1311]: 2025-05-08 00:47:50.222 [INFO][5122] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" iface="eth0" netns="/var/run/netns/cni-9858f845-64e6-523d-2bd7-d69f28c1f756" May 8 00:47:50.271418 env[1311]: 2025-05-08 00:47:50.223 [INFO][5122] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" iface="eth0" netns="/var/run/netns/cni-9858f845-64e6-523d-2bd7-d69f28c1f756" May 8 00:47:50.271418 env[1311]: 2025-05-08 00:47:50.223 [INFO][5122] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" iface="eth0" netns="/var/run/netns/cni-9858f845-64e6-523d-2bd7-d69f28c1f756" May 8 00:47:50.271418 env[1311]: 2025-05-08 00:47:50.223 [INFO][5122] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:47:50.271418 env[1311]: 2025-05-08 00:47:50.223 [INFO][5122] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:47:50.271418 env[1311]: 2025-05-08 00:47:50.257 [INFO][5130] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" HandleID="k8s-pod-network.95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" Workload="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:47:50.271418 env[1311]: 2025-05-08 00:47:50.257 [INFO][5130] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:47:50.271418 env[1311]: 2025-05-08 00:47:50.257 [INFO][5130] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:47:50.271418 env[1311]: 2025-05-08 00:47:50.264 [WARNING][5130] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" HandleID="k8s-pod-network.95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" Workload="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:47:50.271418 env[1311]: 2025-05-08 00:47:50.264 [INFO][5130] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" HandleID="k8s-pod-network.95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" Workload="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:47:50.271418 env[1311]: 2025-05-08 00:47:50.267 [INFO][5130] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:47:50.271418 env[1311]: 2025-05-08 00:47:50.269 [INFO][5122] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:47:50.272036 env[1311]: time="2025-05-08T00:47:50.271624561Z" level=info msg="TearDown network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\" successfully" May 8 00:47:50.272036 env[1311]: time="2025-05-08T00:47:50.271670047Z" level=info msg="StopPodSandbox for \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\" returns successfully" May 8 00:47:50.272433 env[1311]: time="2025-05-08T00:47:50.272408015Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fd97795bb-zmwqb,Uid:ca6c7950-e2c3-4481-994b-8bd117426f80,Namespace:calico-apiserver,Attempt:1,}" May 8 00:47:50.356421 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3765666253.mount: Deactivated successfully. May 8 00:47:50.356593 systemd[1]: run-netns-cni\x2d9858f845\x2d64e6\x2d523d\x2d2bd7\x2dd69f28c1f756.mount: Deactivated successfully. May 8 00:47:50.412588 systemd-networkd[1072]: calidaa7669be5f: Link UP May 8 00:47:50.414283 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): calidaa7669be5f: link becomes ready May 8 00:47:50.414446 systemd-networkd[1072]: calidaa7669be5f: Gained carrier May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.320 [INFO][5138] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0 calico-apiserver-fd97795bb- calico-apiserver ca6c7950-e2c3-4481-994b-8bd117426f80 1211 0 2025-05-08 00:46:38 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:fd97795bb projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-fd97795bb-zmwqb eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calidaa7669be5f [] []}} ContainerID="ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-zmwqb" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-" May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.320 [INFO][5138] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-zmwqb" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.355 [INFO][5152] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" HandleID="k8s-pod-network.ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" Workload="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.366 [INFO][5152] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" HandleID="k8s-pod-network.ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" Workload="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003ac5a0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-fd97795bb-zmwqb", "timestamp":"2025-05-08 00:47:50.355360481 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.366 [INFO][5152] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.366 [INFO][5152] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.366 [INFO][5152] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.369 [INFO][5152] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" host="localhost" May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.373 [INFO][5152] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.382 [INFO][5152] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.385 [INFO][5152] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.387 [INFO][5152] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.388 [INFO][5152] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" host="localhost" May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.389 [INFO][5152] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4 May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.393 [INFO][5152] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" host="localhost" May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.405 [INFO][5152] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" host="localhost" May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.405 [INFO][5152] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" host="localhost" May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.405 [INFO][5152] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:47:50.430086 env[1311]: 2025-05-08 00:47:50.405 [INFO][5152] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" HandleID="k8s-pod-network.ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" Workload="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:47:50.431004 env[1311]: 2025-05-08 00:47:50.408 [INFO][5138] cni-plugin/k8s.go 386: Populated endpoint ContainerID="ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-zmwqb" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0", GenerateName:"calico-apiserver-fd97795bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"ca6c7950-e2c3-4481-994b-8bd117426f80", ResourceVersion:"1211", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"fd97795bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-fd97795bb-zmwqb", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calidaa7669be5f", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:47:50.431004 env[1311]: 2025-05-08 00:47:50.408 [INFO][5138] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.132/32] ContainerID="ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-zmwqb" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:47:50.431004 env[1311]: 2025-05-08 00:47:50.408 [INFO][5138] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calidaa7669be5f ContainerID="ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-zmwqb" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:47:50.431004 env[1311]: 2025-05-08 00:47:50.414 [INFO][5138] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-zmwqb" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:47:50.431004 env[1311]: 2025-05-08 00:47:50.414 [INFO][5138] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-zmwqb" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0", GenerateName:"calico-apiserver-fd97795bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"ca6c7950-e2c3-4481-994b-8bd117426f80", ResourceVersion:"1211", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"fd97795bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4", Pod:"calico-apiserver-fd97795bb-zmwqb", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calidaa7669be5f", MAC:"0a:7d:2a:c4:06:9b", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:47:50.431004 env[1311]: 2025-05-08 00:47:50.428 [INFO][5138] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4" Namespace="calico-apiserver" Pod="calico-apiserver-fd97795bb-zmwqb" WorkloadEndpoint="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:47:50.436083 kubelet[2246]: E0508 00:47:50.435297 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:50.440871 kubelet[2246]: E0508 00:47:50.440805 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:50.459000 audit[5182]: NETFILTER_CFG table=filter:107 family=2 entries=48 op=nft_register_chain pid=5182 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" May 8 00:47:50.459000 audit[5182]: SYSCALL arch=c000003e syscall=46 success=yes exit=25556 a0=3 a1=7fff3d07e4a0 a2=0 a3=7fff3d07e48c items=0 ppid=4459 pid=5182 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:50.459000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 May 8 00:47:50.465189 kubelet[2246]: I0508 00:47:50.464569 2246 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-8xjk8" podStartSLOduration=80.464544187 podStartE2EDuration="1m20.464544187s" podCreationTimestamp="2025-05-08 00:46:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 00:47:50.45182073 +0000 UTC m=+94.392635201" watchObservedRunningTime="2025-05-08 00:47:50.464544187 +0000 UTC m=+94.405358658" May 8 00:47:50.469279 env[1311]: time="2025-05-08T00:47:50.469161662Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 8 00:47:50.470105 env[1311]: time="2025-05-08T00:47:50.470068800Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 8 00:47:50.470328 env[1311]: time="2025-05-08T00:47:50.470266685Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 8 00:47:50.470891 env[1311]: time="2025-05-08T00:47:50.470814462Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4 pid=5181 runtime=io.containerd.runc.v2 May 8 00:47:50.483407 kubelet[2246]: I0508 00:47:50.483276 2246 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-fd97795bb-tzgmm" podStartSLOduration=68.55599969 podStartE2EDuration="1m12.483250866s" podCreationTimestamp="2025-05-08 00:46:38 +0000 UTC" firstStartedPulling="2025-05-08 00:47:45.666508811 +0000 UTC m=+89.607323282" lastFinishedPulling="2025-05-08 00:47:49.593759977 +0000 UTC m=+93.534574458" observedRunningTime="2025-05-08 00:47:50.466268402 +0000 UTC m=+94.407082903" watchObservedRunningTime="2025-05-08 00:47:50.483250866 +0000 UTC m=+94.424065327" May 8 00:47:50.522473 systemd-resolved[1221]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 8 00:47:50.552047 env[1311]: time="2025-05-08T00:47:50.551999261Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-fd97795bb-zmwqb,Uid:ca6c7950-e2c3-4481-994b-8bd117426f80,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4\"" May 8 00:47:50.554642 env[1311]: time="2025-05-08T00:47:50.554620615Z" level=info msg="CreateContainer within sandbox \"ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 8 00:47:50.572030 env[1311]: time="2025-05-08T00:47:50.571991995Z" level=info msg="CreateContainer within sandbox \"ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"601beee62f81a5c285c8b495ccf4c421b9d6e4d58537ce8ece723a3aa1d97b27\"" May 8 00:47:50.573636 env[1311]: time="2025-05-08T00:47:50.573573089Z" level=info msg="StartContainer for \"601beee62f81a5c285c8b495ccf4c421b9d6e4d58537ce8ece723a3aa1d97b27\"" May 8 00:47:50.644055 env[1311]: time="2025-05-08T00:47:50.643983522Z" level=info msg="StartContainer for \"601beee62f81a5c285c8b495ccf4c421b9d6e4d58537ce8ece723a3aa1d97b27\" returns successfully" May 8 00:47:50.702000 audit[5255]: NETFILTER_CFG table=filter:108 family=2 entries=10 op=nft_register_rule pid=5255 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:47:50.702000 audit[5255]: SYSCALL arch=c000003e syscall=46 success=yes exit=3676 a0=3 a1=7ffcfe374000 a2=0 a3=7ffcfe373fec items=0 ppid=2433 pid=5255 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:50.702000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:47:50.732000 audit[5255]: NETFILTER_CFG table=nat:109 family=2 entries=56 op=nft_register_chain pid=5255 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:47:50.732000 audit[5255]: SYSCALL arch=c000003e syscall=46 success=yes exit=19860 a0=3 a1=7ffcfe374000 a2=0 a3=7ffcfe373fec items=0 ppid=2433 pid=5255 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:50.732000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:47:51.255352 systemd-networkd[1072]: calica037e43738: Gained IPv6LL May 8 00:47:51.444831 kubelet[2246]: E0508 00:47:51.444784 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:51.445719 kubelet[2246]: E0508 00:47:51.445692 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:51.467413 kubelet[2246]: I0508 00:47:51.467318 2246 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-fd97795bb-zmwqb" podStartSLOduration=73.467292644 podStartE2EDuration="1m13.467292644s" podCreationTimestamp="2025-05-08 00:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-08 00:47:51.457728284 +0000 UTC m=+95.398542755" watchObservedRunningTime="2025-05-08 00:47:51.467292644 +0000 UTC m=+95.408107136" May 8 00:47:51.752000 audit[5267]: NETFILTER_CFG table=filter:110 family=2 entries=9 op=nft_register_rule pid=5267 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:47:51.752000 audit[5267]: SYSCALL arch=c000003e syscall=46 success=yes exit=2932 a0=3 a1=7ffd2e7176e0 a2=0 a3=7ffd2e7176cc items=0 ppid=2433 pid=5267 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:51.752000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:47:51.758000 audit[5267]: NETFILTER_CFG table=nat:111 family=2 entries=31 op=nft_register_chain pid=5267 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:47:51.758000 audit[5267]: SYSCALL arch=c000003e syscall=46 success=yes exit=11236 a0=3 a1=7ffd2e7176e0 a2=0 a3=7ffd2e7176cc items=0 ppid=2433 pid=5267 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:51.758000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:47:52.165819 env[1311]: time="2025-05-08T00:47:52.165675197Z" level=info msg="StopPodSandbox for \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\"" May 8 00:47:52.345036 systemd-networkd[1072]: calidaa7669be5f: Gained IPv6LL May 8 00:47:52.366089 env[1311]: 2025-05-08 00:47:52.321 [INFO][5285] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:47:52.366089 env[1311]: 2025-05-08 00:47:52.321 [INFO][5285] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" iface="eth0" netns="/var/run/netns/cni-34b93159-6c55-e68a-ca71-74dc7ec29dc0" May 8 00:47:52.366089 env[1311]: 2025-05-08 00:47:52.321 [INFO][5285] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" iface="eth0" netns="/var/run/netns/cni-34b93159-6c55-e68a-ca71-74dc7ec29dc0" May 8 00:47:52.366089 env[1311]: 2025-05-08 00:47:52.321 [INFO][5285] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" iface="eth0" netns="/var/run/netns/cni-34b93159-6c55-e68a-ca71-74dc7ec29dc0" May 8 00:47:52.366089 env[1311]: 2025-05-08 00:47:52.322 [INFO][5285] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:47:52.366089 env[1311]: 2025-05-08 00:47:52.322 [INFO][5285] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:47:52.366089 env[1311]: 2025-05-08 00:47:52.353 [INFO][5293] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" HandleID="k8s-pod-network.b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" Workload="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:47:52.366089 env[1311]: 2025-05-08 00:47:52.353 [INFO][5293] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:47:52.366089 env[1311]: 2025-05-08 00:47:52.353 [INFO][5293] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:47:52.366089 env[1311]: 2025-05-08 00:47:52.359 [WARNING][5293] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" HandleID="k8s-pod-network.b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" Workload="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:47:52.366089 env[1311]: 2025-05-08 00:47:52.360 [INFO][5293] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" HandleID="k8s-pod-network.b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" Workload="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:47:52.366089 env[1311]: 2025-05-08 00:47:52.361 [INFO][5293] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:47:52.366089 env[1311]: 2025-05-08 00:47:52.364 [INFO][5285] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:47:52.367047 env[1311]: time="2025-05-08T00:47:52.366988991Z" level=info msg="TearDown network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\" successfully" May 8 00:47:52.367047 env[1311]: time="2025-05-08T00:47:52.367035759Z" level=info msg="StopPodSandbox for \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\" returns successfully" May 8 00:47:52.367902 env[1311]: time="2025-05-08T00:47:52.367863326Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-2s6tt,Uid:863a2a82-a51b-4afc-a7be-990ee722b050,Namespace:calico-system,Attempt:1,}" May 8 00:47:52.370507 systemd[1]: run-netns-cni\x2d34b93159\x2d6c55\x2de68a\x2dca71\x2d74dc7ec29dc0.mount: Deactivated successfully. May 8 00:47:52.456855 kubelet[2246]: E0508 00:47:52.456214 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:47:52.516682 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready May 8 00:47:52.516828 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cali09845bb6542: link becomes ready May 8 00:47:52.515758 systemd-networkd[1072]: cali09845bb6542: Link UP May 8 00:47:52.517103 systemd-networkd[1072]: cali09845bb6542: Gained carrier May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.418 [INFO][5305] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--2s6tt-eth0 csi-node-driver- calico-system 863a2a82-a51b-4afc-a7be-990ee722b050 1262 0 2025-05-08 00:46:38 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:55b7b4b9d k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-2s6tt eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali09845bb6542 [] []}} ContainerID="c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" Namespace="calico-system" Pod="csi-node-driver-2s6tt" WorkloadEndpoint="localhost-k8s-csi--node--driver--2s6tt-" May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.418 [INFO][5305] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" Namespace="calico-system" Pod="csi-node-driver-2s6tt" WorkloadEndpoint="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.462 [INFO][5315] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" HandleID="k8s-pod-network.c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" Workload="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.471 [INFO][5315] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" HandleID="k8s-pod-network.c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" Workload="localhost-k8s-csi--node--driver--2s6tt-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000313730), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-2s6tt", "timestamp":"2025-05-08 00:47:52.46214366 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.471 [INFO][5315] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.471 [INFO][5315] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.471 [INFO][5315] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.474 [INFO][5315] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" host="localhost" May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.483 [INFO][5315] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.487 [INFO][5315] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.490 [INFO][5315] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.493 [INFO][5315] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.493 [INFO][5315] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" host="localhost" May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.494 [INFO][5315] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565 May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.499 [INFO][5315] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" host="localhost" May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.509 [INFO][5315] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" host="localhost" May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.509 [INFO][5315] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" host="localhost" May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.509 [INFO][5315] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:47:52.536163 env[1311]: 2025-05-08 00:47:52.509 [INFO][5315] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" HandleID="k8s-pod-network.c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" Workload="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:47:52.536948 env[1311]: 2025-05-08 00:47:52.512 [INFO][5305] cni-plugin/k8s.go 386: Populated endpoint ContainerID="c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" Namespace="calico-system" Pod="csi-node-driver-2s6tt" WorkloadEndpoint="localhost-k8s-csi--node--driver--2s6tt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--2s6tt-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"863a2a82-a51b-4afc-a7be-990ee722b050", ResourceVersion:"1262", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b7b4b9d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-2s6tt", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali09845bb6542", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:47:52.536948 env[1311]: 2025-05-08 00:47:52.512 [INFO][5305] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.133/32] ContainerID="c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" Namespace="calico-system" Pod="csi-node-driver-2s6tt" WorkloadEndpoint="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:47:52.536948 env[1311]: 2025-05-08 00:47:52.512 [INFO][5305] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali09845bb6542 ContainerID="c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" Namespace="calico-system" Pod="csi-node-driver-2s6tt" WorkloadEndpoint="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:47:52.536948 env[1311]: 2025-05-08 00:47:52.516 [INFO][5305] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" Namespace="calico-system" Pod="csi-node-driver-2s6tt" WorkloadEndpoint="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:47:52.536948 env[1311]: 2025-05-08 00:47:52.517 [INFO][5305] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" Namespace="calico-system" Pod="csi-node-driver-2s6tt" WorkloadEndpoint="localhost-k8s-csi--node--driver--2s6tt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--2s6tt-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"863a2a82-a51b-4afc-a7be-990ee722b050", ResourceVersion:"1262", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b7b4b9d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565", Pod:"csi-node-driver-2s6tt", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali09845bb6542", MAC:"f2:d5:05:1c:24:62", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:47:52.536948 env[1311]: 2025-05-08 00:47:52.529 [INFO][5305] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565" Namespace="calico-system" Pod="csi-node-driver-2s6tt" WorkloadEndpoint="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:47:52.548000 audit[5335]: NETFILTER_CFG table=filter:112 family=2 entries=46 op=nft_register_chain pid=5335 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" May 8 00:47:52.551335 kernel: kauditd_printk_skb: 34 callbacks suppressed May 8 00:47:52.551432 kernel: audit: type=1325 audit(1746665272.548:510): table=filter:112 family=2 entries=46 op=nft_register_chain pid=5335 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" May 8 00:47:52.548000 audit[5335]: SYSCALL arch=c000003e syscall=46 success=yes exit=22696 a0=3 a1=7ffe16b36e00 a2=0 a3=7ffe16b36dec items=0 ppid=4459 pid=5335 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:52.559342 kernel: audit: type=1300 audit(1746665272.548:510): arch=c000003e syscall=46 success=yes exit=22696 a0=3 a1=7ffe16b36e00 a2=0 a3=7ffe16b36dec items=0 ppid=4459 pid=5335 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:52.566130 env[1311]: time="2025-05-08T00:47:52.561055764Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 8 00:47:52.566130 env[1311]: time="2025-05-08T00:47:52.561135555Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 8 00:47:52.566130 env[1311]: time="2025-05-08T00:47:52.561160272Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 8 00:47:52.566130 env[1311]: time="2025-05-08T00:47:52.561370590Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565 pid=5343 runtime=io.containerd.runc.v2 May 8 00:47:52.548000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 May 8 00:47:52.572668 kernel: audit: type=1327 audit(1746665272.548:510): proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 May 8 00:47:52.595165 systemd-resolved[1221]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 8 00:47:52.610690 env[1311]: time="2025-05-08T00:47:52.610629218Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-2s6tt,Uid:863a2a82-a51b-4afc-a7be-990ee722b050,Namespace:calico-system,Attempt:1,} returns sandbox id \"c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565\"" May 8 00:47:52.612601 env[1311]: time="2025-05-08T00:47:52.612566506Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\"" May 8 00:47:52.717804 systemd[1]: Started sshd@19-10.0.0.82:22-10.0.0.1:33274.service. May 8 00:47:52.716000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.82:22-10.0.0.1:33274 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:52.724175 kernel: audit: type=1130 audit(1746665272.716:511): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.82:22-10.0.0.1:33274 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:52.763000 audit[5380]: USER_ACCT pid=5380 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:52.766388 sshd[5380]: Accepted publickey for core from 10.0.0.1 port 33274 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:52.764000 audit[5380]: CRED_ACQ pid=5380 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:52.769322 sshd[5380]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:52.773574 kernel: audit: type=1101 audit(1746665272.763:512): pid=5380 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:52.773678 kernel: audit: type=1103 audit(1746665272.764:513): pid=5380 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:52.776318 kernel: audit: type=1006 audit(1746665272.765:514): pid=5380 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=20 res=1 May 8 00:47:52.775458 systemd[1]: Started session-20.scope. May 8 00:47:52.775890 systemd-logind[1291]: New session 20 of user core. May 8 00:47:52.765000 audit[5380]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffc9a53fee0 a2=3 a3=0 items=0 ppid=1 pid=5380 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=20 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:52.781757 kernel: audit: type=1300 audit(1746665272.765:514): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffc9a53fee0 a2=3 a3=0 items=0 ppid=1 pid=5380 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=20 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:52.781825 kernel: audit: type=1327 audit(1746665272.765:514): proctitle=737368643A20636F7265205B707269765D May 8 00:47:52.765000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:52.781000 audit[5380]: USER_START pid=5380 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:52.787766 kernel: audit: type=1105 audit(1746665272.781:515): pid=5380 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:52.783000 audit[5383]: CRED_ACQ pid=5383 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:52.901296 sshd[5380]: pam_unix(sshd:session): session closed for user core May 8 00:47:52.901000 audit[5380]: USER_END pid=5380 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:52.901000 audit[5380]: CRED_DISP pid=5380 uid=0 auid=500 ses=20 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:52.904769 systemd[1]: Started sshd@20-10.0.0.82:22-10.0.0.1:33280.service. May 8 00:47:52.905864 systemd[1]: sshd@19-10.0.0.82:22-10.0.0.1:33274.service: Deactivated successfully. May 8 00:47:52.907747 systemd[1]: session-20.scope: Deactivated successfully. May 8 00:47:52.904000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.82:22-10.0.0.1:33280 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:52.905000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@19-10.0.0.82:22-10.0.0.1:33274 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:52.908718 systemd-logind[1291]: Session 20 logged out. Waiting for processes to exit. May 8 00:47:52.910179 systemd-logind[1291]: Removed session 20. May 8 00:47:52.950000 audit[5393]: USER_ACCT pid=5393 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:52.952179 sshd[5393]: Accepted publickey for core from 10.0.0.1 port 33280 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:52.952000 audit[5393]: CRED_ACQ pid=5393 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:52.952000 audit[5393]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffeac65eaa0 a2=3 a3=0 items=0 ppid=1 pid=5393 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=21 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:52.952000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:52.953672 sshd[5393]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:52.958492 systemd-logind[1291]: New session 21 of user core. May 8 00:47:52.959165 systemd[1]: Started session-21.scope. May 8 00:47:52.965000 audit[5393]: USER_START pid=5393 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:52.967000 audit[5398]: CRED_ACQ pid=5398 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:53.166066 env[1311]: time="2025-05-08T00:47:53.166015708Z" level=info msg="StopPodSandbox for \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\"" May 8 00:47:53.369699 systemd[1]: run-containerd-runc-k8s.io-c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565-runc.I6JtTM.mount: Deactivated successfully. May 8 00:47:53.878910 systemd-networkd[1072]: cali09845bb6542: Gained IPv6LL May 8 00:47:54.145091 env[1311]: 2025-05-08 00:47:53.406 [INFO][5421] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:47:54.145091 env[1311]: 2025-05-08 00:47:53.406 [INFO][5421] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" iface="eth0" netns="/var/run/netns/cni-7767e973-86e9-4715-0fc9-079ef00e00f8" May 8 00:47:54.145091 env[1311]: 2025-05-08 00:47:53.406 [INFO][5421] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" iface="eth0" netns="/var/run/netns/cni-7767e973-86e9-4715-0fc9-079ef00e00f8" May 8 00:47:54.145091 env[1311]: 2025-05-08 00:47:53.406 [INFO][5421] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" iface="eth0" netns="/var/run/netns/cni-7767e973-86e9-4715-0fc9-079ef00e00f8" May 8 00:47:54.145091 env[1311]: 2025-05-08 00:47:53.406 [INFO][5421] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:47:54.145091 env[1311]: 2025-05-08 00:47:53.406 [INFO][5421] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:47:54.145091 env[1311]: 2025-05-08 00:47:54.132 [INFO][5429] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" HandleID="k8s-pod-network.d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" Workload="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:47:54.145091 env[1311]: 2025-05-08 00:47:54.132 [INFO][5429] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:47:54.145091 env[1311]: 2025-05-08 00:47:54.132 [INFO][5429] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:47:54.145091 env[1311]: 2025-05-08 00:47:54.139 [WARNING][5429] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" HandleID="k8s-pod-network.d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" Workload="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:47:54.145091 env[1311]: 2025-05-08 00:47:54.139 [INFO][5429] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" HandleID="k8s-pod-network.d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" Workload="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:47:54.145091 env[1311]: 2025-05-08 00:47:54.141 [INFO][5429] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:47:54.145091 env[1311]: 2025-05-08 00:47:54.143 [INFO][5421] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:47:54.145091 env[1311]: time="2025-05-08T00:47:54.145051198Z" level=info msg="TearDown network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\" successfully" May 8 00:47:54.145682 env[1311]: time="2025-05-08T00:47:54.145093879Z" level=info msg="StopPodSandbox for \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\" returns successfully" May 8 00:47:54.146065 env[1311]: time="2025-05-08T00:47:54.146003120Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7f6ff4b949-8b2r4,Uid:191fecaf-17dc-4c3f-a824-7b4430fd6b4f,Namespace:calico-system,Attempt:1,}" May 8 00:47:54.148255 systemd[1]: run-netns-cni\x2d7767e973\x2d86e9\x2d4715\x2d0fc9\x2d079ef00e00f8.mount: Deactivated successfully. May 8 00:47:54.218808 sshd[5393]: pam_unix(sshd:session): session closed for user core May 8 00:47:54.218000 audit[5393]: USER_END pid=5393 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:54.218000 audit[5393]: CRED_DISP pid=5393 uid=0 auid=500 ses=21 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:54.221629 systemd[1]: Started sshd@21-10.0.0.82:22-10.0.0.1:33296.service. May 8 00:47:54.220000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.82:22-10.0.0.1:33296 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:54.222379 systemd[1]: sshd@20-10.0.0.82:22-10.0.0.1:33280.service: Deactivated successfully. May 8 00:47:54.221000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@20-10.0.0.82:22-10.0.0.1:33280 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:54.223402 systemd[1]: session-21.scope: Deactivated successfully. May 8 00:47:54.225439 systemd-logind[1291]: Session 21 logged out. Waiting for processes to exit. May 8 00:47:54.226848 systemd-logind[1291]: Removed session 21. May 8 00:47:54.266000 audit[5450]: USER_ACCT pid=5450 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:54.267864 sshd[5450]: Accepted publickey for core from 10.0.0.1 port 33296 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:54.267000 audit[5450]: CRED_ACQ pid=5450 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:54.268000 audit[5450]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffec7b75730 a2=3 a3=0 items=0 ppid=1 pid=5450 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=22 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:54.268000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:54.269560 sshd[5450]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:54.275445 systemd[1]: Started session-22.scope. May 8 00:47:54.275879 systemd-logind[1291]: New session 22 of user core. May 8 00:47:54.285000 audit[5450]: USER_START pid=5450 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:54.291000 audit[5466]: CRED_ACQ pid=5466 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:54.335316 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready May 8 00:47:54.335468 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): calie6fc0b49f4e: link becomes ready May 8 00:47:54.335764 systemd-networkd[1072]: calie6fc0b49f4e: Link UP May 8 00:47:54.337035 systemd-networkd[1072]: calie6fc0b49f4e: Gained carrier May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.236 [INFO][5437] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0 calico-kube-controllers-7f6ff4b949- calico-system 191fecaf-17dc-4c3f-a824-7b4430fd6b4f 1274 0 2025-05-08 00:46:38 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:7f6ff4b949 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-7f6ff4b949-8b2r4 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calie6fc0b49f4e [] []}} ContainerID="fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" Namespace="calico-system" Pod="calico-kube-controllers-7f6ff4b949-8b2r4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-" May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.236 [INFO][5437] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" Namespace="calico-system" Pod="calico-kube-controllers-7f6ff4b949-8b2r4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.266 [INFO][5457] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" HandleID="k8s-pod-network.fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" Workload="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.275 [INFO][5457] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" HandleID="k8s-pod-network.fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" Workload="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001327d0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-7f6ff4b949-8b2r4", "timestamp":"2025-05-08 00:47:54.266088822 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.275 [INFO][5457] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.275 [INFO][5457] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.275 [INFO][5457] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.277 [INFO][5457] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" host="localhost" May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.290 [INFO][5457] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.299 [INFO][5457] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.303 [INFO][5457] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.307 [INFO][5457] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.308 [INFO][5457] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" host="localhost" May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.310 [INFO][5457] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531 May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.314 [INFO][5457] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" host="localhost" May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.322 [INFO][5457] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" host="localhost" May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.322 [INFO][5457] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" host="localhost" May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.322 [INFO][5457] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:47:54.354520 env[1311]: 2025-05-08 00:47:54.322 [INFO][5457] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" HandleID="k8s-pod-network.fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" Workload="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:47:54.355559 env[1311]: 2025-05-08 00:47:54.326 [INFO][5437] cni-plugin/k8s.go 386: Populated endpoint ContainerID="fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" Namespace="calico-system" Pod="calico-kube-controllers-7f6ff4b949-8b2r4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0", GenerateName:"calico-kube-controllers-7f6ff4b949-", Namespace:"calico-system", SelfLink:"", UID:"191fecaf-17dc-4c3f-a824-7b4430fd6b4f", ResourceVersion:"1274", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7f6ff4b949", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-7f6ff4b949-8b2r4", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calie6fc0b49f4e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:47:54.355559 env[1311]: 2025-05-08 00:47:54.327 [INFO][5437] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.134/32] ContainerID="fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" Namespace="calico-system" Pod="calico-kube-controllers-7f6ff4b949-8b2r4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:47:54.355559 env[1311]: 2025-05-08 00:47:54.327 [INFO][5437] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie6fc0b49f4e ContainerID="fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" Namespace="calico-system" Pod="calico-kube-controllers-7f6ff4b949-8b2r4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:47:54.355559 env[1311]: 2025-05-08 00:47:54.329 [INFO][5437] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" Namespace="calico-system" Pod="calico-kube-controllers-7f6ff4b949-8b2r4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:47:54.355559 env[1311]: 2025-05-08 00:47:54.329 [INFO][5437] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" Namespace="calico-system" Pod="calico-kube-controllers-7f6ff4b949-8b2r4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0", GenerateName:"calico-kube-controllers-7f6ff4b949-", Namespace:"calico-system", SelfLink:"", UID:"191fecaf-17dc-4c3f-a824-7b4430fd6b4f", ResourceVersion:"1274", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7f6ff4b949", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531", Pod:"calico-kube-controllers-7f6ff4b949-8b2r4", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calie6fc0b49f4e", MAC:"aa:f9:a0:31:85:39", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:47:54.355559 env[1311]: 2025-05-08 00:47:54.350 [INFO][5437] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531" Namespace="calico-system" Pod="calico-kube-controllers-7f6ff4b949-8b2r4" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:47:54.374000 audit[5494]: NETFILTER_CFG table=filter:113 family=2 entries=46 op=nft_register_chain pid=5494 subj=system_u:system_r:kernel_t:s0 comm="iptables-nft-re" May 8 00:47:54.374000 audit[5494]: SYSCALL arch=c000003e syscall=46 success=yes exit=22188 a0=3 a1=7ffea73824a0 a2=0 a3=7ffea738248c items=0 ppid=4459 pid=5494 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-nft-re" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:54.374000 audit: PROCTITLE proctitle=69707461626C65732D6E66742D726573746F7265002D2D6E6F666C757368002D2D766572626F7365002D2D77616974003130002D2D776169742D696E74657276616C003530303030 May 8 00:47:54.384488 env[1311]: time="2025-05-08T00:47:54.384406407Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 8 00:47:54.384664 env[1311]: time="2025-05-08T00:47:54.384494183Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 8 00:47:54.384664 env[1311]: time="2025-05-08T00:47:54.384538287Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 8 00:47:54.384853 env[1311]: time="2025-05-08T00:47:54.384789703Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531 pid=5499 runtime=io.containerd.runc.v2 May 8 00:47:54.404379 systemd[1]: run-containerd-runc-k8s.io-fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531-runc.Z4eHQd.mount: Deactivated successfully. May 8 00:47:54.417620 systemd-resolved[1221]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 8 00:47:54.443706 env[1311]: time="2025-05-08T00:47:54.443627367Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7f6ff4b949-8b2r4,Uid:191fecaf-17dc-4c3f-a824-7b4430fd6b4f,Namespace:calico-system,Attempt:1,} returns sandbox id \"fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531\"" May 8 00:47:54.771558 env[1311]: time="2025-05-08T00:47:54.771493140Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/csi:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:54.773415 env[1311]: time="2025-05-08T00:47:54.773384750Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:54.774832 env[1311]: time="2025-05-08T00:47:54.774805809Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/csi:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:54.776457 env[1311]: time="2025-05-08T00:47:54.776427778Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:54.776884 env[1311]: time="2025-05-08T00:47:54.776851530Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\" returns image reference \"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\"" May 8 00:47:54.777804 env[1311]: time="2025-05-08T00:47:54.777770610Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\"" May 8 00:47:54.778723 env[1311]: time="2025-05-08T00:47:54.778699297Z" level=info msg="CreateContainer within sandbox \"c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" May 8 00:47:54.794238 env[1311]: time="2025-05-08T00:47:54.794177832Z" level=info msg="CreateContainer within sandbox \"c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"d02015764f5c38f81dd541f28acffb5d1484b4bac7a2e6929d01446c24a6d1e5\"" May 8 00:47:54.794862 env[1311]: time="2025-05-08T00:47:54.794840656Z" level=info msg="StartContainer for \"d02015764f5c38f81dd541f28acffb5d1484b4bac7a2e6929d01446c24a6d1e5\"" May 8 00:47:54.868098 env[1311]: time="2025-05-08T00:47:54.868032558Z" level=info msg="StartContainer for \"d02015764f5c38f81dd541f28acffb5d1484b4bac7a2e6929d01446c24a6d1e5\" returns successfully" May 8 00:47:55.946000 audit[5581]: NETFILTER_CFG table=filter:114 family=2 entries=20 op=nft_register_rule pid=5581 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:47:55.946000 audit[5581]: SYSCALL arch=c000003e syscall=46 success=yes exit=11860 a0=3 a1=7ffd51822df0 a2=0 a3=7ffd51822ddc items=0 ppid=2433 pid=5581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:55.946000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:47:55.955000 audit[5581]: NETFILTER_CFG table=nat:115 family=2 entries=22 op=nft_register_rule pid=5581 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:47:55.955000 audit[5581]: SYSCALL arch=c000003e syscall=46 success=yes exit=6540 a0=3 a1=7ffd51822df0 a2=0 a3=0 items=0 ppid=2433 pid=5581 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:55.955000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:47:55.962564 sshd[5450]: pam_unix(sshd:session): session closed for user core May 8 00:47:55.963000 audit[5450]: USER_END pid=5450 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:55.963000 audit[5450]: CRED_DISP pid=5450 uid=0 auid=500 ses=22 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:55.965000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.82:22-10.0.0.1:45178 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:55.966405 systemd[1]: Started sshd@22-10.0.0.82:22-10.0.0.1:45178.service. May 8 00:47:55.967104 systemd[1]: sshd@21-10.0.0.82:22-10.0.0.1:33296.service: Deactivated successfully. May 8 00:47:55.967000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@21-10.0.0.82:22-10.0.0.1:33296 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:55.969987 systemd-logind[1291]: Session 22 logged out. Waiting for processes to exit. May 8 00:47:55.971058 systemd[1]: session-22.scope: Deactivated successfully. May 8 00:47:55.973162 systemd-logind[1291]: Removed session 22. May 8 00:47:55.973000 audit[5586]: NETFILTER_CFG table=filter:116 family=2 entries=32 op=nft_register_rule pid=5586 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:47:55.973000 audit[5586]: SYSCALL arch=c000003e syscall=46 success=yes exit=11860 a0=3 a1=7ffdc594df80 a2=0 a3=7ffdc594df6c items=0 ppid=2433 pid=5586 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:55.973000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:47:55.979000 audit[5586]: NETFILTER_CFG table=nat:117 family=2 entries=22 op=nft_register_rule pid=5586 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:47:55.979000 audit[5586]: SYSCALL arch=c000003e syscall=46 success=yes exit=6540 a0=3 a1=7ffdc594df80 a2=0 a3=0 items=0 ppid=2433 pid=5586 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:55.979000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:47:55.990510 systemd-networkd[1072]: calie6fc0b49f4e: Gained IPv6LL May 8 00:47:56.015571 sshd[5583]: Accepted publickey for core from 10.0.0.1 port 45178 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:56.014000 audit[5583]: USER_ACCT pid=5583 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:56.015000 audit[5583]: CRED_ACQ pid=5583 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:56.015000 audit[5583]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffe7b28e9d0 a2=3 a3=0 items=0 ppid=1 pid=5583 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=23 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:56.015000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:56.017129 sshd[5583]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:56.020955 systemd-logind[1291]: New session 23 of user core. May 8 00:47:56.022020 systemd[1]: Started session-23.scope. May 8 00:47:56.029000 audit[5583]: USER_START pid=5583 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:56.030000 audit[5589]: CRED_ACQ pid=5589 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:57.505922 sshd[5583]: pam_unix(sshd:session): session closed for user core May 8 00:47:57.506000 audit[5583]: USER_END pid=5583 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:57.506000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-10.0.0.82:22-10.0.0.1:45188 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:57.506000 audit[5583]: CRED_DISP pid=5583 uid=0 auid=500 ses=23 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:57.508000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@22-10.0.0.82:22-10.0.0.1:45178 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:57.507077 systemd[1]: Started sshd@23-10.0.0.82:22-10.0.0.1:45188.service. May 8 00:47:57.508885 systemd[1]: sshd@22-10.0.0.82:22-10.0.0.1:45178.service: Deactivated successfully. May 8 00:47:57.510050 systemd[1]: session-23.scope: Deactivated successfully. May 8 00:47:57.511151 systemd-logind[1291]: Session 23 logged out. Waiting for processes to exit. May 8 00:47:57.515966 systemd-logind[1291]: Removed session 23. May 8 00:47:57.569685 kernel: kauditd_printk_skb: 53 callbacks suppressed May 8 00:47:57.569845 kernel: audit: type=1101 audit(1746665277.557:553): pid=5596 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:57.569882 kernel: audit: type=1103 audit(1746665277.558:554): pid=5596 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:57.569903 kernel: audit: type=1006 audit(1746665277.558:555): pid=5596 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=24 res=1 May 8 00:47:57.557000 audit[5596]: USER_ACCT pid=5596 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:57.558000 audit[5596]: CRED_ACQ pid=5596 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:57.560428 sshd[5596]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:47:57.570388 sshd[5596]: Accepted publickey for core from 10.0.0.1 port 45188 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:47:57.558000 audit[5596]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffc85ee1050 a2=3 a3=0 items=0 ppid=1 pid=5596 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=24 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:57.571958 systemd-logind[1291]: New session 24 of user core. May 8 00:47:57.572498 systemd[1]: Started session-24.scope. May 8 00:47:57.575856 kernel: audit: type=1300 audit(1746665277.558:555): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffc85ee1050 a2=3 a3=0 items=0 ppid=1 pid=5596 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=24 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:47:57.575925 kernel: audit: type=1327 audit(1746665277.558:555): proctitle=737368643A20636F7265205B707269765D May 8 00:47:57.558000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:47:57.576000 audit[5596]: USER_START pid=5596 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:57.578000 audit[5601]: CRED_ACQ pid=5601 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:57.586180 kernel: audit: type=1105 audit(1746665277.576:556): pid=5596 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:57.586288 kernel: audit: type=1103 audit(1746665277.578:557): pid=5601 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:58.280325 env[1311]: time="2025-05-08T00:47:58.280244980Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/kube-controllers:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:58.282839 env[1311]: time="2025-05-08T00:47:58.282806234Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:58.285148 env[1311]: time="2025-05-08T00:47:58.285119519Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/kube-controllers:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:58.286976 env[1311]: time="2025-05-08T00:47:58.286919012Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:47:58.287291 env[1311]: time="2025-05-08T00:47:58.287262312Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" returns image reference \"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\"" May 8 00:47:58.289093 env[1311]: time="2025-05-08T00:47:58.289057377Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\"" May 8 00:47:58.306312 env[1311]: time="2025-05-08T00:47:58.306217855Z" level=info msg="CreateContainer within sandbox \"fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" May 8 00:47:58.328170 env[1311]: time="2025-05-08T00:47:58.328040459Z" level=info msg="CreateContainer within sandbox \"fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"17cd06d969354194d48f416bcae0a8966b438de64e0d4fd9c9a7ad8d6dc8bdf7\"" May 8 00:47:58.328723 env[1311]: time="2025-05-08T00:47:58.328666123Z" level=info msg="StartContainer for \"17cd06d969354194d48f416bcae0a8966b438de64e0d4fd9c9a7ad8d6dc8bdf7\"" May 8 00:47:58.334125 sshd[5596]: pam_unix(sshd:session): session closed for user core May 8 00:47:58.334000 audit[5596]: USER_END pid=5596 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:58.336825 systemd[1]: sshd@23-10.0.0.82:22-10.0.0.1:45188.service: Deactivated successfully. May 8 00:47:58.338144 systemd[1]: session-24.scope: Deactivated successfully. May 8 00:47:58.338242 systemd-logind[1291]: Session 24 logged out. Waiting for processes to exit. May 8 00:47:58.339217 systemd-logind[1291]: Removed session 24. May 8 00:47:58.334000 audit[5596]: CRED_DISP pid=5596 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:58.348184 kernel: audit: type=1106 audit(1746665278.334:558): pid=5596 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:58.348308 kernel: audit: type=1104 audit(1746665278.334:559): pid=5596 uid=0 auid=500 ses=24 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:47:58.348561 kernel: audit: type=1131 audit(1746665278.335:560): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-10.0.0.82:22-10.0.0.1:45188 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:58.335000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@23-10.0.0.82:22-10.0.0.1:45188 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:47:58.398198 env[1311]: time="2025-05-08T00:47:58.398151698Z" level=info msg="StartContainer for \"17cd06d969354194d48f416bcae0a8966b438de64e0d4fd9c9a7ad8d6dc8bdf7\" returns successfully" May 8 00:47:58.483639 kubelet[2246]: I0508 00:47:58.483261 2246 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-7f6ff4b949-8b2r4" podStartSLOduration=76.640168355 podStartE2EDuration="1m20.483236035s" podCreationTimestamp="2025-05-08 00:46:38 +0000 UTC" firstStartedPulling="2025-05-08 00:47:54.445237254 +0000 UTC m=+98.386051725" lastFinishedPulling="2025-05-08 00:47:58.288304924 +0000 UTC m=+102.229119405" observedRunningTime="2025-05-08 00:47:58.481925136 +0000 UTC m=+102.422739607" watchObservedRunningTime="2025-05-08 00:47:58.483236035 +0000 UTC m=+102.424050506" May 8 00:47:59.303802 systemd[1]: run-containerd-runc-k8s.io-17cd06d969354194d48f416bcae0a8966b438de64e0d4fd9c9a7ad8d6dc8bdf7-runc.afJLQx.mount: Deactivated successfully. May 8 00:48:00.169641 env[1311]: time="2025-05-08T00:48:00.169582261Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:48:00.171880 env[1311]: time="2025-05-08T00:48:00.171814702Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:48:00.173613 env[1311]: time="2025-05-08T00:48:00.173580901Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:48:00.175345 env[1311]: time="2025-05-08T00:48:00.175316103Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 8 00:48:00.175925 env[1311]: time="2025-05-08T00:48:00.175850413Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" returns image reference \"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\"" May 8 00:48:00.178424 env[1311]: time="2025-05-08T00:48:00.178357613Z" level=info msg="CreateContainer within sandbox \"c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" May 8 00:48:00.193748 env[1311]: time="2025-05-08T00:48:00.193666660Z" level=info msg="CreateContainer within sandbox \"c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"ad902852e419073198cebe4c41074e8d90e26a23986230fbcc32081092800c68\"" May 8 00:48:00.194344 env[1311]: time="2025-05-08T00:48:00.194305429Z" level=info msg="StartContainer for \"ad902852e419073198cebe4c41074e8d90e26a23986230fbcc32081092800c68\"" May 8 00:48:00.250652 env[1311]: time="2025-05-08T00:48:00.250552704Z" level=info msg="StartContainer for \"ad902852e419073198cebe4c41074e8d90e26a23986230fbcc32081092800c68\" returns successfully" May 8 00:48:00.311067 kubelet[2246]: I0508 00:48:00.310994 2246 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 May 8 00:48:00.311818 kubelet[2246]: I0508 00:48:00.311789 2246 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock May 8 00:48:03.336480 systemd[1]: Started sshd@24-10.0.0.82:22-10.0.0.1:45190.service. May 8 00:48:03.335000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@24-10.0.0.82:22-10.0.0.1:45190 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:03.341153 kernel: audit: type=1130 audit(1746665283.335:561): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@24-10.0.0.82:22-10.0.0.1:45190 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:03.384000 audit[5710]: USER_ACCT pid=5710 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:03.386430 sshd[5710]: Accepted publickey for core from 10.0.0.1 port 45190 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:48:03.388077 sshd[5710]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:48:03.393189 systemd[1]: Started session-25.scope. May 8 00:48:03.393656 systemd-logind[1291]: New session 25 of user core. May 8 00:48:03.386000 audit[5710]: CRED_ACQ pid=5710 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:03.400198 kernel: audit: type=1101 audit(1746665283.384:562): pid=5710 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:03.400314 kernel: audit: type=1103 audit(1746665283.386:563): pid=5710 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:03.400348 kernel: audit: type=1006 audit(1746665283.386:564): pid=5710 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=25 res=1 May 8 00:48:03.386000 audit[5710]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffd42826270 a2=3 a3=0 items=0 ppid=1 pid=5710 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=25 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:48:03.406968 kernel: audit: type=1300 audit(1746665283.386:564): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffd42826270 a2=3 a3=0 items=0 ppid=1 pid=5710 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=25 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:48:03.407038 kernel: audit: type=1327 audit(1746665283.386:564): proctitle=737368643A20636F7265205B707269765D May 8 00:48:03.386000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:48:03.397000 audit[5710]: USER_START pid=5710 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:03.424103 kernel: audit: type=1105 audit(1746665283.397:565): pid=5710 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:03.424299 kernel: audit: type=1103 audit(1746665283.399:566): pid=5713 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:03.399000 audit[5713]: CRED_ACQ pid=5713 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:03.572842 sshd[5710]: pam_unix(sshd:session): session closed for user core May 8 00:48:03.572000 audit[5710]: USER_END pid=5710 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:03.575481 systemd[1]: sshd@24-10.0.0.82:22-10.0.0.1:45190.service: Deactivated successfully. May 8 00:48:03.576500 systemd[1]: session-25.scope: Deactivated successfully. May 8 00:48:03.572000 audit[5710]: CRED_DISP pid=5710 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:03.579704 systemd-logind[1291]: Session 25 logged out. Waiting for processes to exit. May 8 00:48:03.580653 systemd-logind[1291]: Removed session 25. May 8 00:48:03.596408 kernel: audit: type=1106 audit(1746665283.572:567): pid=5710 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:03.596641 kernel: audit: type=1104 audit(1746665283.572:568): pid=5710 uid=0 auid=500 ses=25 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:03.572000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@24-10.0.0.82:22-10.0.0.1:45190 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:04.166193 kubelet[2246]: E0508 00:48:04.166145 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:48:08.575000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-10.0.0.82:22-10.0.0.1:33168 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:08.576588 systemd[1]: Started sshd@25-10.0.0.82:22-10.0.0.1:33168.service. May 8 00:48:08.593167 kernel: kauditd_printk_skb: 1 callbacks suppressed May 8 00:48:08.593306 kernel: audit: type=1130 audit(1746665288.575:570): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-10.0.0.82:22-10.0.0.1:33168 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:08.629000 audit[5732]: USER_ACCT pid=5732 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:08.631327 sshd[5732]: Accepted publickey for core from 10.0.0.1 port 33168 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:48:08.633560 sshd[5732]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:48:08.632000 audit[5732]: CRED_ACQ pid=5732 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:08.638937 systemd-logind[1291]: New session 26 of user core. May 8 00:48:08.639914 kernel: audit: type=1101 audit(1746665288.629:571): pid=5732 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:08.639997 kernel: audit: type=1103 audit(1746665288.632:572): pid=5732 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:08.640030 kernel: audit: type=1006 audit(1746665288.632:573): pid=5732 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=26 res=1 May 8 00:48:08.640244 systemd[1]: Started session-26.scope. May 8 00:48:08.642621 kernel: audit: type=1300 audit(1746665288.632:573): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffefd8d5190 a2=3 a3=0 items=0 ppid=1 pid=5732 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=26 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:48:08.632000 audit[5732]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffefd8d5190 a2=3 a3=0 items=0 ppid=1 pid=5732 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=26 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:48:08.646982 kernel: audit: type=1327 audit(1746665288.632:573): proctitle=737368643A20636F7265205B707269765D May 8 00:48:08.632000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:48:08.644000 audit[5732]: USER_START pid=5732 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:08.652957 kernel: audit: type=1105 audit(1746665288.644:574): pid=5732 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:08.653010 kernel: audit: type=1103 audit(1746665288.645:575): pid=5735 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:08.645000 audit[5735]: CRED_ACQ pid=5735 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:08.771096 sshd[5732]: pam_unix(sshd:session): session closed for user core May 8 00:48:08.771000 audit[5732]: USER_END pid=5732 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:08.775361 systemd[1]: sshd@25-10.0.0.82:22-10.0.0.1:33168.service: Deactivated successfully. May 8 00:48:08.776839 systemd-logind[1291]: Session 26 logged out. Waiting for processes to exit. May 8 00:48:08.776869 systemd[1]: session-26.scope: Deactivated successfully. May 8 00:48:08.778309 kernel: audit: type=1106 audit(1746665288.771:576): pid=5732 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:08.778431 kernel: audit: type=1104 audit(1746665288.771:577): pid=5732 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:08.771000 audit[5732]: CRED_DISP pid=5732 uid=0 auid=500 ses=26 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:08.777979 systemd-logind[1291]: Removed session 26. May 8 00:48:08.774000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@25-10.0.0.82:22-10.0.0.1:33168 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:09.553000 audit[5747]: NETFILTER_CFG table=filter:118 family=2 entries=20 op=nft_register_rule pid=5747 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:48:09.553000 audit[5747]: SYSCALL arch=c000003e syscall=46 success=yes exit=2932 a0=3 a1=7fffdd2620f0 a2=0 a3=7fffdd2620dc items=0 ppid=2433 pid=5747 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:48:09.553000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:48:09.559000 audit[5747]: NETFILTER_CFG table=nat:119 family=2 entries=106 op=nft_register_chain pid=5747 subj=system_u:system_r:kernel_t:s0 comm="iptables-restor" May 8 00:48:09.559000 audit[5747]: SYSCALL arch=c000003e syscall=46 success=yes exit=49452 a0=3 a1=7fffdd2620f0 a2=0 a3=7fffdd2620dc items=0 ppid=2433 pid=5747 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="iptables-restor" exe="/usr/sbin/xtables-nft-multi" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:48:09.559000 audit: PROCTITLE proctitle=69707461626C65732D726573746F7265002D770035002D5700313030303030002D2D6E6F666C757368002D2D636F756E74657273 May 8 00:48:09.862150 kubelet[2246]: E0508 00:48:09.862035 2246 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 8 00:48:09.923581 kubelet[2246]: I0508 00:48:09.923482 2246 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-2s6tt" podStartSLOduration=84.352705626 podStartE2EDuration="1m31.917675201s" podCreationTimestamp="2025-05-08 00:46:38 +0000 UTC" firstStartedPulling="2025-05-08 00:47:52.612020472 +0000 UTC m=+96.552834943" lastFinishedPulling="2025-05-08 00:48:00.176990047 +0000 UTC m=+104.117804518" observedRunningTime="2025-05-08 00:48:00.491809675 +0000 UTC m=+104.432624146" watchObservedRunningTime="2025-05-08 00:48:09.917675201 +0000 UTC m=+113.858489672" May 8 00:48:13.773000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-10.0.0.82:22-10.0.0.1:33176 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:13.774395 systemd[1]: Started sshd@26-10.0.0.82:22-10.0.0.1:33176.service. May 8 00:48:13.775704 kernel: kauditd_printk_skb: 7 callbacks suppressed May 8 00:48:13.775768 kernel: audit: type=1130 audit(1746665293.773:581): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-10.0.0.82:22-10.0.0.1:33176 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:13.819000 audit[5771]: USER_ACCT pid=5771 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:13.821022 sshd[5771]: Accepted publickey for core from 10.0.0.1 port 33176 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:48:13.822735 sshd[5771]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:48:14.194155 kernel: audit: type=1101 audit(1746665293.819:582): pid=5771 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:14.194524 kernel: audit: type=1103 audit(1746665293.821:583): pid=5771 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:13.821000 audit[5771]: CRED_ACQ pid=5771 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:14.198592 systemd-logind[1291]: New session 27 of user core. May 8 00:48:14.199719 systemd[1]: Started session-27.scope. May 8 00:48:14.201153 kernel: audit: type=1006 audit(1746665293.821:584): pid=5771 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=27 res=1 May 8 00:48:14.201207 kernel: audit: type=1300 audit(1746665293.821:584): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffecc4666e0 a2=3 a3=0 items=0 ppid=1 pid=5771 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=27 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:48:13.821000 audit[5771]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffecc4666e0 a2=3 a3=0 items=0 ppid=1 pid=5771 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=27 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:48:13.821000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:48:14.227277 kernel: audit: type=1327 audit(1746665293.821:584): proctitle=737368643A20636F7265205B707269765D May 8 00:48:14.227436 kernel: audit: type=1105 audit(1746665294.204:585): pid=5771 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:14.204000 audit[5771]: USER_START pid=5771 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:14.206000 audit[5774]: CRED_ACQ pid=5774 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:14.235094 kernel: audit: type=1103 audit(1746665294.206:586): pid=5774 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:14.450444 sshd[5771]: pam_unix(sshd:session): session closed for user core May 8 00:48:14.450000 audit[5771]: USER_END pid=5771 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:14.453268 systemd[1]: sshd@26-10.0.0.82:22-10.0.0.1:33176.service: Deactivated successfully. May 8 00:48:14.454786 systemd[1]: session-27.scope: Deactivated successfully. May 8 00:48:14.455265 systemd-logind[1291]: Session 27 logged out. Waiting for processes to exit. May 8 00:48:14.456271 systemd-logind[1291]: Removed session 27. May 8 00:48:14.475163 kernel: audit: type=1106 audit(1746665294.450:587): pid=5771 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:14.450000 audit[5771]: CRED_DISP pid=5771 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:14.450000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@26-10.0.0.82:22-10.0.0.1:33176 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:14.480153 kernel: audit: type=1104 audit(1746665294.450:588): pid=5771 uid=0 auid=500 ses=27 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:16.189786 env[1311]: time="2025-05-08T00:48:16.189712558Z" level=info msg="StopPodSandbox for \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\"" May 8 00:48:16.521465 env[1311]: 2025-05-08 00:48:16.435 [WARNING][5804] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"9c26b328-0da1-4b8c-b690-cd29958643ac", ResourceVersion:"1191", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388", Pod:"coredns-7db6d8ff4d-c2qjt", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali62884932f3c", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:48:16.521465 env[1311]: 2025-05-08 00:48:16.437 [INFO][5804] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:48:16.521465 env[1311]: 2025-05-08 00:48:16.437 [INFO][5804] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" iface="eth0" netns="" May 8 00:48:16.521465 env[1311]: 2025-05-08 00:48:16.438 [INFO][5804] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:48:16.521465 env[1311]: 2025-05-08 00:48:16.438 [INFO][5804] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:48:16.521465 env[1311]: 2025-05-08 00:48:16.501 [INFO][5812] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" HandleID="k8s-pod-network.85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" Workload="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:48:16.521465 env[1311]: 2025-05-08 00:48:16.502 [INFO][5812] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:48:16.521465 env[1311]: 2025-05-08 00:48:16.502 [INFO][5812] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:48:16.521465 env[1311]: 2025-05-08 00:48:16.508 [WARNING][5812] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" HandleID="k8s-pod-network.85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" Workload="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:48:16.521465 env[1311]: 2025-05-08 00:48:16.508 [INFO][5812] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" HandleID="k8s-pod-network.85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" Workload="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:48:16.521465 env[1311]: 2025-05-08 00:48:16.510 [INFO][5812] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:48:16.521465 env[1311]: 2025-05-08 00:48:16.518 [INFO][5804] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:48:16.522308 env[1311]: time="2025-05-08T00:48:16.522259525Z" level=info msg="TearDown network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\" successfully" May 8 00:48:16.522443 env[1311]: time="2025-05-08T00:48:16.522414046Z" level=info msg="StopPodSandbox for \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\" returns successfully" May 8 00:48:16.522968 env[1311]: time="2025-05-08T00:48:16.522944168Z" level=info msg="RemovePodSandbox for \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\"" May 8 00:48:16.523088 env[1311]: time="2025-05-08T00:48:16.523043004Z" level=info msg="Forcibly stopping sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\"" May 8 00:48:16.627796 env[1311]: 2025-05-08 00:48:16.572 [WARNING][5835] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"9c26b328-0da1-4b8c-b690-cd29958643ac", ResourceVersion:"1191", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"6c0eeb85372953c8871f37b49376b04304f42d245d4c1993eda3f477f8881388", Pod:"coredns-7db6d8ff4d-c2qjt", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali62884932f3c", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:48:16.627796 env[1311]: 2025-05-08 00:48:16.586 [INFO][5835] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:48:16.627796 env[1311]: 2025-05-08 00:48:16.586 [INFO][5835] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" iface="eth0" netns="" May 8 00:48:16.627796 env[1311]: 2025-05-08 00:48:16.586 [INFO][5835] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:48:16.627796 env[1311]: 2025-05-08 00:48:16.586 [INFO][5835] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:48:16.627796 env[1311]: 2025-05-08 00:48:16.618 [INFO][5844] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" HandleID="k8s-pod-network.85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" Workload="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:48:16.627796 env[1311]: 2025-05-08 00:48:16.618 [INFO][5844] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:48:16.627796 env[1311]: 2025-05-08 00:48:16.618 [INFO][5844] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:48:16.627796 env[1311]: 2025-05-08 00:48:16.623 [WARNING][5844] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" HandleID="k8s-pod-network.85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" Workload="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:48:16.627796 env[1311]: 2025-05-08 00:48:16.623 [INFO][5844] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" HandleID="k8s-pod-network.85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" Workload="localhost-k8s-coredns--7db6d8ff4d--c2qjt-eth0" May 8 00:48:16.627796 env[1311]: 2025-05-08 00:48:16.624 [INFO][5844] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:48:16.627796 env[1311]: 2025-05-08 00:48:16.626 [INFO][5835] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380" May 8 00:48:16.628436 env[1311]: time="2025-05-08T00:48:16.628391357Z" level=info msg="TearDown network for sandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\" successfully" May 8 00:48:16.767866 env[1311]: time="2025-05-08T00:48:16.767754265Z" level=info msg="RemovePodSandbox \"85fbe591ed0af0be380fab77e5c071b23ac14fc37c0e8a4d5b7288047a93e380\" returns successfully" May 8 00:48:16.768443 env[1311]: time="2025-05-08T00:48:16.768416104Z" level=info msg="StopPodSandbox for \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\"" May 8 00:48:16.768559 env[1311]: time="2025-05-08T00:48:16.768516053Z" level=info msg="TearDown network for sandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" successfully" May 8 00:48:16.768601 env[1311]: time="2025-05-08T00:48:16.768557752Z" level=info msg="StopPodSandbox for \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" returns successfully" May 8 00:48:16.768885 env[1311]: time="2025-05-08T00:48:16.768852839Z" level=info msg="RemovePodSandbox for \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\"" May 8 00:48:16.769062 env[1311]: time="2025-05-08T00:48:16.768883616Z" level=info msg="Forcibly stopping sandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\"" May 8 00:48:16.769062 env[1311]: time="2025-05-08T00:48:16.768963357Z" level=info msg="TearDown network for sandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" successfully" May 8 00:48:16.806140 env[1311]: time="2025-05-08T00:48:16.805919094Z" level=info msg="RemovePodSandbox \"c7c77083b4a8aa2ae124c051c81f6a47c2d3b4e5d7110597476de7931460b808\" returns successfully" May 8 00:48:16.816149 env[1311]: time="2025-05-08T00:48:16.816062450Z" level=info msg="StopPodSandbox for \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\"" May 8 00:48:16.916036 env[1311]: 2025-05-08 00:48:16.877 [WARNING][5868] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"435e64db-0f2e-4f05-8fcf-eefdac14188d", ResourceVersion:"1226", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98", Pod:"coredns-7db6d8ff4d-8xjk8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calica037e43738", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:48:16.916036 env[1311]: 2025-05-08 00:48:16.877 [INFO][5868] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:48:16.916036 env[1311]: 2025-05-08 00:48:16.877 [INFO][5868] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" iface="eth0" netns="" May 8 00:48:16.916036 env[1311]: 2025-05-08 00:48:16.877 [INFO][5868] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:48:16.916036 env[1311]: 2025-05-08 00:48:16.877 [INFO][5868] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:48:16.916036 env[1311]: 2025-05-08 00:48:16.904 [INFO][5877] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" HandleID="k8s-pod-network.df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" Workload="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:48:16.916036 env[1311]: 2025-05-08 00:48:16.904 [INFO][5877] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:48:16.916036 env[1311]: 2025-05-08 00:48:16.904 [INFO][5877] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:48:16.916036 env[1311]: 2025-05-08 00:48:16.910 [WARNING][5877] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" HandleID="k8s-pod-network.df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" Workload="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:48:16.916036 env[1311]: 2025-05-08 00:48:16.911 [INFO][5877] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" HandleID="k8s-pod-network.df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" Workload="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:48:16.916036 env[1311]: 2025-05-08 00:48:16.912 [INFO][5877] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:48:16.916036 env[1311]: 2025-05-08 00:48:16.914 [INFO][5868] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:48:16.916602 env[1311]: time="2025-05-08T00:48:16.916073539Z" level=info msg="TearDown network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\" successfully" May 8 00:48:16.916602 env[1311]: time="2025-05-08T00:48:16.916121359Z" level=info msg="StopPodSandbox for \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\" returns successfully" May 8 00:48:16.916753 env[1311]: time="2025-05-08T00:48:16.916718506Z" level=info msg="RemovePodSandbox for \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\"" May 8 00:48:16.916811 env[1311]: time="2025-05-08T00:48:16.916763501Z" level=info msg="Forcibly stopping sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\"" May 8 00:48:16.988736 env[1311]: 2025-05-08 00:48:16.953 [WARNING][5900] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"435e64db-0f2e-4f05-8fcf-eefdac14188d", ResourceVersion:"1226", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"7d7e0f8efa5df88f30afef922f7841215a3bcb68089c5380926ba81b38f50e98", Pod:"coredns-7db6d8ff4d-8xjk8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calica037e43738", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:48:16.988736 env[1311]: 2025-05-08 00:48:16.953 [INFO][5900] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:48:16.988736 env[1311]: 2025-05-08 00:48:16.953 [INFO][5900] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" iface="eth0" netns="" May 8 00:48:16.988736 env[1311]: 2025-05-08 00:48:16.953 [INFO][5900] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:48:16.988736 env[1311]: 2025-05-08 00:48:16.953 [INFO][5900] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:48:16.988736 env[1311]: 2025-05-08 00:48:16.977 [INFO][5908] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" HandleID="k8s-pod-network.df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" Workload="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:48:16.988736 env[1311]: 2025-05-08 00:48:16.977 [INFO][5908] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:48:16.988736 env[1311]: 2025-05-08 00:48:16.977 [INFO][5908] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:48:16.988736 env[1311]: 2025-05-08 00:48:16.983 [WARNING][5908] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" HandleID="k8s-pod-network.df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" Workload="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:48:16.988736 env[1311]: 2025-05-08 00:48:16.983 [INFO][5908] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" HandleID="k8s-pod-network.df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" Workload="localhost-k8s-coredns--7db6d8ff4d--8xjk8-eth0" May 8 00:48:16.988736 env[1311]: 2025-05-08 00:48:16.985 [INFO][5908] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:48:16.988736 env[1311]: 2025-05-08 00:48:16.987 [INFO][5900] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2" May 8 00:48:16.989384 env[1311]: time="2025-05-08T00:48:16.988755181Z" level=info msg="TearDown network for sandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\" successfully" May 8 00:48:16.993522 env[1311]: time="2025-05-08T00:48:16.993454952Z" level=info msg="RemovePodSandbox \"df90765664fee0b6ad69fef5b2dcfd7d2cb2b1cca4c89645e6e81f36e1e38cd2\" returns successfully" May 8 00:48:16.994367 env[1311]: time="2025-05-08T00:48:16.994333110Z" level=info msg="StopPodSandbox for \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\"" May 8 00:48:17.074887 env[1311]: 2025-05-08 00:48:17.036 [WARNING][5930] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0", GenerateName:"calico-kube-controllers-7f6ff4b949-", Namespace:"calico-system", SelfLink:"", UID:"191fecaf-17dc-4c3f-a824-7b4430fd6b4f", ResourceVersion:"1336", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7f6ff4b949", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531", Pod:"calico-kube-controllers-7f6ff4b949-8b2r4", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calie6fc0b49f4e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:48:17.074887 env[1311]: 2025-05-08 00:48:17.037 [INFO][5930] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:48:17.074887 env[1311]: 2025-05-08 00:48:17.037 [INFO][5930] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" iface="eth0" netns="" May 8 00:48:17.074887 env[1311]: 2025-05-08 00:48:17.037 [INFO][5930] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:48:17.074887 env[1311]: 2025-05-08 00:48:17.037 [INFO][5930] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:48:17.074887 env[1311]: 2025-05-08 00:48:17.062 [INFO][5938] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" HandleID="k8s-pod-network.d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" Workload="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:48:17.074887 env[1311]: 2025-05-08 00:48:17.062 [INFO][5938] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:48:17.074887 env[1311]: 2025-05-08 00:48:17.062 [INFO][5938] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:48:17.074887 env[1311]: 2025-05-08 00:48:17.069 [WARNING][5938] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" HandleID="k8s-pod-network.d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" Workload="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:48:17.074887 env[1311]: 2025-05-08 00:48:17.069 [INFO][5938] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" HandleID="k8s-pod-network.d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" Workload="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:48:17.074887 env[1311]: 2025-05-08 00:48:17.071 [INFO][5938] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:48:17.074887 env[1311]: 2025-05-08 00:48:17.072 [INFO][5930] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:48:17.074887 env[1311]: time="2025-05-08T00:48:17.074820555Z" level=info msg="TearDown network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\" successfully" May 8 00:48:17.074887 env[1311]: time="2025-05-08T00:48:17.074852625Z" level=info msg="StopPodSandbox for \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\" returns successfully" May 8 00:48:17.075915 env[1311]: time="2025-05-08T00:48:17.075336709Z" level=info msg="RemovePodSandbox for \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\"" May 8 00:48:17.075915 env[1311]: time="2025-05-08T00:48:17.075365333Z" level=info msg="Forcibly stopping sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\"" May 8 00:48:17.152480 env[1311]: 2025-05-08 00:48:17.116 [WARNING][5960] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0", GenerateName:"calico-kube-controllers-7f6ff4b949-", Namespace:"calico-system", SelfLink:"", UID:"191fecaf-17dc-4c3f-a824-7b4430fd6b4f", ResourceVersion:"1336", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7f6ff4b949", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"fdde066016dc10ae520f0803d631719f5bbaa647b6e7100d37ad5af0f81f1531", Pod:"calico-kube-controllers-7f6ff4b949-8b2r4", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calie6fc0b49f4e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:48:17.152480 env[1311]: 2025-05-08 00:48:17.116 [INFO][5960] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:48:17.152480 env[1311]: 2025-05-08 00:48:17.116 [INFO][5960] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" iface="eth0" netns="" May 8 00:48:17.152480 env[1311]: 2025-05-08 00:48:17.116 [INFO][5960] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:48:17.152480 env[1311]: 2025-05-08 00:48:17.116 [INFO][5960] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:48:17.152480 env[1311]: 2025-05-08 00:48:17.138 [INFO][5968] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" HandleID="k8s-pod-network.d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" Workload="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:48:17.152480 env[1311]: 2025-05-08 00:48:17.138 [INFO][5968] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:48:17.152480 env[1311]: 2025-05-08 00:48:17.138 [INFO][5968] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:48:17.152480 env[1311]: 2025-05-08 00:48:17.145 [WARNING][5968] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" HandleID="k8s-pod-network.d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" Workload="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:48:17.152480 env[1311]: 2025-05-08 00:48:17.145 [INFO][5968] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" HandleID="k8s-pod-network.d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" Workload="localhost-k8s-calico--kube--controllers--7f6ff4b949--8b2r4-eth0" May 8 00:48:17.152480 env[1311]: 2025-05-08 00:48:17.148 [INFO][5968] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:48:17.152480 env[1311]: 2025-05-08 00:48:17.150 [INFO][5960] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4" May 8 00:48:17.153303 env[1311]: time="2025-05-08T00:48:17.153231809Z" level=info msg="TearDown network for sandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\" successfully" May 8 00:48:17.159973 env[1311]: time="2025-05-08T00:48:17.159907900Z" level=info msg="RemovePodSandbox \"d7e2cb1492ce9591921643ae444934f2b7453fe0b38b3a9e9765da30ed1125c4\" returns successfully" May 8 00:48:17.161998 env[1311]: time="2025-05-08T00:48:17.161949233Z" level=info msg="StopPodSandbox for \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\"" May 8 00:48:17.251870 env[1311]: 2025-05-08 00:48:17.206 [WARNING][5992] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0", GenerateName:"calico-apiserver-fd97795bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"c5c7ed4b-9783-4fe2-b85a-8452a359cb74", ResourceVersion:"1235", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"fd97795bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f", Pod:"calico-apiserver-fd97795bb-tzgmm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calidaae6f6c0e9", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:48:17.251870 env[1311]: 2025-05-08 00:48:17.206 [INFO][5992] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:48:17.251870 env[1311]: 2025-05-08 00:48:17.206 [INFO][5992] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" iface="eth0" netns="" May 8 00:48:17.251870 env[1311]: 2025-05-08 00:48:17.206 [INFO][5992] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:48:17.251870 env[1311]: 2025-05-08 00:48:17.206 [INFO][5992] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:48:17.251870 env[1311]: 2025-05-08 00:48:17.239 [INFO][6001] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" HandleID="k8s-pod-network.cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" Workload="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:48:17.251870 env[1311]: 2025-05-08 00:48:17.240 [INFO][6001] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:48:17.251870 env[1311]: 2025-05-08 00:48:17.240 [INFO][6001] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:48:17.251870 env[1311]: 2025-05-08 00:48:17.247 [WARNING][6001] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" HandleID="k8s-pod-network.cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" Workload="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:48:17.251870 env[1311]: 2025-05-08 00:48:17.247 [INFO][6001] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" HandleID="k8s-pod-network.cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" Workload="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:48:17.251870 env[1311]: 2025-05-08 00:48:17.248 [INFO][6001] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:48:17.251870 env[1311]: 2025-05-08 00:48:17.250 [INFO][5992] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:48:17.252648 env[1311]: time="2025-05-08T00:48:17.251915808Z" level=info msg="TearDown network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\" successfully" May 8 00:48:17.252648 env[1311]: time="2025-05-08T00:48:17.251956875Z" level=info msg="StopPodSandbox for \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\" returns successfully" May 8 00:48:17.252698 env[1311]: time="2025-05-08T00:48:17.252652638Z" level=info msg="RemovePodSandbox for \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\"" May 8 00:48:17.252729 env[1311]: time="2025-05-08T00:48:17.252691572Z" level=info msg="Forcibly stopping sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\"" May 8 00:48:17.326146 env[1311]: 2025-05-08 00:48:17.289 [WARNING][6023] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0", GenerateName:"calico-apiserver-fd97795bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"c5c7ed4b-9783-4fe2-b85a-8452a359cb74", ResourceVersion:"1235", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"fd97795bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"3397f9148e1d8b2652321d60a669c8b6e7b27623990bcc659e8144318b5b2a4f", Pod:"calico-apiserver-fd97795bb-tzgmm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calidaae6f6c0e9", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:48:17.326146 env[1311]: 2025-05-08 00:48:17.289 [INFO][6023] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:48:17.326146 env[1311]: 2025-05-08 00:48:17.289 [INFO][6023] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" iface="eth0" netns="" May 8 00:48:17.326146 env[1311]: 2025-05-08 00:48:17.289 [INFO][6023] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:48:17.326146 env[1311]: 2025-05-08 00:48:17.289 [INFO][6023] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:48:17.326146 env[1311]: 2025-05-08 00:48:17.315 [INFO][6031] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" HandleID="k8s-pod-network.cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" Workload="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:48:17.326146 env[1311]: 2025-05-08 00:48:17.315 [INFO][6031] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:48:17.326146 env[1311]: 2025-05-08 00:48:17.315 [INFO][6031] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:48:17.326146 env[1311]: 2025-05-08 00:48:17.321 [WARNING][6031] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" HandleID="k8s-pod-network.cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" Workload="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:48:17.326146 env[1311]: 2025-05-08 00:48:17.321 [INFO][6031] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" HandleID="k8s-pod-network.cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" Workload="localhost-k8s-calico--apiserver--fd97795bb--tzgmm-eth0" May 8 00:48:17.326146 env[1311]: 2025-05-08 00:48:17.322 [INFO][6031] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:48:17.326146 env[1311]: 2025-05-08 00:48:17.324 [INFO][6023] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd" May 8 00:48:17.327495 env[1311]: time="2025-05-08T00:48:17.326092701Z" level=info msg="TearDown network for sandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\" successfully" May 8 00:48:17.331652 env[1311]: time="2025-05-08T00:48:17.331595777Z" level=info msg="RemovePodSandbox \"cd01e992a0d272d65a715b880e2f43a8ac997bdb6f2c65b4c52fead20a1874bd\" returns successfully" May 8 00:48:17.332165 env[1311]: time="2025-05-08T00:48:17.332137340Z" level=info msg="StopPodSandbox for \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\"" May 8 00:48:17.402173 env[1311]: 2025-05-08 00:48:17.367 [WARNING][6053] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--2s6tt-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"863a2a82-a51b-4afc-a7be-990ee722b050", ResourceVersion:"1353", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b7b4b9d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565", Pod:"csi-node-driver-2s6tt", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali09845bb6542", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:48:17.402173 env[1311]: 2025-05-08 00:48:17.368 [INFO][6053] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:48:17.402173 env[1311]: 2025-05-08 00:48:17.368 [INFO][6053] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" iface="eth0" netns="" May 8 00:48:17.402173 env[1311]: 2025-05-08 00:48:17.368 [INFO][6053] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:48:17.402173 env[1311]: 2025-05-08 00:48:17.368 [INFO][6053] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:48:17.402173 env[1311]: 2025-05-08 00:48:17.389 [INFO][6062] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" HandleID="k8s-pod-network.b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" Workload="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:48:17.402173 env[1311]: 2025-05-08 00:48:17.389 [INFO][6062] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:48:17.402173 env[1311]: 2025-05-08 00:48:17.389 [INFO][6062] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:48:17.402173 env[1311]: 2025-05-08 00:48:17.395 [WARNING][6062] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" HandleID="k8s-pod-network.b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" Workload="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:48:17.402173 env[1311]: 2025-05-08 00:48:17.396 [INFO][6062] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" HandleID="k8s-pod-network.b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" Workload="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:48:17.402173 env[1311]: 2025-05-08 00:48:17.398 [INFO][6062] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:48:17.402173 env[1311]: 2025-05-08 00:48:17.400 [INFO][6053] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:48:17.403346 env[1311]: time="2025-05-08T00:48:17.403089105Z" level=info msg="TearDown network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\" successfully" May 8 00:48:17.403465 env[1311]: time="2025-05-08T00:48:17.403444626Z" level=info msg="StopPodSandbox for \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\" returns successfully" May 8 00:48:17.404211 env[1311]: time="2025-05-08T00:48:17.404162090Z" level=info msg="RemovePodSandbox for \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\"" May 8 00:48:17.404413 env[1311]: time="2025-05-08T00:48:17.404212395Z" level=info msg="Forcibly stopping sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\"" May 8 00:48:17.472242 env[1311]: 2025-05-08 00:48:17.441 [WARNING][6085] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--2s6tt-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"863a2a82-a51b-4afc-a7be-990ee722b050", ResourceVersion:"1353", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b7b4b9d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"c775eb658b21772d3c7eb30700344a76a6dee3a42da7efa0ed4c9e966b3ae565", Pod:"csi-node-driver-2s6tt", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali09845bb6542", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:48:17.472242 env[1311]: 2025-05-08 00:48:17.441 [INFO][6085] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:48:17.472242 env[1311]: 2025-05-08 00:48:17.441 [INFO][6085] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" iface="eth0" netns="" May 8 00:48:17.472242 env[1311]: 2025-05-08 00:48:17.441 [INFO][6085] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:48:17.472242 env[1311]: 2025-05-08 00:48:17.441 [INFO][6085] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:48:17.472242 env[1311]: 2025-05-08 00:48:17.461 [INFO][6093] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" HandleID="k8s-pod-network.b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" Workload="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:48:17.472242 env[1311]: 2025-05-08 00:48:17.461 [INFO][6093] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:48:17.472242 env[1311]: 2025-05-08 00:48:17.462 [INFO][6093] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:48:17.472242 env[1311]: 2025-05-08 00:48:17.468 [WARNING][6093] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" HandleID="k8s-pod-network.b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" Workload="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:48:17.472242 env[1311]: 2025-05-08 00:48:17.468 [INFO][6093] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" HandleID="k8s-pod-network.b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" Workload="localhost-k8s-csi--node--driver--2s6tt-eth0" May 8 00:48:17.472242 env[1311]: 2025-05-08 00:48:17.469 [INFO][6093] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:48:17.472242 env[1311]: 2025-05-08 00:48:17.470 [INFO][6085] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce" May 8 00:48:17.472854 env[1311]: time="2025-05-08T00:48:17.472276000Z" level=info msg="TearDown network for sandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\" successfully" May 8 00:48:17.524160 env[1311]: time="2025-05-08T00:48:17.524103751Z" level=info msg="RemovePodSandbox \"b1efef17f6486c2adb6583901096099625b5235f609ea5e7d806d650d767ebce\" returns successfully" May 8 00:48:17.524388 env[1311]: time="2025-05-08T00:48:17.524363250Z" level=info msg="StopPodSandbox for \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\"" May 8 00:48:17.594749 env[1311]: 2025-05-08 00:48:17.563 [WARNING][6116] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0", GenerateName:"calico-apiserver-fd97795bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"ca6c7950-e2c3-4481-994b-8bd117426f80", ResourceVersion:"1248", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"fd97795bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4", Pod:"calico-apiserver-fd97795bb-zmwqb", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calidaa7669be5f", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:48:17.594749 env[1311]: 2025-05-08 00:48:17.564 [INFO][6116] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:48:17.594749 env[1311]: 2025-05-08 00:48:17.564 [INFO][6116] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" iface="eth0" netns="" May 8 00:48:17.594749 env[1311]: 2025-05-08 00:48:17.564 [INFO][6116] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:48:17.594749 env[1311]: 2025-05-08 00:48:17.564 [INFO][6116] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:48:17.594749 env[1311]: 2025-05-08 00:48:17.583 [INFO][6124] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" HandleID="k8s-pod-network.95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" Workload="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:48:17.594749 env[1311]: 2025-05-08 00:48:17.583 [INFO][6124] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:48:17.594749 env[1311]: 2025-05-08 00:48:17.583 [INFO][6124] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:48:17.594749 env[1311]: 2025-05-08 00:48:17.588 [WARNING][6124] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" HandleID="k8s-pod-network.95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" Workload="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:48:17.594749 env[1311]: 2025-05-08 00:48:17.588 [INFO][6124] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" HandleID="k8s-pod-network.95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" Workload="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:48:17.594749 env[1311]: 2025-05-08 00:48:17.589 [INFO][6124] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:48:17.594749 env[1311]: 2025-05-08 00:48:17.591 [INFO][6116] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:48:17.595842 env[1311]: time="2025-05-08T00:48:17.595758905Z" level=info msg="TearDown network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\" successfully" May 8 00:48:17.596029 env[1311]: time="2025-05-08T00:48:17.595824808Z" level=info msg="StopPodSandbox for \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\" returns successfully" May 8 00:48:17.596571 env[1311]: time="2025-05-08T00:48:17.596537023Z" level=info msg="RemovePodSandbox for \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\"" May 8 00:48:17.596650 env[1311]: time="2025-05-08T00:48:17.596571388Z" level=info msg="Forcibly stopping sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\"" May 8 00:48:17.657944 env[1311]: 2025-05-08 00:48:17.628 [WARNING][6148] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0", GenerateName:"calico-apiserver-fd97795bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"ca6c7950-e2c3-4481-994b-8bd117426f80", ResourceVersion:"1248", Generation:0, CreationTimestamp:time.Date(2025, time.May, 8, 0, 46, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"fd97795bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"ffcd4135e18193a43b139fb3413fa0b88c2688254540432d46b5504bf67b24c4", Pod:"calico-apiserver-fd97795bb-zmwqb", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calidaa7669be5f", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 8 00:48:17.657944 env[1311]: 2025-05-08 00:48:17.628 [INFO][6148] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:48:17.657944 env[1311]: 2025-05-08 00:48:17.628 [INFO][6148] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" iface="eth0" netns="" May 8 00:48:17.657944 env[1311]: 2025-05-08 00:48:17.628 [INFO][6148] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:48:17.657944 env[1311]: 2025-05-08 00:48:17.628 [INFO][6148] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:48:17.657944 env[1311]: 2025-05-08 00:48:17.648 [INFO][6157] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" HandleID="k8s-pod-network.95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" Workload="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:48:17.657944 env[1311]: 2025-05-08 00:48:17.648 [INFO][6157] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 8 00:48:17.657944 env[1311]: 2025-05-08 00:48:17.648 [INFO][6157] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 8 00:48:17.657944 env[1311]: 2025-05-08 00:48:17.653 [WARNING][6157] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" HandleID="k8s-pod-network.95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" Workload="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:48:17.657944 env[1311]: 2025-05-08 00:48:17.653 [INFO][6157] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" HandleID="k8s-pod-network.95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" Workload="localhost-k8s-calico--apiserver--fd97795bb--zmwqb-eth0" May 8 00:48:17.657944 env[1311]: 2025-05-08 00:48:17.654 [INFO][6157] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 8 00:48:17.657944 env[1311]: 2025-05-08 00:48:17.656 [INFO][6148] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b" May 8 00:48:17.658453 env[1311]: time="2025-05-08T00:48:17.657984570Z" level=info msg="TearDown network for sandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\" successfully" May 8 00:48:17.805908 env[1311]: time="2025-05-08T00:48:17.805833915Z" level=info msg="RemovePodSandbox \"95a73d4a18c3e05ba15eb4cbb513f75f25ea87b0b70bee4b9380b4c0a781562b\" returns successfully" May 8 00:48:19.454200 systemd[1]: Started sshd@27-10.0.0.82:22-10.0.0.1:37212.service. May 8 00:48:19.453000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@27-10.0.0.82:22-10.0.0.1:37212 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:19.459133 kernel: kauditd_printk_skb: 1 callbacks suppressed May 8 00:48:19.459215 kernel: audit: type=1130 audit(1746665299.453:590): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@27-10.0.0.82:22-10.0.0.1:37212 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:19.501000 audit[6164]: USER_ACCT pid=6164 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:19.502411 sshd[6164]: Accepted publickey for core from 10.0.0.1 port 37212 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:48:19.504224 sshd[6164]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:48:19.502000 audit[6164]: CRED_ACQ pid=6164 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:19.512130 kernel: audit: type=1101 audit(1746665299.501:591): pid=6164 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:19.512190 kernel: audit: type=1103 audit(1746665299.502:592): pid=6164 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:19.512211 kernel: audit: type=1006 audit(1746665299.502:593): pid=6164 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=28 res=1 May 8 00:48:19.510306 systemd-logind[1291]: New session 28 of user core. May 8 00:48:19.502000 audit[6164]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7fff1598cbb0 a2=3 a3=0 items=0 ppid=1 pid=6164 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=28 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:48:19.511023 systemd[1]: Started session-28.scope. May 8 00:48:19.516196 kernel: audit: type=1300 audit(1746665299.502:593): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7fff1598cbb0 a2=3 a3=0 items=0 ppid=1 pid=6164 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=28 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:48:19.502000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:48:19.517694 kernel: audit: type=1327 audit(1746665299.502:593): proctitle=737368643A20636F7265205B707269765D May 8 00:48:19.522098 kernel: audit: type=1105 audit(1746665299.515:594): pid=6164 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:19.515000 audit[6164]: USER_START pid=6164 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:19.517000 audit[6168]: CRED_ACQ pid=6168 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:19.526130 kernel: audit: type=1103 audit(1746665299.517:595): pid=6168 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:19.718865 sshd[6164]: pam_unix(sshd:session): session closed for user core May 8 00:48:19.718000 audit[6164]: USER_END pid=6164 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:19.721702 systemd[1]: sshd@27-10.0.0.82:22-10.0.0.1:37212.service: Deactivated successfully. May 8 00:48:19.722762 systemd[1]: session-28.scope: Deactivated successfully. May 8 00:48:19.722824 systemd-logind[1291]: Session 28 logged out. Waiting for processes to exit. May 8 00:48:19.723908 systemd-logind[1291]: Removed session 28. May 8 00:48:19.719000 audit[6164]: CRED_DISP pid=6164 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:19.729360 kernel: audit: type=1106 audit(1746665299.718:596): pid=6164 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:19.729411 kernel: audit: type=1104 audit(1746665299.719:597): pid=6164 uid=0 auid=500 ses=28 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:19.720000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@27-10.0.0.82:22-10.0.0.1:37212 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:24.722246 systemd[1]: Started sshd@28-10.0.0.82:22-10.0.0.1:37224.service. May 8 00:48:24.721000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@28-10.0.0.82:22-10.0.0.1:37224 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:24.723538 kernel: kauditd_printk_skb: 1 callbacks suppressed May 8 00:48:24.723667 kernel: audit: type=1130 audit(1746665304.721:599): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@28-10.0.0.82:22-10.0.0.1:37224 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:24.766000 audit[6180]: USER_ACCT pid=6180 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:24.768350 sshd[6180]: Accepted publickey for core from 10.0.0.1 port 37224 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:48:24.769477 sshd[6180]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:48:24.766000 audit[6180]: CRED_ACQ pid=6180 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:24.775831 kernel: audit: type=1101 audit(1746665304.766:600): pid=6180 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:24.775958 kernel: audit: type=1103 audit(1746665304.766:601): pid=6180 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:24.775977 kernel: audit: type=1006 audit(1746665304.766:602): pid=6180 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=29 res=1 May 8 00:48:24.776278 systemd-logind[1291]: New session 29 of user core. May 8 00:48:24.777230 systemd[1]: Started session-29.scope. May 8 00:48:24.766000 audit[6180]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffee676c2a0 a2=3 a3=0 items=0 ppid=1 pid=6180 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=29 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:48:24.783607 kernel: audit: type=1300 audit(1746665304.766:602): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7ffee676c2a0 a2=3 a3=0 items=0 ppid=1 pid=6180 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=29 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:48:24.783764 kernel: audit: type=1327 audit(1746665304.766:602): proctitle=737368643A20636F7265205B707269765D May 8 00:48:24.766000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:48:24.781000 audit[6180]: USER_START pid=6180 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:24.788478 kernel: audit: type=1105 audit(1746665304.781:603): pid=6180 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:24.788555 kernel: audit: type=1103 audit(1746665304.782:604): pid=6184 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:24.782000 audit[6184]: CRED_ACQ pid=6184 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:24.923011 systemd[1]: run-containerd-runc-k8s.io-17cd06d969354194d48f416bcae0a8966b438de64e0d4fd9c9a7ad8d6dc8bdf7-runc.PM3xMw.mount: Deactivated successfully. May 8 00:48:24.926843 sshd[6180]: pam_unix(sshd:session): session closed for user core May 8 00:48:24.926000 audit[6180]: USER_END pid=6180 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:24.926000 audit[6180]: CRED_DISP pid=6180 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:24.932735 systemd[1]: sshd@28-10.0.0.82:22-10.0.0.1:37224.service: Deactivated successfully. May 8 00:48:24.933915 systemd[1]: session-29.scope: Deactivated successfully. May 8 00:48:24.934526 systemd-logind[1291]: Session 29 logged out. Waiting for processes to exit. May 8 00:48:24.935399 systemd-logind[1291]: Removed session 29. May 8 00:48:24.936224 kernel: audit: type=1106 audit(1746665304.926:605): pid=6180 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:24.936290 kernel: audit: type=1104 audit(1746665304.926:606): pid=6180 uid=0 auid=500 ses=29 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:24.931000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@28-10.0.0.82:22-10.0.0.1:37224 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:29.929000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@29-10.0.0.82:22-10.0.0.1:59592 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:29.930897 systemd[1]: Started sshd@29-10.0.0.82:22-10.0.0.1:59592.service. May 8 00:48:29.932510 kernel: kauditd_printk_skb: 1 callbacks suppressed May 8 00:48:29.932572 kernel: audit: type=1130 audit(1746665309.929:608): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@29-10.0.0.82:22-10.0.0.1:59592 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 8 00:48:29.976000 audit[6223]: USER_ACCT pid=6223 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:29.977638 sshd[6223]: Accepted publickey for core from 10.0.0.1 port 59592 ssh2: RSA SHA256:1LBxu83eHkdm4X8dsk4zPTne32Wp9pee2vrXUZ4T9Dg May 8 00:48:29.980189 sshd[6223]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 8 00:48:29.978000 audit[6223]: CRED_ACQ pid=6223 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:29.983881 systemd-logind[1291]: New session 30 of user core. May 8 00:48:29.984682 systemd[1]: Started session-30.scope. May 8 00:48:29.987591 kernel: audit: type=1101 audit(1746665309.976:609): pid=6223 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_faillock,pam_permit acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:29.987667 kernel: audit: type=1103 audit(1746665309.978:610): pid=6223 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:29.990786 kernel: audit: type=1006 audit(1746665309.978:611): pid=6223 uid=0 subj=system_u:system_r:kernel_t:s0 old-auid=4294967295 auid=500 tty=(none) old-ses=4294967295 ses=30 res=1 May 8 00:48:29.990851 kernel: audit: type=1300 audit(1746665309.978:611): arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7fffb1f9a840 a2=3 a3=0 items=0 ppid=1 pid=6223 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=30 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:48:29.978000 audit[6223]: SYSCALL arch=c000003e syscall=1 success=yes exit=3 a0=5 a1=7fffb1f9a840 a2=3 a3=0 items=0 ppid=1 pid=6223 auid=500 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=30 comm="sshd" exe="/usr/sbin/sshd" subj=system_u:system_r:kernel_t:s0 key=(null) May 8 00:48:29.978000 audit: PROCTITLE proctitle=737368643A20636F7265205B707269765D May 8 00:48:29.997922 kernel: audit: type=1327 audit(1746665309.978:611): proctitle=737368643A20636F7265205B707269765D May 8 00:48:29.997970 kernel: audit: type=1105 audit(1746665309.988:612): pid=6223 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:29.988000 audit[6223]: USER_START pid=6223 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_open grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:29.990000 audit[6226]: CRED_ACQ pid=6226 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:30.007322 kernel: audit: type=1103 audit(1746665309.990:613): pid=6226 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:30.106159 sshd[6223]: pam_unix(sshd:session): session closed for user core May 8 00:48:30.106000 audit[6223]: USER_END pid=6223 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:30.108750 systemd[1]: sshd@29-10.0.0.82:22-10.0.0.1:59592.service: Deactivated successfully. May 8 00:48:30.109691 systemd[1]: session-30.scope: Deactivated successfully. May 8 00:48:30.106000 audit[6223]: CRED_DISP pid=6223 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:30.113846 systemd-logind[1291]: Session 30 logged out. Waiting for processes to exit. May 8 00:48:30.114656 systemd-logind[1291]: Removed session 30. May 8 00:48:30.116935 kernel: audit: type=1106 audit(1746665310.106:614): pid=6223 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:session_close grantors=pam_loginuid,pam_env,pam_lastlog,pam_limits,pam_env,pam_unix,pam_permit,pam_systemd,pam_mail acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:30.117015 kernel: audit: type=1104 audit(1746665310.106:615): pid=6223 uid=0 auid=500 ses=30 subj=system_u:system_r:kernel_t:s0 msg='op=PAM:setcred grantors=pam_env,pam_faillock,pam_unix acct="core" exe="/usr/sbin/sshd" hostname=10.0.0.1 addr=10.0.0.1 terminal=ssh res=success' May 8 00:48:30.107000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=sshd@29-10.0.0.82:22-10.0.0.1:59592 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'