Jul 7 00:05:09.915577 kernel: Linux version 6.6.95-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Sun Jul 6 22:23:50 -00 2025 Jul 7 00:05:09.915596 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=akamai verity.usrhash=65c65ff9d50198f0ae5c37458dc3ff85c6a690e7aa124bb306a2f4c63a54d876 Jul 7 00:05:09.915605 kernel: BIOS-provided physical RAM map: Jul 7 00:05:09.915610 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009f7ff] usable Jul 7 00:05:09.915615 kernel: BIOS-e820: [mem 0x000000000009f800-0x000000000009ffff] reserved Jul 7 00:05:09.915624 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Jul 7 00:05:09.915630 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdcfff] usable Jul 7 00:05:09.915636 kernel: BIOS-e820: [mem 0x000000007ffdd000-0x000000007fffffff] reserved Jul 7 00:05:09.915641 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Jul 7 00:05:09.915646 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Jul 7 00:05:09.915652 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Jul 7 00:05:09.915657 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Jul 7 00:05:09.915663 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000017fffffff] usable Jul 7 00:05:09.915668 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Jul 7 00:05:09.915678 kernel: NX (Execute Disable) protection: active Jul 7 00:05:09.915684 kernel: APIC: Static calls initialized Jul 7 00:05:09.915689 kernel: SMBIOS 2.8 present. Jul 7 00:05:09.915695 kernel: DMI: Linode Compute Instance, BIOS Not Specified Jul 7 00:05:09.915701 kernel: Hypervisor detected: KVM Jul 7 00:05:09.915709 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Jul 7 00:05:09.915715 kernel: kvm-clock: using sched offset of 4506436040 cycles Jul 7 00:05:09.915721 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Jul 7 00:05:09.915727 kernel: tsc: Detected 2000.000 MHz processor Jul 7 00:05:09.915733 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jul 7 00:05:09.915740 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jul 7 00:05:09.915746 kernel: last_pfn = 0x180000 max_arch_pfn = 0x400000000 Jul 7 00:05:09.915752 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Jul 7 00:05:09.915758 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jul 7 00:05:09.915766 kernel: last_pfn = 0x7ffdd max_arch_pfn = 0x400000000 Jul 7 00:05:09.915772 kernel: Using GB pages for direct mapping Jul 7 00:05:09.915778 kernel: ACPI: Early table checksum verification disabled Jul 7 00:05:09.915784 kernel: ACPI: RSDP 0x00000000000F51B0 000014 (v00 BOCHS ) Jul 7 00:05:09.915790 kernel: ACPI: RSDT 0x000000007FFE2307 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jul 7 00:05:09.915796 kernel: ACPI: FACP 0x000000007FFE20F7 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Jul 7 00:05:09.915802 kernel: ACPI: DSDT 0x000000007FFE0040 0020B7 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jul 7 00:05:09.915807 kernel: ACPI: FACS 0x000000007FFE0000 000040 Jul 7 00:05:09.915813 kernel: ACPI: APIC 0x000000007FFE21EB 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jul 7 00:05:09.915822 kernel: ACPI: HPET 0x000000007FFE226B 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jul 7 00:05:09.915828 kernel: ACPI: MCFG 0x000000007FFE22A3 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Jul 7 00:05:09.915834 kernel: ACPI: WAET 0x000000007FFE22DF 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jul 7 00:05:09.915843 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe20f7-0x7ffe21ea] Jul 7 00:05:09.915850 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe20f6] Jul 7 00:05:09.915856 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Jul 7 00:05:09.915862 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe21eb-0x7ffe226a] Jul 7 00:05:09.915871 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe226b-0x7ffe22a2] Jul 7 00:05:09.915878 kernel: ACPI: Reserving MCFG table memory at [mem 0x7ffe22a3-0x7ffe22de] Jul 7 00:05:09.915884 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe22df-0x7ffe2306] Jul 7 00:05:09.915890 kernel: No NUMA configuration found Jul 7 00:05:09.915896 kernel: Faking a node at [mem 0x0000000000000000-0x000000017fffffff] Jul 7 00:05:09.915902 kernel: NODE_DATA(0) allocated [mem 0x17fffa000-0x17fffffff] Jul 7 00:05:09.915908 kernel: Zone ranges: Jul 7 00:05:09.915914 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jul 7 00:05:09.915923 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Jul 7 00:05:09.915929 kernel: Normal [mem 0x0000000100000000-0x000000017fffffff] Jul 7 00:05:09.915935 kernel: Movable zone start for each node Jul 7 00:05:09.915941 kernel: Early memory node ranges Jul 7 00:05:09.915948 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Jul 7 00:05:09.915954 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdcfff] Jul 7 00:05:09.915960 kernel: node 0: [mem 0x0000000100000000-0x000000017fffffff] Jul 7 00:05:09.915966 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000017fffffff] Jul 7 00:05:09.915972 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jul 7 00:05:09.915981 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Jul 7 00:05:09.915987 kernel: On node 0, zone Normal: 35 pages in unavailable ranges Jul 7 00:05:09.915993 kernel: ACPI: PM-Timer IO Port: 0x608 Jul 7 00:05:09.915999 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Jul 7 00:05:09.916005 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Jul 7 00:05:09.916011 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Jul 7 00:05:09.916018 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Jul 7 00:05:09.916024 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Jul 7 00:05:09.916030 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Jul 7 00:05:09.916039 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Jul 7 00:05:09.916045 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jul 7 00:05:09.916051 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Jul 7 00:05:09.916057 kernel: TSC deadline timer available Jul 7 00:05:09.916063 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Jul 7 00:05:09.916069 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Jul 7 00:05:09.916075 kernel: kvm-guest: KVM setup pv remote TLB flush Jul 7 00:05:09.916081 kernel: kvm-guest: setup PV sched yield Jul 7 00:05:09.916087 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Jul 7 00:05:09.916096 kernel: Booting paravirtualized kernel on KVM Jul 7 00:05:09.916102 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jul 7 00:05:09.916109 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Jul 7 00:05:09.916115 kernel: percpu: Embedded 58 pages/cpu s197096 r8192 d32280 u1048576 Jul 7 00:05:09.916121 kernel: pcpu-alloc: s197096 r8192 d32280 u1048576 alloc=1*2097152 Jul 7 00:05:09.916127 kernel: pcpu-alloc: [0] 0 1 Jul 7 00:05:09.916133 kernel: kvm-guest: PV spinlocks enabled Jul 7 00:05:09.916139 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Jul 7 00:05:09.916146 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=akamai verity.usrhash=65c65ff9d50198f0ae5c37458dc3ff85c6a690e7aa124bb306a2f4c63a54d876 Jul 7 00:05:09.916155 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jul 7 00:05:09.916161 kernel: random: crng init done Jul 7 00:05:09.916167 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jul 7 00:05:09.916174 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jul 7 00:05:09.916180 kernel: Fallback order for Node 0: 0 Jul 7 00:05:09.916186 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1031901 Jul 7 00:05:09.916192 kernel: Policy zone: Normal Jul 7 00:05:09.916198 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jul 7 00:05:09.916207 kernel: software IO TLB: area num 2. Jul 7 00:05:09.916213 kernel: Memory: 3966212K/4193772K available (12288K kernel code, 2295K rwdata, 22748K rodata, 42868K init, 2324K bss, 227300K reserved, 0K cma-reserved) Jul 7 00:05:09.916219 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Jul 7 00:05:09.916225 kernel: ftrace: allocating 37966 entries in 149 pages Jul 7 00:05:09.916232 kernel: ftrace: allocated 149 pages with 4 groups Jul 7 00:05:09.916238 kernel: Dynamic Preempt: voluntary Jul 7 00:05:09.916244 kernel: rcu: Preemptible hierarchical RCU implementation. Jul 7 00:05:09.916268 kernel: rcu: RCU event tracing is enabled. Jul 7 00:05:09.916275 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Jul 7 00:05:09.916284 kernel: Trampoline variant of Tasks RCU enabled. Jul 7 00:05:09.916291 kernel: Rude variant of Tasks RCU enabled. Jul 7 00:05:09.916297 kernel: Tracing variant of Tasks RCU enabled. Jul 7 00:05:09.916303 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jul 7 00:05:09.916309 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Jul 7 00:05:09.916315 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Jul 7 00:05:09.916321 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jul 7 00:05:09.916327 kernel: Console: colour VGA+ 80x25 Jul 7 00:05:09.916334 kernel: printk: console [tty0] enabled Jul 7 00:05:09.916341 kernel: printk: console [ttyS0] enabled Jul 7 00:05:09.916350 kernel: ACPI: Core revision 20230628 Jul 7 00:05:09.916357 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Jul 7 00:05:09.916363 kernel: APIC: Switch to symmetric I/O mode setup Jul 7 00:05:09.916377 kernel: x2apic enabled Jul 7 00:05:09.916387 kernel: APIC: Switched APIC routing to: physical x2apic Jul 7 00:05:09.916393 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Jul 7 00:05:09.916400 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Jul 7 00:05:09.916406 kernel: kvm-guest: setup PV IPIs Jul 7 00:05:09.916413 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jul 7 00:05:09.916419 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Jul 7 00:05:09.916426 kernel: Calibrating delay loop (skipped) preset value.. 4000.00 BogoMIPS (lpj=2000000) Jul 7 00:05:09.916435 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Jul 7 00:05:09.916442 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Jul 7 00:05:09.916448 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Jul 7 00:05:09.916455 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jul 7 00:05:09.916461 kernel: Spectre V2 : Mitigation: Retpolines Jul 7 00:05:09.916470 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Jul 7 00:05:09.916477 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Jul 7 00:05:09.916483 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Jul 7 00:05:09.916490 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Jul 7 00:05:09.916496 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Jul 7 00:05:09.916503 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Jul 7 00:05:09.916510 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Jul 7 00:05:09.916516 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jul 7 00:05:09.916525 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jul 7 00:05:09.916532 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jul 7 00:05:09.916538 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Jul 7 00:05:09.916544 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jul 7 00:05:09.916551 kernel: x86/fpu: xstate_offset[9]: 832, xstate_sizes[9]: 8 Jul 7 00:05:09.916558 kernel: x86/fpu: Enabled xstate features 0x207, context size is 840 bytes, using 'compacted' format. Jul 7 00:05:09.916564 kernel: Freeing SMP alternatives memory: 32K Jul 7 00:05:09.916570 kernel: pid_max: default: 32768 minimum: 301 Jul 7 00:05:09.916577 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Jul 7 00:05:09.916586 kernel: landlock: Up and running. Jul 7 00:05:09.916592 kernel: SELinux: Initializing. Jul 7 00:05:09.916599 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jul 7 00:05:09.916605 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jul 7 00:05:09.916612 kernel: smpboot: CPU0: AMD EPYC 7713 64-Core Processor (family: 0x19, model: 0x1, stepping: 0x1) Jul 7 00:05:09.916618 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jul 7 00:05:09.916625 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jul 7 00:05:09.916631 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jul 7 00:05:09.916638 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Jul 7 00:05:09.916647 kernel: ... version: 0 Jul 7 00:05:09.916653 kernel: ... bit width: 48 Jul 7 00:05:09.916660 kernel: ... generic registers: 6 Jul 7 00:05:09.916666 kernel: ... value mask: 0000ffffffffffff Jul 7 00:05:09.916673 kernel: ... max period: 00007fffffffffff Jul 7 00:05:09.916679 kernel: ... fixed-purpose events: 0 Jul 7 00:05:09.916685 kernel: ... event mask: 000000000000003f Jul 7 00:05:09.916692 kernel: signal: max sigframe size: 3376 Jul 7 00:05:09.916698 kernel: rcu: Hierarchical SRCU implementation. Jul 7 00:05:09.916707 kernel: rcu: Max phase no-delay instances is 400. Jul 7 00:05:09.916714 kernel: smp: Bringing up secondary CPUs ... Jul 7 00:05:09.916720 kernel: smpboot: x86: Booting SMP configuration: Jul 7 00:05:09.916727 kernel: .... node #0, CPUs: #1 Jul 7 00:05:09.916733 kernel: smp: Brought up 1 node, 2 CPUs Jul 7 00:05:09.916740 kernel: smpboot: Max logical packages: 1 Jul 7 00:05:09.916746 kernel: smpboot: Total of 2 processors activated (8000.00 BogoMIPS) Jul 7 00:05:09.916753 kernel: devtmpfs: initialized Jul 7 00:05:09.916759 kernel: x86/mm: Memory block size: 128MB Jul 7 00:05:09.916765 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jul 7 00:05:09.916775 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Jul 7 00:05:09.916781 kernel: pinctrl core: initialized pinctrl subsystem Jul 7 00:05:09.916788 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jul 7 00:05:09.916794 kernel: audit: initializing netlink subsys (disabled) Jul 7 00:05:09.916801 kernel: audit: type=2000 audit(1751846709.061:1): state=initialized audit_enabled=0 res=1 Jul 7 00:05:09.916807 kernel: thermal_sys: Registered thermal governor 'step_wise' Jul 7 00:05:09.916813 kernel: thermal_sys: Registered thermal governor 'user_space' Jul 7 00:05:09.916820 kernel: cpuidle: using governor menu Jul 7 00:05:09.916826 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jul 7 00:05:09.916835 kernel: dca service started, version 1.12.1 Jul 7 00:05:09.916842 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Jul 7 00:05:09.916848 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Jul 7 00:05:09.916855 kernel: PCI: Using configuration type 1 for base access Jul 7 00:05:09.916862 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jul 7 00:05:09.916868 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jul 7 00:05:09.916874 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jul 7 00:05:09.916881 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jul 7 00:05:09.916890 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jul 7 00:05:09.916896 kernel: ACPI: Added _OSI(Module Device) Jul 7 00:05:09.916903 kernel: ACPI: Added _OSI(Processor Device) Jul 7 00:05:09.916909 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jul 7 00:05:09.916916 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jul 7 00:05:09.916922 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Jul 7 00:05:09.916928 kernel: ACPI: Interpreter enabled Jul 7 00:05:09.916935 kernel: ACPI: PM: (supports S0 S3 S5) Jul 7 00:05:09.916941 kernel: ACPI: Using IOAPIC for interrupt routing Jul 7 00:05:09.916948 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jul 7 00:05:09.916957 kernel: PCI: Using E820 reservations for host bridge windows Jul 7 00:05:09.916963 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Jul 7 00:05:09.916969 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jul 7 00:05:09.917151 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jul 7 00:05:09.919344 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Jul 7 00:05:09.919473 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Jul 7 00:05:09.919483 kernel: PCI host bridge to bus 0000:00 Jul 7 00:05:09.919609 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jul 7 00:05:09.919711 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Jul 7 00:05:09.919816 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jul 7 00:05:09.919919 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xafffffff window] Jul 7 00:05:09.920022 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Jul 7 00:05:09.920125 kernel: pci_bus 0000:00: root bus resource [mem 0x180000000-0x97fffffff window] Jul 7 00:05:09.920229 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jul 7 00:05:09.921207 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Jul 7 00:05:09.921375 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Jul 7 00:05:09.921489 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfd000000-0xfdffffff pref] Jul 7 00:05:09.921607 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfebd0000-0xfebd0fff] Jul 7 00:05:09.921721 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfebc0000-0xfebcffff pref] Jul 7 00:05:09.923662 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jul 7 00:05:09.923803 kernel: pci 0000:00:02.0: [1af4:1004] type 00 class 0x010000 Jul 7 00:05:09.923917 kernel: pci 0000:00:02.0: reg 0x10: [io 0xc000-0xc03f] Jul 7 00:05:09.924028 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xfebd1000-0xfebd1fff] Jul 7 00:05:09.924140 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfe000000-0xfe003fff 64bit pref] Jul 7 00:05:09.924303 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Jul 7 00:05:09.924421 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc040-0xc07f] Jul 7 00:05:09.924530 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebd2000-0xfebd2fff] Jul 7 00:05:09.924645 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe004000-0xfe007fff 64bit pref] Jul 7 00:05:09.924754 kernel: pci 0000:00:03.0: reg 0x30: [mem 0xfeb80000-0xfebbffff pref] Jul 7 00:05:09.924871 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Jul 7 00:05:09.924979 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Jul 7 00:05:09.925106 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Jul 7 00:05:09.925217 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc0c0-0xc0df] Jul 7 00:05:09.927833 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfebd3000-0xfebd3fff] Jul 7 00:05:09.927965 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Jul 7 00:05:09.928075 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Jul 7 00:05:09.928085 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Jul 7 00:05:09.928092 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Jul 7 00:05:09.928099 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jul 7 00:05:09.928106 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Jul 7 00:05:09.928112 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Jul 7 00:05:09.928124 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Jul 7 00:05:09.928130 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Jul 7 00:05:09.928137 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Jul 7 00:05:09.928143 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Jul 7 00:05:09.928149 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Jul 7 00:05:09.928156 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Jul 7 00:05:09.928162 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Jul 7 00:05:09.928169 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Jul 7 00:05:09.928175 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Jul 7 00:05:09.928184 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Jul 7 00:05:09.928191 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Jul 7 00:05:09.928197 kernel: iommu: Default domain type: Translated Jul 7 00:05:09.928204 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jul 7 00:05:09.928210 kernel: PCI: Using ACPI for IRQ routing Jul 7 00:05:09.928217 kernel: PCI: pci_cache_line_size set to 64 bytes Jul 7 00:05:09.928224 kernel: e820: reserve RAM buffer [mem 0x0009f800-0x0009ffff] Jul 7 00:05:09.928230 kernel: e820: reserve RAM buffer [mem 0x7ffdd000-0x7fffffff] Jul 7 00:05:09.930107 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Jul 7 00:05:09.930232 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Jul 7 00:05:09.930379 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jul 7 00:05:09.930390 kernel: vgaarb: loaded Jul 7 00:05:09.930397 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Jul 7 00:05:09.930404 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Jul 7 00:05:09.930410 kernel: clocksource: Switched to clocksource kvm-clock Jul 7 00:05:09.930417 kernel: VFS: Disk quotas dquot_6.6.0 Jul 7 00:05:09.930424 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jul 7 00:05:09.930430 kernel: pnp: PnP ACPI init Jul 7 00:05:09.930560 kernel: system 00:04: [mem 0xb0000000-0xbfffffff window] has been reserved Jul 7 00:05:09.930570 kernel: pnp: PnP ACPI: found 5 devices Jul 7 00:05:09.930577 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jul 7 00:05:09.930584 kernel: NET: Registered PF_INET protocol family Jul 7 00:05:09.930590 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jul 7 00:05:09.930597 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jul 7 00:05:09.930603 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jul 7 00:05:09.930610 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jul 7 00:05:09.930620 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jul 7 00:05:09.930626 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jul 7 00:05:09.930633 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jul 7 00:05:09.930639 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jul 7 00:05:09.930646 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jul 7 00:05:09.930652 kernel: NET: Registered PF_XDP protocol family Jul 7 00:05:09.930755 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Jul 7 00:05:09.930855 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Jul 7 00:05:09.930959 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Jul 7 00:05:09.931056 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xafffffff window] Jul 7 00:05:09.931154 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Jul 7 00:05:09.931299 kernel: pci_bus 0000:00: resource 9 [mem 0x180000000-0x97fffffff window] Jul 7 00:05:09.931310 kernel: PCI: CLS 0 bytes, default 64 Jul 7 00:05:09.931317 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Jul 7 00:05:09.931324 kernel: software IO TLB: mapped [mem 0x000000007bfdd000-0x000000007ffdd000] (64MB) Jul 7 00:05:09.931331 kernel: Initialise system trusted keyrings Jul 7 00:05:09.931337 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jul 7 00:05:09.931348 kernel: Key type asymmetric registered Jul 7 00:05:09.931355 kernel: Asymmetric key parser 'x509' registered Jul 7 00:05:09.931361 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Jul 7 00:05:09.931368 kernel: io scheduler mq-deadline registered Jul 7 00:05:09.931375 kernel: io scheduler kyber registered Jul 7 00:05:09.931381 kernel: io scheduler bfq registered Jul 7 00:05:09.931388 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jul 7 00:05:09.931394 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Jul 7 00:05:09.931401 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Jul 7 00:05:09.931411 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jul 7 00:05:09.931417 kernel: 00:02: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jul 7 00:05:09.931424 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Jul 7 00:05:09.931430 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jul 7 00:05:09.931437 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jul 7 00:05:09.931559 kernel: rtc_cmos 00:03: RTC can wake from S4 Jul 7 00:05:09.931569 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jul 7 00:05:09.931670 kernel: rtc_cmos 00:03: registered as rtc0 Jul 7 00:05:09.931777 kernel: rtc_cmos 00:03: setting system clock to 2025-07-07T00:05:09 UTC (1751846709) Jul 7 00:05:09.931880 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Jul 7 00:05:09.931889 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Jul 7 00:05:09.931895 kernel: NET: Registered PF_INET6 protocol family Jul 7 00:05:09.931902 kernel: Segment Routing with IPv6 Jul 7 00:05:09.931908 kernel: In-situ OAM (IOAM) with IPv6 Jul 7 00:05:09.931915 kernel: NET: Registered PF_PACKET protocol family Jul 7 00:05:09.931921 kernel: Key type dns_resolver registered Jul 7 00:05:09.931927 kernel: IPI shorthand broadcast: enabled Jul 7 00:05:09.931938 kernel: sched_clock: Marking stable (739004807, 230949109)->(1039360891, -69406975) Jul 7 00:05:09.931944 kernel: registered taskstats version 1 Jul 7 00:05:09.931951 kernel: Loading compiled-in X.509 certificates Jul 7 00:05:09.931958 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.95-flatcar: 6372c48ca52cc7f7bbee5675b604584c1c68ec5b' Jul 7 00:05:09.931964 kernel: Key type .fscrypt registered Jul 7 00:05:09.931970 kernel: Key type fscrypt-provisioning registered Jul 7 00:05:09.931977 kernel: ima: No TPM chip found, activating TPM-bypass! Jul 7 00:05:09.931984 kernel: ima: Allocated hash algorithm: sha1 Jul 7 00:05:09.931990 kernel: ima: No architecture policies found Jul 7 00:05:09.931999 kernel: clk: Disabling unused clocks Jul 7 00:05:09.932006 kernel: Freeing unused kernel image (initmem) memory: 42868K Jul 7 00:05:09.932012 kernel: Write protecting the kernel read-only data: 36864k Jul 7 00:05:09.932019 kernel: Freeing unused kernel image (rodata/data gap) memory: 1828K Jul 7 00:05:09.932025 kernel: Run /init as init process Jul 7 00:05:09.932032 kernel: with arguments: Jul 7 00:05:09.932038 kernel: /init Jul 7 00:05:09.932044 kernel: with environment: Jul 7 00:05:09.932051 kernel: HOME=/ Jul 7 00:05:09.932060 kernel: TERM=linux Jul 7 00:05:09.932066 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jul 7 00:05:09.932075 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jul 7 00:05:09.932083 systemd[1]: Detected virtualization kvm. Jul 7 00:05:09.932090 systemd[1]: Detected architecture x86-64. Jul 7 00:05:09.932097 systemd[1]: Running in initrd. Jul 7 00:05:09.932104 systemd[1]: No hostname configured, using default hostname. Jul 7 00:05:09.932113 systemd[1]: Hostname set to . Jul 7 00:05:09.932120 systemd[1]: Initializing machine ID from random generator. Jul 7 00:05:09.932127 systemd[1]: Queued start job for default target initrd.target. Jul 7 00:05:09.932134 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 7 00:05:09.932141 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 7 00:05:09.932163 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jul 7 00:05:09.932176 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jul 7 00:05:09.932184 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jul 7 00:05:09.932191 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jul 7 00:05:09.932199 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jul 7 00:05:09.932207 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jul 7 00:05:09.932214 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 7 00:05:09.932221 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jul 7 00:05:09.932231 systemd[1]: Reached target paths.target - Path Units. Jul 7 00:05:09.932238 systemd[1]: Reached target slices.target - Slice Units. Jul 7 00:05:09.932246 systemd[1]: Reached target swap.target - Swaps. Jul 7 00:05:09.933482 systemd[1]: Reached target timers.target - Timer Units. Jul 7 00:05:09.933491 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jul 7 00:05:09.933499 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jul 7 00:05:09.933506 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jul 7 00:05:09.933514 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jul 7 00:05:09.933526 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jul 7 00:05:09.933533 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jul 7 00:05:09.933540 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jul 7 00:05:09.933547 systemd[1]: Reached target sockets.target - Socket Units. Jul 7 00:05:09.933555 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jul 7 00:05:09.933562 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jul 7 00:05:09.933569 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jul 7 00:05:09.933577 systemd[1]: Starting systemd-fsck-usr.service... Jul 7 00:05:09.933584 systemd[1]: Starting systemd-journald.service - Journal Service... Jul 7 00:05:09.933594 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jul 7 00:05:09.933622 systemd-journald[177]: Collecting audit messages is disabled. Jul 7 00:05:09.933639 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 7 00:05:09.933647 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jul 7 00:05:09.933657 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jul 7 00:05:09.933664 systemd[1]: Finished systemd-fsck-usr.service. Jul 7 00:05:09.933674 systemd-journald[177]: Journal started Jul 7 00:05:09.933692 systemd-journald[177]: Runtime Journal (/run/log/journal/f45ce32254a84bb1a3346ddec5b181cf) is 8.0M, max 78.3M, 70.3M free. Jul 7 00:05:09.937453 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jul 7 00:05:09.932339 systemd-modules-load[178]: Inserted module 'overlay' Jul 7 00:05:09.987153 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jul 7 00:05:09.987173 kernel: Bridge firewalling registered Jul 7 00:05:09.957849 systemd-modules-load[178]: Inserted module 'br_netfilter' Jul 7 00:05:09.994420 systemd[1]: Started systemd-journald.service - Journal Service. Jul 7 00:05:09.994861 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jul 7 00:05:09.995684 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 7 00:05:09.999636 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jul 7 00:05:10.005396 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 7 00:05:10.012396 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 7 00:05:10.016420 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jul 7 00:05:10.024283 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jul 7 00:05:10.034951 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 7 00:05:10.049143 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 7 00:05:10.051492 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 7 00:05:10.059438 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jul 7 00:05:10.060632 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 7 00:05:10.072430 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jul 7 00:05:10.080958 dracut-cmdline[210]: dracut-dracut-053 Jul 7 00:05:10.085734 dracut-cmdline[210]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=akamai verity.usrhash=65c65ff9d50198f0ae5c37458dc3ff85c6a690e7aa124bb306a2f4c63a54d876 Jul 7 00:05:10.113764 systemd-resolved[212]: Positive Trust Anchors: Jul 7 00:05:10.114515 systemd-resolved[212]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 7 00:05:10.114545 systemd-resolved[212]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jul 7 00:05:10.117856 systemd-resolved[212]: Defaulting to hostname 'linux'. Jul 7 00:05:10.119115 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jul 7 00:05:10.120167 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jul 7 00:05:10.167313 kernel: SCSI subsystem initialized Jul 7 00:05:10.176274 kernel: Loading iSCSI transport class v2.0-870. Jul 7 00:05:10.188284 kernel: iscsi: registered transport (tcp) Jul 7 00:05:10.209293 kernel: iscsi: registered transport (qla4xxx) Jul 7 00:05:10.209329 kernel: QLogic iSCSI HBA Driver Jul 7 00:05:10.262742 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jul 7 00:05:10.268456 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jul 7 00:05:10.294424 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jul 7 00:05:10.294508 kernel: device-mapper: uevent: version 1.0.3 Jul 7 00:05:10.297294 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jul 7 00:05:10.339291 kernel: raid6: avx2x4 gen() 31712 MB/s Jul 7 00:05:10.357276 kernel: raid6: avx2x2 gen() 28085 MB/s Jul 7 00:05:10.375739 kernel: raid6: avx2x1 gen() 24966 MB/s Jul 7 00:05:10.375756 kernel: raid6: using algorithm avx2x4 gen() 31712 MB/s Jul 7 00:05:10.394815 kernel: raid6: .... xor() 5162 MB/s, rmw enabled Jul 7 00:05:10.394830 kernel: raid6: using avx2x2 recovery algorithm Jul 7 00:05:10.415284 kernel: xor: automatically using best checksumming function avx Jul 7 00:05:10.548319 kernel: Btrfs loaded, zoned=no, fsverity=no Jul 7 00:05:10.564873 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jul 7 00:05:10.573516 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 7 00:05:10.588005 systemd-udevd[395]: Using default interface naming scheme 'v255'. Jul 7 00:05:10.593753 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 7 00:05:10.600390 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jul 7 00:05:10.618560 dracut-pre-trigger[402]: rd.md=0: removing MD RAID activation Jul 7 00:05:10.652132 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jul 7 00:05:10.656377 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jul 7 00:05:10.714732 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jul 7 00:05:10.721405 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jul 7 00:05:10.733804 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jul 7 00:05:10.736759 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jul 7 00:05:10.740336 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 7 00:05:10.740884 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jul 7 00:05:10.750378 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jul 7 00:05:10.762169 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jul 7 00:05:10.788751 kernel: scsi host0: Virtio SCSI HBA Jul 7 00:05:10.929309 kernel: scsi 0:0:0:0: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Jul 7 00:05:10.929387 kernel: cryptd: max_cpu_qlen set to 1000 Jul 7 00:05:10.942641 kernel: libata version 3.00 loaded. Jul 7 00:05:10.979622 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jul 7 00:05:10.979829 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 7 00:05:10.981813 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 7 00:05:10.983204 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 7 00:05:10.983395 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 7 00:05:10.988189 kernel: AVX2 version of gcm_enc/dec engaged. Jul 7 00:05:10.987081 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jul 7 00:05:10.991329 kernel: AES CTR mode by8 optimization enabled Jul 7 00:05:10.995061 kernel: ahci 0000:00:1f.2: version 3.0 Jul 7 00:05:10.994516 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 7 00:05:11.002718 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Jul 7 00:05:11.010925 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Jul 7 00:05:11.011124 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Jul 7 00:05:11.017097 kernel: sd 0:0:0:0: Power-on or device reset occurred Jul 7 00:05:11.021188 kernel: sd 0:0:0:0: [sda] 167739392 512-byte logical blocks: (85.9 GB/80.0 GiB) Jul 7 00:05:11.021370 kernel: sd 0:0:0:0: [sda] Write Protect is off Jul 7 00:05:11.021512 kernel: sd 0:0:0:0: [sda] Mode Sense: 63 00 00 08 Jul 7 00:05:11.021648 kernel: sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Jul 7 00:05:11.021781 kernel: scsi host1: ahci Jul 7 00:05:11.022272 kernel: scsi host2: ahci Jul 7 00:05:11.023449 kernel: scsi host3: ahci Jul 7 00:05:11.025497 kernel: scsi host4: ahci Jul 7 00:05:11.026268 kernel: scsi host5: ahci Jul 7 00:05:11.030364 kernel: scsi host6: ahci Jul 7 00:05:11.030521 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jul 7 00:05:11.030538 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3100 irq 46 Jul 7 00:05:11.030548 kernel: GPT:9289727 != 167739391 Jul 7 00:05:11.030557 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3180 irq 46 Jul 7 00:05:11.030567 kernel: GPT:Alternate GPT header not at the end of the disk. Jul 7 00:05:11.030577 kernel: GPT:9289727 != 167739391 Jul 7 00:05:11.030586 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3200 irq 46 Jul 7 00:05:11.030595 kernel: GPT: Use GNU Parted to correct GPT errors. Jul 7 00:05:11.030604 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 7 00:05:11.030613 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3280 irq 46 Jul 7 00:05:11.035292 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Jul 7 00:05:11.035743 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3300 irq 46 Jul 7 00:05:11.052356 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3380 irq 46 Jul 7 00:05:11.106377 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 7 00:05:11.112447 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 7 00:05:11.127171 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 7 00:05:11.361822 kernel: ata3: SATA link down (SStatus 0 SControl 300) Jul 7 00:05:11.361889 kernel: ata1: SATA link down (SStatus 0 SControl 300) Jul 7 00:05:11.361902 kernel: ata2: SATA link down (SStatus 0 SControl 300) Jul 7 00:05:11.362265 kernel: ata4: SATA link down (SStatus 0 SControl 300) Jul 7 00:05:11.364275 kernel: ata5: SATA link down (SStatus 0 SControl 300) Jul 7 00:05:11.365269 kernel: ata6: SATA link down (SStatus 0 SControl 300) Jul 7 00:05:11.407634 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Jul 7 00:05:11.416614 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/sda6 scanned by (udev-worker) (441) Jul 7 00:05:11.416634 kernel: BTRFS: device fsid 01287863-c21f-4cbb-820d-bbae8208f32f devid 1 transid 34 /dev/sda3 scanned by (udev-worker) (439) Jul 7 00:05:11.418466 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Jul 7 00:05:11.429220 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Jul 7 00:05:11.434048 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Jul 7 00:05:11.434714 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Jul 7 00:05:11.448445 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jul 7 00:05:11.453797 disk-uuid[565]: Primary Header is updated. Jul 7 00:05:11.453797 disk-uuid[565]: Secondary Entries is updated. Jul 7 00:05:11.453797 disk-uuid[565]: Secondary Header is updated. Jul 7 00:05:11.459321 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 7 00:05:11.464273 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 7 00:05:12.468213 disk-uuid[566]: The operation has completed successfully. Jul 7 00:05:12.470302 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 7 00:05:12.512243 systemd[1]: disk-uuid.service: Deactivated successfully. Jul 7 00:05:12.512511 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jul 7 00:05:12.530377 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jul 7 00:05:12.534726 sh[580]: Success Jul 7 00:05:12.547332 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Jul 7 00:05:12.588195 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jul 7 00:05:12.602335 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jul 7 00:05:12.603667 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jul 7 00:05:12.622909 kernel: BTRFS info (device dm-0): first mount of filesystem 01287863-c21f-4cbb-820d-bbae8208f32f Jul 7 00:05:12.622936 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jul 7 00:05:12.622948 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jul 7 00:05:12.627197 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jul 7 00:05:12.627212 kernel: BTRFS info (device dm-0): using free space tree Jul 7 00:05:12.636321 kernel: BTRFS info (device dm-0): enabling ssd optimizations Jul 7 00:05:12.638280 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jul 7 00:05:12.639309 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jul 7 00:05:12.643358 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jul 7 00:05:12.646374 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jul 7 00:05:12.663578 kernel: BTRFS info (device sda6): first mount of filesystem 11f56a79-b29d-47db-ad8e-56effe5ac41b Jul 7 00:05:12.663603 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 7 00:05:12.663614 kernel: BTRFS info (device sda6): using free space tree Jul 7 00:05:12.669944 kernel: BTRFS info (device sda6): enabling ssd optimizations Jul 7 00:05:12.669967 kernel: BTRFS info (device sda6): auto enabling async discard Jul 7 00:05:12.681857 systemd[1]: mnt-oem.mount: Deactivated successfully. Jul 7 00:05:12.682821 kernel: BTRFS info (device sda6): last unmount of filesystem 11f56a79-b29d-47db-ad8e-56effe5ac41b Jul 7 00:05:12.689548 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jul 7 00:05:12.696103 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jul 7 00:05:12.762512 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jul 7 00:05:12.768480 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jul 7 00:05:12.779430 ignition[694]: Ignition 2.19.0 Jul 7 00:05:12.780177 ignition[694]: Stage: fetch-offline Jul 7 00:05:12.780730 ignition[694]: no configs at "/usr/lib/ignition/base.d" Jul 7 00:05:12.780741 ignition[694]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jul 7 00:05:12.780832 ignition[694]: parsed url from cmdline: "" Jul 7 00:05:12.780836 ignition[694]: no config URL provided Jul 7 00:05:12.780841 ignition[694]: reading system config file "/usr/lib/ignition/user.ign" Jul 7 00:05:12.784505 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jul 7 00:05:12.780850 ignition[694]: no config at "/usr/lib/ignition/user.ign" Jul 7 00:05:12.780856 ignition[694]: failed to fetch config: resource requires networking Jul 7 00:05:12.781026 ignition[694]: Ignition finished successfully Jul 7 00:05:12.797840 systemd-networkd[765]: lo: Link UP Jul 7 00:05:12.797854 systemd-networkd[765]: lo: Gained carrier Jul 7 00:05:12.799418 systemd-networkd[765]: Enumeration completed Jul 7 00:05:12.799827 systemd-networkd[765]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 7 00:05:12.799831 systemd-networkd[765]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jul 7 00:05:12.800924 systemd[1]: Started systemd-networkd.service - Network Configuration. Jul 7 00:05:12.802123 systemd-networkd[765]: eth0: Link UP Jul 7 00:05:12.802127 systemd-networkd[765]: eth0: Gained carrier Jul 7 00:05:12.802143 systemd-networkd[765]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 7 00:05:12.802472 systemd[1]: Reached target network.target - Network. Jul 7 00:05:12.812552 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Jul 7 00:05:12.826225 ignition[769]: Ignition 2.19.0 Jul 7 00:05:12.826239 ignition[769]: Stage: fetch Jul 7 00:05:12.826416 ignition[769]: no configs at "/usr/lib/ignition/base.d" Jul 7 00:05:12.826428 ignition[769]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jul 7 00:05:12.826506 ignition[769]: parsed url from cmdline: "" Jul 7 00:05:12.826510 ignition[769]: no config URL provided Jul 7 00:05:12.826515 ignition[769]: reading system config file "/usr/lib/ignition/user.ign" Jul 7 00:05:12.826523 ignition[769]: no config at "/usr/lib/ignition/user.ign" Jul 7 00:05:12.826543 ignition[769]: PUT http://169.254.169.254/v1/token: attempt #1 Jul 7 00:05:12.826697 ignition[769]: PUT error: Put "http://169.254.169.254/v1/token": dial tcp 169.254.169.254:80: connect: network is unreachable Jul 7 00:05:13.026916 ignition[769]: PUT http://169.254.169.254/v1/token: attempt #2 Jul 7 00:05:13.027087 ignition[769]: PUT error: Put "http://169.254.169.254/v1/token": dial tcp 169.254.169.254:80: connect: network is unreachable Jul 7 00:05:13.293331 systemd-networkd[765]: eth0: DHCPv4 address 172.234.31.125/24, gateway 172.234.31.1 acquired from 23.205.167.142 Jul 7 00:05:13.427713 ignition[769]: PUT http://169.254.169.254/v1/token: attempt #3 Jul 7 00:05:13.519316 ignition[769]: PUT result: OK Jul 7 00:05:13.519392 ignition[769]: GET http://169.254.169.254/v1/user-data: attempt #1 Jul 7 00:05:13.631918 ignition[769]: GET result: OK Jul 7 00:05:13.632085 ignition[769]: parsing config with SHA512: 794185a1569c43f9981f60e78de291d07968e5187c2a4e45ef7ca9597f09bff4ef6475b3edfaa89241ae8685e498f874a0df9f7bc251debfd8023f77d7c744c9 Jul 7 00:05:13.636605 unknown[769]: fetched base config from "system" Jul 7 00:05:13.636618 unknown[769]: fetched base config from "system" Jul 7 00:05:13.636870 ignition[769]: fetch: fetch complete Jul 7 00:05:13.636631 unknown[769]: fetched user config from "akamai" Jul 7 00:05:13.636875 ignition[769]: fetch: fetch passed Jul 7 00:05:13.636919 ignition[769]: Ignition finished successfully Jul 7 00:05:13.646736 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Jul 7 00:05:13.652435 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jul 7 00:05:13.670929 ignition[777]: Ignition 2.19.0 Jul 7 00:05:13.670943 ignition[777]: Stage: kargs Jul 7 00:05:13.671096 ignition[777]: no configs at "/usr/lib/ignition/base.d" Jul 7 00:05:13.671109 ignition[777]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jul 7 00:05:13.671824 ignition[777]: kargs: kargs passed Jul 7 00:05:13.673410 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jul 7 00:05:13.671865 ignition[777]: Ignition finished successfully Jul 7 00:05:13.679485 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jul 7 00:05:13.696730 ignition[783]: Ignition 2.19.0 Jul 7 00:05:13.696745 ignition[783]: Stage: disks Jul 7 00:05:13.696889 ignition[783]: no configs at "/usr/lib/ignition/base.d" Jul 7 00:05:13.701079 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jul 7 00:05:13.696903 ignition[783]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jul 7 00:05:13.702216 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jul 7 00:05:13.697808 ignition[783]: disks: disks passed Jul 7 00:05:13.703129 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jul 7 00:05:13.697846 ignition[783]: Ignition finished successfully Jul 7 00:05:13.704457 systemd[1]: Reached target local-fs.target - Local File Systems. Jul 7 00:05:13.727770 systemd[1]: Reached target sysinit.target - System Initialization. Jul 7 00:05:13.728999 systemd[1]: Reached target basic.target - Basic System. Jul 7 00:05:13.735461 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jul 7 00:05:13.752318 systemd-fsck[791]: ROOT: clean, 14/553520 files, 52654/553472 blocks Jul 7 00:05:13.754797 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jul 7 00:05:13.760365 systemd[1]: Mounting sysroot.mount - /sysroot... Jul 7 00:05:13.854280 kernel: EXT4-fs (sda9): mounted filesystem c3eefe20-4a42-420d-8034-4d5498275b2f r/w with ordered data mode. Quota mode: none. Jul 7 00:05:13.855431 systemd[1]: Mounted sysroot.mount - /sysroot. Jul 7 00:05:13.856808 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jul 7 00:05:13.874364 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jul 7 00:05:13.877536 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jul 7 00:05:13.879339 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jul 7 00:05:13.879398 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jul 7 00:05:13.879423 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jul 7 00:05:13.891318 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/sda6 scanned by mount (799) Jul 7 00:05:13.894053 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jul 7 00:05:13.902618 kernel: BTRFS info (device sda6): first mount of filesystem 11f56a79-b29d-47db-ad8e-56effe5ac41b Jul 7 00:05:13.902635 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 7 00:05:13.902646 kernel: BTRFS info (device sda6): using free space tree Jul 7 00:05:13.902656 kernel: BTRFS info (device sda6): enabling ssd optimizations Jul 7 00:05:13.902666 kernel: BTRFS info (device sda6): auto enabling async discard Jul 7 00:05:13.904286 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jul 7 00:05:13.910384 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jul 7 00:05:13.957340 initrd-setup-root[823]: cut: /sysroot/etc/passwd: No such file or directory Jul 7 00:05:13.964981 initrd-setup-root[830]: cut: /sysroot/etc/group: No such file or directory Jul 7 00:05:13.970662 initrd-setup-root[837]: cut: /sysroot/etc/shadow: No such file or directory Jul 7 00:05:13.976352 initrd-setup-root[844]: cut: /sysroot/etc/gshadow: No such file or directory Jul 7 00:05:14.080280 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jul 7 00:05:14.085410 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jul 7 00:05:14.088455 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jul 7 00:05:14.099287 kernel: BTRFS info (device sda6): last unmount of filesystem 11f56a79-b29d-47db-ad8e-56effe5ac41b Jul 7 00:05:14.099506 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jul 7 00:05:14.121983 ignition[913]: INFO : Ignition 2.19.0 Jul 7 00:05:14.122592 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jul 7 00:05:14.124869 ignition[913]: INFO : Stage: mount Jul 7 00:05:14.124869 ignition[913]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 7 00:05:14.124869 ignition[913]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jul 7 00:05:14.127264 ignition[913]: INFO : mount: mount passed Jul 7 00:05:14.127264 ignition[913]: INFO : Ignition finished successfully Jul 7 00:05:14.127043 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jul 7 00:05:14.133373 systemd[1]: Starting ignition-files.service - Ignition (files)... Jul 7 00:05:14.559631 systemd-networkd[765]: eth0: Gained IPv6LL Jul 7 00:05:14.862404 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jul 7 00:05:14.876294 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 scanned by mount (924) Jul 7 00:05:14.876330 kernel: BTRFS info (device sda6): first mount of filesystem 11f56a79-b29d-47db-ad8e-56effe5ac41b Jul 7 00:05:14.880641 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 7 00:05:14.880668 kernel: BTRFS info (device sda6): using free space tree Jul 7 00:05:14.887721 kernel: BTRFS info (device sda6): enabling ssd optimizations Jul 7 00:05:14.887742 kernel: BTRFS info (device sda6): auto enabling async discard Jul 7 00:05:14.891842 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jul 7 00:05:14.917025 ignition[941]: INFO : Ignition 2.19.0 Jul 7 00:05:14.917025 ignition[941]: INFO : Stage: files Jul 7 00:05:14.918454 ignition[941]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 7 00:05:14.918454 ignition[941]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jul 7 00:05:14.918454 ignition[941]: DEBUG : files: compiled without relabeling support, skipping Jul 7 00:05:14.923389 ignition[941]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jul 7 00:05:14.923389 ignition[941]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jul 7 00:05:14.923389 ignition[941]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jul 7 00:05:14.923389 ignition[941]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jul 7 00:05:14.926704 ignition[941]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jul 7 00:05:14.926704 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jul 7 00:05:14.926704 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Jul 7 00:05:14.923527 unknown[941]: wrote ssh authorized keys file for user: core Jul 7 00:05:15.273064 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jul 7 00:05:16.251853 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jul 7 00:05:16.251853 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-x86-64.raw: attempt #1 Jul 7 00:05:16.793557 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jul 7 00:05:17.091590 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Jul 7 00:05:17.091590 ignition[941]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(d): op(e): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(f): [started] setting preset to enabled for "prepare-helm.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(f): [finished] setting preset to enabled for "prepare-helm.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: createResultFile: createFiles: op(10): [started] writing file "/sysroot/etc/.ignition-result.json" Jul 7 00:05:17.094666 ignition[941]: INFO : files: createResultFile: createFiles: op(10): [finished] writing file "/sysroot/etc/.ignition-result.json" Jul 7 00:05:17.094666 ignition[941]: INFO : files: files passed Jul 7 00:05:17.094666 ignition[941]: INFO : Ignition finished successfully Jul 7 00:05:17.097728 systemd[1]: Finished ignition-files.service - Ignition (files). Jul 7 00:05:17.129504 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jul 7 00:05:17.135401 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jul 7 00:05:17.137118 systemd[1]: ignition-quench.service: Deactivated successfully. Jul 7 00:05:17.137230 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jul 7 00:05:17.152513 initrd-setup-root-after-ignition[969]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 7 00:05:17.152513 initrd-setup-root-after-ignition[969]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jul 7 00:05:17.154766 initrd-setup-root-after-ignition[973]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 7 00:05:17.156175 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jul 7 00:05:17.157016 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jul 7 00:05:17.162406 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jul 7 00:05:17.193961 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jul 7 00:05:17.194109 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jul 7 00:05:17.196138 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jul 7 00:05:17.198097 systemd[1]: Reached target initrd.target - Initrd Default Target. Jul 7 00:05:17.199060 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jul 7 00:05:17.206478 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jul 7 00:05:17.223025 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jul 7 00:05:17.230423 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jul 7 00:05:17.240038 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jul 7 00:05:17.241585 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 7 00:05:17.243100 systemd[1]: Stopped target timers.target - Timer Units. Jul 7 00:05:17.244648 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jul 7 00:05:17.244776 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jul 7 00:05:17.246980 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jul 7 00:05:17.248028 systemd[1]: Stopped target basic.target - Basic System. Jul 7 00:05:17.249222 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jul 7 00:05:17.250370 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jul 7 00:05:17.251695 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jul 7 00:05:17.253415 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jul 7 00:05:17.254878 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jul 7 00:05:17.256166 systemd[1]: Stopped target sysinit.target - System Initialization. Jul 7 00:05:17.257674 systemd[1]: Stopped target local-fs.target - Local File Systems. Jul 7 00:05:17.258930 systemd[1]: Stopped target swap.target - Swaps. Jul 7 00:05:17.260170 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jul 7 00:05:17.260494 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jul 7 00:05:17.261978 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jul 7 00:05:17.262899 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 7 00:05:17.264645 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jul 7 00:05:17.264754 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 7 00:05:17.266137 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jul 7 00:05:17.266236 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jul 7 00:05:17.267818 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jul 7 00:05:17.267927 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jul 7 00:05:17.268881 systemd[1]: ignition-files.service: Deactivated successfully. Jul 7 00:05:17.269050 systemd[1]: Stopped ignition-files.service - Ignition (files). Jul 7 00:05:17.279747 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jul 7 00:05:17.280325 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jul 7 00:05:17.280478 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jul 7 00:05:17.283474 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jul 7 00:05:17.284018 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jul 7 00:05:17.284165 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jul 7 00:05:17.284874 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jul 7 00:05:17.285008 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jul 7 00:05:17.296805 ignition[993]: INFO : Ignition 2.19.0 Jul 7 00:05:17.296805 ignition[993]: INFO : Stage: umount Jul 7 00:05:17.299701 ignition[993]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 7 00:05:17.299701 ignition[993]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jul 7 00:05:17.297711 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jul 7 00:05:17.305030 ignition[993]: INFO : umount: umount passed Jul 7 00:05:17.305030 ignition[993]: INFO : Ignition finished successfully Jul 7 00:05:17.297820 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jul 7 00:05:17.307544 systemd[1]: ignition-mount.service: Deactivated successfully. Jul 7 00:05:17.307665 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jul 7 00:05:17.309544 systemd[1]: ignition-disks.service: Deactivated successfully. Jul 7 00:05:17.309630 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jul 7 00:05:17.313562 systemd[1]: ignition-kargs.service: Deactivated successfully. Jul 7 00:05:17.313613 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jul 7 00:05:17.314479 systemd[1]: ignition-fetch.service: Deactivated successfully. Jul 7 00:05:17.314526 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Jul 7 00:05:17.315052 systemd[1]: Stopped target network.target - Network. Jul 7 00:05:17.315613 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jul 7 00:05:17.315666 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jul 7 00:05:17.317614 systemd[1]: Stopped target paths.target - Path Units. Jul 7 00:05:17.318914 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jul 7 00:05:17.319725 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 7 00:05:17.320324 systemd[1]: Stopped target slices.target - Slice Units. Jul 7 00:05:17.321475 systemd[1]: Stopped target sockets.target - Socket Units. Jul 7 00:05:17.323031 systemd[1]: iscsid.socket: Deactivated successfully. Jul 7 00:05:17.323106 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jul 7 00:05:17.346050 systemd[1]: iscsiuio.socket: Deactivated successfully. Jul 7 00:05:17.346112 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jul 7 00:05:17.347316 systemd[1]: ignition-setup.service: Deactivated successfully. Jul 7 00:05:17.347382 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jul 7 00:05:17.348439 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jul 7 00:05:17.348489 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jul 7 00:05:17.349858 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jul 7 00:05:17.351352 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jul 7 00:05:17.353519 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jul 7 00:05:17.354182 systemd[1]: sysroot-boot.service: Deactivated successfully. Jul 7 00:05:17.354336 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jul 7 00:05:17.355363 systemd-networkd[765]: eth0: DHCPv6 lease lost Jul 7 00:05:17.356031 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jul 7 00:05:17.356121 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jul 7 00:05:17.358221 systemd[1]: systemd-networkd.service: Deactivated successfully. Jul 7 00:05:17.358410 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jul 7 00:05:17.362380 systemd[1]: systemd-resolved.service: Deactivated successfully. Jul 7 00:05:17.362518 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jul 7 00:05:17.366848 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jul 7 00:05:17.366910 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jul 7 00:05:17.373396 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jul 7 00:05:17.374630 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jul 7 00:05:17.374690 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jul 7 00:05:17.376962 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jul 7 00:05:17.377017 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jul 7 00:05:17.378802 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jul 7 00:05:17.378852 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jul 7 00:05:17.379894 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jul 7 00:05:17.379941 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 7 00:05:17.382974 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 7 00:05:17.404991 systemd[1]: systemd-udevd.service: Deactivated successfully. Jul 7 00:05:17.405179 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 7 00:05:17.407008 systemd[1]: network-cleanup.service: Deactivated successfully. Jul 7 00:05:17.407111 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jul 7 00:05:17.409592 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jul 7 00:05:17.409667 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jul 7 00:05:17.410563 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jul 7 00:05:17.410603 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jul 7 00:05:17.411758 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jul 7 00:05:17.411809 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jul 7 00:05:17.413696 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jul 7 00:05:17.413743 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jul 7 00:05:17.414934 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jul 7 00:05:17.414983 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 7 00:05:17.422501 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jul 7 00:05:17.424010 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jul 7 00:05:17.424090 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 7 00:05:17.425493 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 7 00:05:17.425545 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 7 00:05:17.429137 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jul 7 00:05:17.429299 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jul 7 00:05:17.430420 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jul 7 00:05:17.443395 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jul 7 00:05:17.450744 systemd[1]: Switching root. Jul 7 00:05:17.482914 systemd-journald[177]: Journal stopped Jul 7 00:05:09.915577 kernel: Linux version 6.6.95-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Sun Jul 6 22:23:50 -00 2025 Jul 7 00:05:09.915596 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=akamai verity.usrhash=65c65ff9d50198f0ae5c37458dc3ff85c6a690e7aa124bb306a2f4c63a54d876 Jul 7 00:05:09.915605 kernel: BIOS-provided physical RAM map: Jul 7 00:05:09.915610 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009f7ff] usable Jul 7 00:05:09.915615 kernel: BIOS-e820: [mem 0x000000000009f800-0x000000000009ffff] reserved Jul 7 00:05:09.915624 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Jul 7 00:05:09.915630 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdcfff] usable Jul 7 00:05:09.915636 kernel: BIOS-e820: [mem 0x000000007ffdd000-0x000000007fffffff] reserved Jul 7 00:05:09.915641 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Jul 7 00:05:09.915646 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Jul 7 00:05:09.915652 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Jul 7 00:05:09.915657 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Jul 7 00:05:09.915663 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000017fffffff] usable Jul 7 00:05:09.915668 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Jul 7 00:05:09.915678 kernel: NX (Execute Disable) protection: active Jul 7 00:05:09.915684 kernel: APIC: Static calls initialized Jul 7 00:05:09.915689 kernel: SMBIOS 2.8 present. Jul 7 00:05:09.915695 kernel: DMI: Linode Compute Instance, BIOS Not Specified Jul 7 00:05:09.915701 kernel: Hypervisor detected: KVM Jul 7 00:05:09.915709 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Jul 7 00:05:09.915715 kernel: kvm-clock: using sched offset of 4506436040 cycles Jul 7 00:05:09.915721 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Jul 7 00:05:09.915727 kernel: tsc: Detected 2000.000 MHz processor Jul 7 00:05:09.915733 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jul 7 00:05:09.915740 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jul 7 00:05:09.915746 kernel: last_pfn = 0x180000 max_arch_pfn = 0x400000000 Jul 7 00:05:09.915752 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Jul 7 00:05:09.915758 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jul 7 00:05:09.915766 kernel: last_pfn = 0x7ffdd max_arch_pfn = 0x400000000 Jul 7 00:05:09.915772 kernel: Using GB pages for direct mapping Jul 7 00:05:09.915778 kernel: ACPI: Early table checksum verification disabled Jul 7 00:05:09.915784 kernel: ACPI: RSDP 0x00000000000F51B0 000014 (v00 BOCHS ) Jul 7 00:05:09.915790 kernel: ACPI: RSDT 0x000000007FFE2307 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jul 7 00:05:09.915796 kernel: ACPI: FACP 0x000000007FFE20F7 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Jul 7 00:05:09.915802 kernel: ACPI: DSDT 0x000000007FFE0040 0020B7 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jul 7 00:05:09.915807 kernel: ACPI: FACS 0x000000007FFE0000 000040 Jul 7 00:05:09.915813 kernel: ACPI: APIC 0x000000007FFE21EB 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jul 7 00:05:09.915822 kernel: ACPI: HPET 0x000000007FFE226B 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jul 7 00:05:09.915828 kernel: ACPI: MCFG 0x000000007FFE22A3 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Jul 7 00:05:09.915834 kernel: ACPI: WAET 0x000000007FFE22DF 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jul 7 00:05:09.915843 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe20f7-0x7ffe21ea] Jul 7 00:05:09.915850 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe20f6] Jul 7 00:05:09.915856 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Jul 7 00:05:09.915862 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe21eb-0x7ffe226a] Jul 7 00:05:09.915871 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe226b-0x7ffe22a2] Jul 7 00:05:09.915878 kernel: ACPI: Reserving MCFG table memory at [mem 0x7ffe22a3-0x7ffe22de] Jul 7 00:05:09.915884 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe22df-0x7ffe2306] Jul 7 00:05:09.915890 kernel: No NUMA configuration found Jul 7 00:05:09.915896 kernel: Faking a node at [mem 0x0000000000000000-0x000000017fffffff] Jul 7 00:05:09.915902 kernel: NODE_DATA(0) allocated [mem 0x17fffa000-0x17fffffff] Jul 7 00:05:09.915908 kernel: Zone ranges: Jul 7 00:05:09.915914 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jul 7 00:05:09.915923 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Jul 7 00:05:09.915929 kernel: Normal [mem 0x0000000100000000-0x000000017fffffff] Jul 7 00:05:09.915935 kernel: Movable zone start for each node Jul 7 00:05:09.915941 kernel: Early memory node ranges Jul 7 00:05:09.915948 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Jul 7 00:05:09.915954 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdcfff] Jul 7 00:05:09.915960 kernel: node 0: [mem 0x0000000100000000-0x000000017fffffff] Jul 7 00:05:09.915966 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000017fffffff] Jul 7 00:05:09.915972 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jul 7 00:05:09.915981 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Jul 7 00:05:09.915987 kernel: On node 0, zone Normal: 35 pages in unavailable ranges Jul 7 00:05:09.915993 kernel: ACPI: PM-Timer IO Port: 0x608 Jul 7 00:05:09.915999 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Jul 7 00:05:09.916005 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Jul 7 00:05:09.916011 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Jul 7 00:05:09.916018 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Jul 7 00:05:09.916024 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Jul 7 00:05:09.916030 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Jul 7 00:05:09.916039 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Jul 7 00:05:09.916045 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jul 7 00:05:09.916051 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Jul 7 00:05:09.916057 kernel: TSC deadline timer available Jul 7 00:05:09.916063 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Jul 7 00:05:09.916069 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Jul 7 00:05:09.916075 kernel: kvm-guest: KVM setup pv remote TLB flush Jul 7 00:05:09.916081 kernel: kvm-guest: setup PV sched yield Jul 7 00:05:09.916087 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Jul 7 00:05:09.916096 kernel: Booting paravirtualized kernel on KVM Jul 7 00:05:09.916102 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jul 7 00:05:09.916109 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Jul 7 00:05:09.916115 kernel: percpu: Embedded 58 pages/cpu s197096 r8192 d32280 u1048576 Jul 7 00:05:09.916121 kernel: pcpu-alloc: s197096 r8192 d32280 u1048576 alloc=1*2097152 Jul 7 00:05:09.916127 kernel: pcpu-alloc: [0] 0 1 Jul 7 00:05:09.916133 kernel: kvm-guest: PV spinlocks enabled Jul 7 00:05:09.916139 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Jul 7 00:05:09.916146 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=akamai verity.usrhash=65c65ff9d50198f0ae5c37458dc3ff85c6a690e7aa124bb306a2f4c63a54d876 Jul 7 00:05:09.916155 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jul 7 00:05:09.916161 kernel: random: crng init done Jul 7 00:05:09.916167 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jul 7 00:05:09.916174 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jul 7 00:05:09.916180 kernel: Fallback order for Node 0: 0 Jul 7 00:05:09.916186 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1031901 Jul 7 00:05:09.916192 kernel: Policy zone: Normal Jul 7 00:05:09.916198 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jul 7 00:05:09.916207 kernel: software IO TLB: area num 2. Jul 7 00:05:09.916213 kernel: Memory: 3966212K/4193772K available (12288K kernel code, 2295K rwdata, 22748K rodata, 42868K init, 2324K bss, 227300K reserved, 0K cma-reserved) Jul 7 00:05:09.916219 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Jul 7 00:05:09.916225 kernel: ftrace: allocating 37966 entries in 149 pages Jul 7 00:05:09.916232 kernel: ftrace: allocated 149 pages with 4 groups Jul 7 00:05:09.916238 kernel: Dynamic Preempt: voluntary Jul 7 00:05:09.916244 kernel: rcu: Preemptible hierarchical RCU implementation. Jul 7 00:05:09.916268 kernel: rcu: RCU event tracing is enabled. Jul 7 00:05:09.916275 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Jul 7 00:05:09.916284 kernel: Trampoline variant of Tasks RCU enabled. Jul 7 00:05:09.916291 kernel: Rude variant of Tasks RCU enabled. Jul 7 00:05:09.916297 kernel: Tracing variant of Tasks RCU enabled. Jul 7 00:05:09.916303 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jul 7 00:05:09.916309 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Jul 7 00:05:09.916315 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Jul 7 00:05:09.916321 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jul 7 00:05:09.916327 kernel: Console: colour VGA+ 80x25 Jul 7 00:05:09.916334 kernel: printk: console [tty0] enabled Jul 7 00:05:09.916341 kernel: printk: console [ttyS0] enabled Jul 7 00:05:09.916350 kernel: ACPI: Core revision 20230628 Jul 7 00:05:09.916357 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Jul 7 00:05:09.916363 kernel: APIC: Switch to symmetric I/O mode setup Jul 7 00:05:09.916377 kernel: x2apic enabled Jul 7 00:05:09.916387 kernel: APIC: Switched APIC routing to: physical x2apic Jul 7 00:05:09.916393 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Jul 7 00:05:09.916400 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Jul 7 00:05:09.916406 kernel: kvm-guest: setup PV IPIs Jul 7 00:05:09.916413 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jul 7 00:05:09.916419 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Jul 7 00:05:09.916426 kernel: Calibrating delay loop (skipped) preset value.. 4000.00 BogoMIPS (lpj=2000000) Jul 7 00:05:09.916435 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Jul 7 00:05:09.916442 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Jul 7 00:05:09.916448 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Jul 7 00:05:09.916455 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jul 7 00:05:09.916461 kernel: Spectre V2 : Mitigation: Retpolines Jul 7 00:05:09.916470 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Jul 7 00:05:09.916477 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Jul 7 00:05:09.916483 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Jul 7 00:05:09.916490 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Jul 7 00:05:09.916496 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Jul 7 00:05:09.916503 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Jul 7 00:05:09.916510 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Jul 7 00:05:09.916516 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jul 7 00:05:09.916525 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jul 7 00:05:09.916532 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jul 7 00:05:09.916538 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Jul 7 00:05:09.916544 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jul 7 00:05:09.916551 kernel: x86/fpu: xstate_offset[9]: 832, xstate_sizes[9]: 8 Jul 7 00:05:09.916558 kernel: x86/fpu: Enabled xstate features 0x207, context size is 840 bytes, using 'compacted' format. Jul 7 00:05:09.916564 kernel: Freeing SMP alternatives memory: 32K Jul 7 00:05:09.916570 kernel: pid_max: default: 32768 minimum: 301 Jul 7 00:05:09.916577 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Jul 7 00:05:09.916586 kernel: landlock: Up and running. Jul 7 00:05:09.916592 kernel: SELinux: Initializing. Jul 7 00:05:09.916599 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jul 7 00:05:09.916605 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jul 7 00:05:09.916612 kernel: smpboot: CPU0: AMD EPYC 7713 64-Core Processor (family: 0x19, model: 0x1, stepping: 0x1) Jul 7 00:05:09.916618 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jul 7 00:05:09.916625 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jul 7 00:05:09.916631 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jul 7 00:05:09.916638 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Jul 7 00:05:09.916647 kernel: ... version: 0 Jul 7 00:05:09.916653 kernel: ... bit width: 48 Jul 7 00:05:09.916660 kernel: ... generic registers: 6 Jul 7 00:05:09.916666 kernel: ... value mask: 0000ffffffffffff Jul 7 00:05:09.916673 kernel: ... max period: 00007fffffffffff Jul 7 00:05:09.916679 kernel: ... fixed-purpose events: 0 Jul 7 00:05:09.916685 kernel: ... event mask: 000000000000003f Jul 7 00:05:09.916692 kernel: signal: max sigframe size: 3376 Jul 7 00:05:09.916698 kernel: rcu: Hierarchical SRCU implementation. Jul 7 00:05:09.916707 kernel: rcu: Max phase no-delay instances is 400. Jul 7 00:05:09.916714 kernel: smp: Bringing up secondary CPUs ... Jul 7 00:05:09.916720 kernel: smpboot: x86: Booting SMP configuration: Jul 7 00:05:09.916727 kernel: .... node #0, CPUs: #1 Jul 7 00:05:09.916733 kernel: smp: Brought up 1 node, 2 CPUs Jul 7 00:05:09.916740 kernel: smpboot: Max logical packages: 1 Jul 7 00:05:09.916746 kernel: smpboot: Total of 2 processors activated (8000.00 BogoMIPS) Jul 7 00:05:09.916753 kernel: devtmpfs: initialized Jul 7 00:05:09.916759 kernel: x86/mm: Memory block size: 128MB Jul 7 00:05:09.916765 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jul 7 00:05:09.916775 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Jul 7 00:05:09.916781 kernel: pinctrl core: initialized pinctrl subsystem Jul 7 00:05:09.916788 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jul 7 00:05:09.916794 kernel: audit: initializing netlink subsys (disabled) Jul 7 00:05:09.916801 kernel: audit: type=2000 audit(1751846709.061:1): state=initialized audit_enabled=0 res=1 Jul 7 00:05:09.916807 kernel: thermal_sys: Registered thermal governor 'step_wise' Jul 7 00:05:09.916813 kernel: thermal_sys: Registered thermal governor 'user_space' Jul 7 00:05:09.916820 kernel: cpuidle: using governor menu Jul 7 00:05:09.916826 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jul 7 00:05:09.916835 kernel: dca service started, version 1.12.1 Jul 7 00:05:09.916842 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Jul 7 00:05:09.916848 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved as E820 entry Jul 7 00:05:09.916855 kernel: PCI: Using configuration type 1 for base access Jul 7 00:05:09.916862 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jul 7 00:05:09.916868 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jul 7 00:05:09.916874 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jul 7 00:05:09.916881 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jul 7 00:05:09.916890 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jul 7 00:05:09.916896 kernel: ACPI: Added _OSI(Module Device) Jul 7 00:05:09.916903 kernel: ACPI: Added _OSI(Processor Device) Jul 7 00:05:09.916909 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jul 7 00:05:09.916916 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jul 7 00:05:09.916922 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Jul 7 00:05:09.916928 kernel: ACPI: Interpreter enabled Jul 7 00:05:09.916935 kernel: ACPI: PM: (supports S0 S3 S5) Jul 7 00:05:09.916941 kernel: ACPI: Using IOAPIC for interrupt routing Jul 7 00:05:09.916948 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jul 7 00:05:09.916957 kernel: PCI: Using E820 reservations for host bridge windows Jul 7 00:05:09.916963 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Jul 7 00:05:09.916969 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jul 7 00:05:09.917151 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jul 7 00:05:09.919344 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Jul 7 00:05:09.919473 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Jul 7 00:05:09.919483 kernel: PCI host bridge to bus 0000:00 Jul 7 00:05:09.919609 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jul 7 00:05:09.919711 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Jul 7 00:05:09.919816 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jul 7 00:05:09.919919 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xafffffff window] Jul 7 00:05:09.920022 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Jul 7 00:05:09.920125 kernel: pci_bus 0000:00: root bus resource [mem 0x180000000-0x97fffffff window] Jul 7 00:05:09.920229 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jul 7 00:05:09.921207 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Jul 7 00:05:09.921375 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Jul 7 00:05:09.921489 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfd000000-0xfdffffff pref] Jul 7 00:05:09.921607 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfebd0000-0xfebd0fff] Jul 7 00:05:09.921721 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfebc0000-0xfebcffff pref] Jul 7 00:05:09.923662 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jul 7 00:05:09.923803 kernel: pci 0000:00:02.0: [1af4:1004] type 00 class 0x010000 Jul 7 00:05:09.923917 kernel: pci 0000:00:02.0: reg 0x10: [io 0xc000-0xc03f] Jul 7 00:05:09.924028 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xfebd1000-0xfebd1fff] Jul 7 00:05:09.924140 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfe000000-0xfe003fff 64bit pref] Jul 7 00:05:09.924303 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Jul 7 00:05:09.924421 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc040-0xc07f] Jul 7 00:05:09.924530 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebd2000-0xfebd2fff] Jul 7 00:05:09.924645 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe004000-0xfe007fff 64bit pref] Jul 7 00:05:09.924754 kernel: pci 0000:00:03.0: reg 0x30: [mem 0xfeb80000-0xfebbffff pref] Jul 7 00:05:09.924871 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Jul 7 00:05:09.924979 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Jul 7 00:05:09.925106 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Jul 7 00:05:09.925217 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc0c0-0xc0df] Jul 7 00:05:09.927833 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfebd3000-0xfebd3fff] Jul 7 00:05:09.927965 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Jul 7 00:05:09.928075 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Jul 7 00:05:09.928085 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Jul 7 00:05:09.928092 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Jul 7 00:05:09.928099 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jul 7 00:05:09.928106 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Jul 7 00:05:09.928112 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Jul 7 00:05:09.928124 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Jul 7 00:05:09.928130 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Jul 7 00:05:09.928137 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Jul 7 00:05:09.928143 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Jul 7 00:05:09.928149 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Jul 7 00:05:09.928156 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Jul 7 00:05:09.928162 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Jul 7 00:05:09.928169 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Jul 7 00:05:09.928175 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Jul 7 00:05:09.928184 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Jul 7 00:05:09.928191 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Jul 7 00:05:09.928197 kernel: iommu: Default domain type: Translated Jul 7 00:05:09.928204 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jul 7 00:05:09.928210 kernel: PCI: Using ACPI for IRQ routing Jul 7 00:05:09.928217 kernel: PCI: pci_cache_line_size set to 64 bytes Jul 7 00:05:09.928224 kernel: e820: reserve RAM buffer [mem 0x0009f800-0x0009ffff] Jul 7 00:05:09.928230 kernel: e820: reserve RAM buffer [mem 0x7ffdd000-0x7fffffff] Jul 7 00:05:09.930107 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Jul 7 00:05:09.930232 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Jul 7 00:05:09.930379 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jul 7 00:05:09.930390 kernel: vgaarb: loaded Jul 7 00:05:09.930397 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Jul 7 00:05:09.930404 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Jul 7 00:05:09.930410 kernel: clocksource: Switched to clocksource kvm-clock Jul 7 00:05:09.930417 kernel: VFS: Disk quotas dquot_6.6.0 Jul 7 00:05:09.930424 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jul 7 00:05:09.930430 kernel: pnp: PnP ACPI init Jul 7 00:05:09.930560 kernel: system 00:04: [mem 0xb0000000-0xbfffffff window] has been reserved Jul 7 00:05:09.930570 kernel: pnp: PnP ACPI: found 5 devices Jul 7 00:05:09.930577 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jul 7 00:05:09.930584 kernel: NET: Registered PF_INET protocol family Jul 7 00:05:09.930590 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jul 7 00:05:09.930597 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jul 7 00:05:09.930603 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jul 7 00:05:09.930610 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jul 7 00:05:09.930620 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jul 7 00:05:09.930626 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jul 7 00:05:09.930633 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jul 7 00:05:09.930639 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jul 7 00:05:09.930646 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jul 7 00:05:09.930652 kernel: NET: Registered PF_XDP protocol family Jul 7 00:05:09.930755 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Jul 7 00:05:09.930855 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Jul 7 00:05:09.930959 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Jul 7 00:05:09.931056 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xafffffff window] Jul 7 00:05:09.931154 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Jul 7 00:05:09.931299 kernel: pci_bus 0000:00: resource 9 [mem 0x180000000-0x97fffffff window] Jul 7 00:05:09.931310 kernel: PCI: CLS 0 bytes, default 64 Jul 7 00:05:09.931317 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Jul 7 00:05:09.931324 kernel: software IO TLB: mapped [mem 0x000000007bfdd000-0x000000007ffdd000] (64MB) Jul 7 00:05:09.931331 kernel: Initialise system trusted keyrings Jul 7 00:05:09.931337 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jul 7 00:05:09.931348 kernel: Key type asymmetric registered Jul 7 00:05:09.931355 kernel: Asymmetric key parser 'x509' registered Jul 7 00:05:09.931361 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Jul 7 00:05:09.931368 kernel: io scheduler mq-deadline registered Jul 7 00:05:09.931375 kernel: io scheduler kyber registered Jul 7 00:05:09.931381 kernel: io scheduler bfq registered Jul 7 00:05:09.931388 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jul 7 00:05:09.931394 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Jul 7 00:05:09.931401 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Jul 7 00:05:09.931411 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jul 7 00:05:09.931417 kernel: 00:02: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jul 7 00:05:09.931424 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Jul 7 00:05:09.931430 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jul 7 00:05:09.931437 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jul 7 00:05:09.931559 kernel: rtc_cmos 00:03: RTC can wake from S4 Jul 7 00:05:09.931569 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jul 7 00:05:09.931670 kernel: rtc_cmos 00:03: registered as rtc0 Jul 7 00:05:09.931777 kernel: rtc_cmos 00:03: setting system clock to 2025-07-07T00:05:09 UTC (1751846709) Jul 7 00:05:09.931880 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Jul 7 00:05:09.931889 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Jul 7 00:05:09.931895 kernel: NET: Registered PF_INET6 protocol family Jul 7 00:05:09.931902 kernel: Segment Routing with IPv6 Jul 7 00:05:09.931908 kernel: In-situ OAM (IOAM) with IPv6 Jul 7 00:05:09.931915 kernel: NET: Registered PF_PACKET protocol family Jul 7 00:05:09.931921 kernel: Key type dns_resolver registered Jul 7 00:05:09.931927 kernel: IPI shorthand broadcast: enabled Jul 7 00:05:09.931938 kernel: sched_clock: Marking stable (739004807, 230949109)->(1039360891, -69406975) Jul 7 00:05:09.931944 kernel: registered taskstats version 1 Jul 7 00:05:09.931951 kernel: Loading compiled-in X.509 certificates Jul 7 00:05:09.931958 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.95-flatcar: 6372c48ca52cc7f7bbee5675b604584c1c68ec5b' Jul 7 00:05:09.931964 kernel: Key type .fscrypt registered Jul 7 00:05:09.931970 kernel: Key type fscrypt-provisioning registered Jul 7 00:05:09.931977 kernel: ima: No TPM chip found, activating TPM-bypass! Jul 7 00:05:09.931984 kernel: ima: Allocated hash algorithm: sha1 Jul 7 00:05:09.931990 kernel: ima: No architecture policies found Jul 7 00:05:09.931999 kernel: clk: Disabling unused clocks Jul 7 00:05:09.932006 kernel: Freeing unused kernel image (initmem) memory: 42868K Jul 7 00:05:09.932012 kernel: Write protecting the kernel read-only data: 36864k Jul 7 00:05:09.932019 kernel: Freeing unused kernel image (rodata/data gap) memory: 1828K Jul 7 00:05:09.932025 kernel: Run /init as init process Jul 7 00:05:09.932032 kernel: with arguments: Jul 7 00:05:09.932038 kernel: /init Jul 7 00:05:09.932044 kernel: with environment: Jul 7 00:05:09.932051 kernel: HOME=/ Jul 7 00:05:09.932060 kernel: TERM=linux Jul 7 00:05:09.932066 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jul 7 00:05:09.932075 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jul 7 00:05:09.932083 systemd[1]: Detected virtualization kvm. Jul 7 00:05:09.932090 systemd[1]: Detected architecture x86-64. Jul 7 00:05:09.932097 systemd[1]: Running in initrd. Jul 7 00:05:09.932104 systemd[1]: No hostname configured, using default hostname. Jul 7 00:05:09.932113 systemd[1]: Hostname set to . Jul 7 00:05:09.932120 systemd[1]: Initializing machine ID from random generator. Jul 7 00:05:09.932127 systemd[1]: Queued start job for default target initrd.target. Jul 7 00:05:09.932134 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 7 00:05:09.932141 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 7 00:05:09.932163 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jul 7 00:05:09.932176 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jul 7 00:05:09.932184 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jul 7 00:05:09.932191 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jul 7 00:05:09.932199 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jul 7 00:05:09.932207 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jul 7 00:05:09.932214 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 7 00:05:09.932221 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jul 7 00:05:09.932231 systemd[1]: Reached target paths.target - Path Units. Jul 7 00:05:09.932238 systemd[1]: Reached target slices.target - Slice Units. Jul 7 00:05:09.932246 systemd[1]: Reached target swap.target - Swaps. Jul 7 00:05:09.933482 systemd[1]: Reached target timers.target - Timer Units. Jul 7 00:05:09.933491 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jul 7 00:05:09.933499 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jul 7 00:05:09.933506 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jul 7 00:05:09.933514 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jul 7 00:05:09.933526 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jul 7 00:05:09.933533 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jul 7 00:05:09.933540 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jul 7 00:05:09.933547 systemd[1]: Reached target sockets.target - Socket Units. Jul 7 00:05:09.933555 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jul 7 00:05:09.933562 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jul 7 00:05:09.933569 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jul 7 00:05:09.933577 systemd[1]: Starting systemd-fsck-usr.service... Jul 7 00:05:09.933584 systemd[1]: Starting systemd-journald.service - Journal Service... Jul 7 00:05:09.933594 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jul 7 00:05:09.933622 systemd-journald[177]: Collecting audit messages is disabled. Jul 7 00:05:09.933639 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 7 00:05:09.933647 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jul 7 00:05:09.933657 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jul 7 00:05:09.933664 systemd[1]: Finished systemd-fsck-usr.service. Jul 7 00:05:09.933674 systemd-journald[177]: Journal started Jul 7 00:05:09.933692 systemd-journald[177]: Runtime Journal (/run/log/journal/f45ce32254a84bb1a3346ddec5b181cf) is 8.0M, max 78.3M, 70.3M free. Jul 7 00:05:09.937453 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jul 7 00:05:09.932339 systemd-modules-load[178]: Inserted module 'overlay' Jul 7 00:05:09.987153 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jul 7 00:05:09.987173 kernel: Bridge firewalling registered Jul 7 00:05:09.957849 systemd-modules-load[178]: Inserted module 'br_netfilter' Jul 7 00:05:09.994420 systemd[1]: Started systemd-journald.service - Journal Service. Jul 7 00:05:09.994861 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jul 7 00:05:09.995684 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 7 00:05:09.999636 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jul 7 00:05:10.005396 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 7 00:05:10.012396 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 7 00:05:10.016420 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jul 7 00:05:10.024283 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jul 7 00:05:10.034951 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 7 00:05:10.049143 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 7 00:05:10.051492 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 7 00:05:10.059438 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jul 7 00:05:10.060632 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 7 00:05:10.072430 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jul 7 00:05:10.080958 dracut-cmdline[210]: dracut-dracut-053 Jul 7 00:05:10.085734 dracut-cmdline[210]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=akamai verity.usrhash=65c65ff9d50198f0ae5c37458dc3ff85c6a690e7aa124bb306a2f4c63a54d876 Jul 7 00:05:10.113764 systemd-resolved[212]: Positive Trust Anchors: Jul 7 00:05:10.114515 systemd-resolved[212]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 7 00:05:10.114545 systemd-resolved[212]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jul 7 00:05:10.117856 systemd-resolved[212]: Defaulting to hostname 'linux'. Jul 7 00:05:10.119115 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jul 7 00:05:10.120167 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jul 7 00:05:10.167313 kernel: SCSI subsystem initialized Jul 7 00:05:10.176274 kernel: Loading iSCSI transport class v2.0-870. Jul 7 00:05:10.188284 kernel: iscsi: registered transport (tcp) Jul 7 00:05:10.209293 kernel: iscsi: registered transport (qla4xxx) Jul 7 00:05:10.209329 kernel: QLogic iSCSI HBA Driver Jul 7 00:05:10.262742 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jul 7 00:05:10.268456 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jul 7 00:05:10.294424 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jul 7 00:05:10.294508 kernel: device-mapper: uevent: version 1.0.3 Jul 7 00:05:10.297294 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jul 7 00:05:10.339291 kernel: raid6: avx2x4 gen() 31712 MB/s Jul 7 00:05:10.357276 kernel: raid6: avx2x2 gen() 28085 MB/s Jul 7 00:05:10.375739 kernel: raid6: avx2x1 gen() 24966 MB/s Jul 7 00:05:10.375756 kernel: raid6: using algorithm avx2x4 gen() 31712 MB/s Jul 7 00:05:10.394815 kernel: raid6: .... xor() 5162 MB/s, rmw enabled Jul 7 00:05:10.394830 kernel: raid6: using avx2x2 recovery algorithm Jul 7 00:05:10.415284 kernel: xor: automatically using best checksumming function avx Jul 7 00:05:10.548319 kernel: Btrfs loaded, zoned=no, fsverity=no Jul 7 00:05:10.564873 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jul 7 00:05:10.573516 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 7 00:05:10.588005 systemd-udevd[395]: Using default interface naming scheme 'v255'. Jul 7 00:05:10.593753 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 7 00:05:10.600390 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jul 7 00:05:10.618560 dracut-pre-trigger[402]: rd.md=0: removing MD RAID activation Jul 7 00:05:10.652132 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jul 7 00:05:10.656377 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jul 7 00:05:10.714732 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jul 7 00:05:10.721405 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jul 7 00:05:10.733804 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jul 7 00:05:10.736759 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jul 7 00:05:10.740336 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 7 00:05:10.740884 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jul 7 00:05:10.750378 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jul 7 00:05:10.762169 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jul 7 00:05:10.788751 kernel: scsi host0: Virtio SCSI HBA Jul 7 00:05:10.929309 kernel: scsi 0:0:0:0: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Jul 7 00:05:10.929387 kernel: cryptd: max_cpu_qlen set to 1000 Jul 7 00:05:10.942641 kernel: libata version 3.00 loaded. Jul 7 00:05:10.979622 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jul 7 00:05:10.979829 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 7 00:05:10.981813 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 7 00:05:10.983204 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 7 00:05:10.983395 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 7 00:05:10.988189 kernel: AVX2 version of gcm_enc/dec engaged. Jul 7 00:05:10.987081 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jul 7 00:05:10.991329 kernel: AES CTR mode by8 optimization enabled Jul 7 00:05:10.995061 kernel: ahci 0000:00:1f.2: version 3.0 Jul 7 00:05:10.994516 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 7 00:05:11.002718 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Jul 7 00:05:11.010925 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Jul 7 00:05:11.011124 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Jul 7 00:05:11.017097 kernel: sd 0:0:0:0: Power-on or device reset occurred Jul 7 00:05:11.021188 kernel: sd 0:0:0:0: [sda] 167739392 512-byte logical blocks: (85.9 GB/80.0 GiB) Jul 7 00:05:11.021370 kernel: sd 0:0:0:0: [sda] Write Protect is off Jul 7 00:05:11.021512 kernel: sd 0:0:0:0: [sda] Mode Sense: 63 00 00 08 Jul 7 00:05:11.021648 kernel: sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Jul 7 00:05:11.021781 kernel: scsi host1: ahci Jul 7 00:05:11.022272 kernel: scsi host2: ahci Jul 7 00:05:11.023449 kernel: scsi host3: ahci Jul 7 00:05:11.025497 kernel: scsi host4: ahci Jul 7 00:05:11.026268 kernel: scsi host5: ahci Jul 7 00:05:11.030364 kernel: scsi host6: ahci Jul 7 00:05:11.030521 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jul 7 00:05:11.030538 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3100 irq 46 Jul 7 00:05:11.030548 kernel: GPT:9289727 != 167739391 Jul 7 00:05:11.030557 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3180 irq 46 Jul 7 00:05:11.030567 kernel: GPT:Alternate GPT header not at the end of the disk. Jul 7 00:05:11.030577 kernel: GPT:9289727 != 167739391 Jul 7 00:05:11.030586 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3200 irq 46 Jul 7 00:05:11.030595 kernel: GPT: Use GNU Parted to correct GPT errors. Jul 7 00:05:11.030604 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 7 00:05:11.030613 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3280 irq 46 Jul 7 00:05:11.035292 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Jul 7 00:05:11.035743 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3300 irq 46 Jul 7 00:05:11.052356 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd3000 port 0xfebd3380 irq 46 Jul 7 00:05:11.106377 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 7 00:05:11.112447 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 7 00:05:11.127171 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 7 00:05:11.361822 kernel: ata3: SATA link down (SStatus 0 SControl 300) Jul 7 00:05:11.361889 kernel: ata1: SATA link down (SStatus 0 SControl 300) Jul 7 00:05:11.361902 kernel: ata2: SATA link down (SStatus 0 SControl 300) Jul 7 00:05:11.362265 kernel: ata4: SATA link down (SStatus 0 SControl 300) Jul 7 00:05:11.364275 kernel: ata5: SATA link down (SStatus 0 SControl 300) Jul 7 00:05:11.365269 kernel: ata6: SATA link down (SStatus 0 SControl 300) Jul 7 00:05:11.407634 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Jul 7 00:05:11.416614 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/sda6 scanned by (udev-worker) (441) Jul 7 00:05:11.416634 kernel: BTRFS: device fsid 01287863-c21f-4cbb-820d-bbae8208f32f devid 1 transid 34 /dev/sda3 scanned by (udev-worker) (439) Jul 7 00:05:11.418466 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Jul 7 00:05:11.429220 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Jul 7 00:05:11.434048 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Jul 7 00:05:11.434714 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Jul 7 00:05:11.448445 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jul 7 00:05:11.453797 disk-uuid[565]: Primary Header is updated. Jul 7 00:05:11.453797 disk-uuid[565]: Secondary Entries is updated. Jul 7 00:05:11.453797 disk-uuid[565]: Secondary Header is updated. Jul 7 00:05:11.459321 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 7 00:05:11.464273 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 7 00:05:12.468213 disk-uuid[566]: The operation has completed successfully. Jul 7 00:05:12.470302 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 7 00:05:12.512243 systemd[1]: disk-uuid.service: Deactivated successfully. Jul 7 00:05:12.512511 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jul 7 00:05:12.530377 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jul 7 00:05:12.534726 sh[580]: Success Jul 7 00:05:12.547332 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Jul 7 00:05:12.588195 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jul 7 00:05:12.602335 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jul 7 00:05:12.603667 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jul 7 00:05:12.622909 kernel: BTRFS info (device dm-0): first mount of filesystem 01287863-c21f-4cbb-820d-bbae8208f32f Jul 7 00:05:12.622936 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jul 7 00:05:12.622948 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jul 7 00:05:12.627197 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jul 7 00:05:12.627212 kernel: BTRFS info (device dm-0): using free space tree Jul 7 00:05:12.636321 kernel: BTRFS info (device dm-0): enabling ssd optimizations Jul 7 00:05:12.638280 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jul 7 00:05:12.639309 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jul 7 00:05:12.643358 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jul 7 00:05:12.646374 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jul 7 00:05:12.663578 kernel: BTRFS info (device sda6): first mount of filesystem 11f56a79-b29d-47db-ad8e-56effe5ac41b Jul 7 00:05:12.663603 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 7 00:05:12.663614 kernel: BTRFS info (device sda6): using free space tree Jul 7 00:05:12.669944 kernel: BTRFS info (device sda6): enabling ssd optimizations Jul 7 00:05:12.669967 kernel: BTRFS info (device sda6): auto enabling async discard Jul 7 00:05:12.681857 systemd[1]: mnt-oem.mount: Deactivated successfully. Jul 7 00:05:12.682821 kernel: BTRFS info (device sda6): last unmount of filesystem 11f56a79-b29d-47db-ad8e-56effe5ac41b Jul 7 00:05:12.689548 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jul 7 00:05:12.696103 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jul 7 00:05:12.762512 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jul 7 00:05:12.768480 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jul 7 00:05:12.779430 ignition[694]: Ignition 2.19.0 Jul 7 00:05:12.780177 ignition[694]: Stage: fetch-offline Jul 7 00:05:12.780730 ignition[694]: no configs at "/usr/lib/ignition/base.d" Jul 7 00:05:12.780741 ignition[694]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jul 7 00:05:12.780832 ignition[694]: parsed url from cmdline: "" Jul 7 00:05:12.780836 ignition[694]: no config URL provided Jul 7 00:05:12.780841 ignition[694]: reading system config file "/usr/lib/ignition/user.ign" Jul 7 00:05:12.784505 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jul 7 00:05:12.780850 ignition[694]: no config at "/usr/lib/ignition/user.ign" Jul 7 00:05:12.780856 ignition[694]: failed to fetch config: resource requires networking Jul 7 00:05:12.781026 ignition[694]: Ignition finished successfully Jul 7 00:05:12.797840 systemd-networkd[765]: lo: Link UP Jul 7 00:05:12.797854 systemd-networkd[765]: lo: Gained carrier Jul 7 00:05:12.799418 systemd-networkd[765]: Enumeration completed Jul 7 00:05:12.799827 systemd-networkd[765]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 7 00:05:12.799831 systemd-networkd[765]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jul 7 00:05:12.800924 systemd[1]: Started systemd-networkd.service - Network Configuration. Jul 7 00:05:12.802123 systemd-networkd[765]: eth0: Link UP Jul 7 00:05:12.802127 systemd-networkd[765]: eth0: Gained carrier Jul 7 00:05:12.802143 systemd-networkd[765]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 7 00:05:12.802472 systemd[1]: Reached target network.target - Network. Jul 7 00:05:12.812552 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Jul 7 00:05:12.826225 ignition[769]: Ignition 2.19.0 Jul 7 00:05:12.826239 ignition[769]: Stage: fetch Jul 7 00:05:12.826416 ignition[769]: no configs at "/usr/lib/ignition/base.d" Jul 7 00:05:12.826428 ignition[769]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jul 7 00:05:12.826506 ignition[769]: parsed url from cmdline: "" Jul 7 00:05:12.826510 ignition[769]: no config URL provided Jul 7 00:05:12.826515 ignition[769]: reading system config file "/usr/lib/ignition/user.ign" Jul 7 00:05:12.826523 ignition[769]: no config at "/usr/lib/ignition/user.ign" Jul 7 00:05:12.826543 ignition[769]: PUT http://169.254.169.254/v1/token: attempt #1 Jul 7 00:05:12.826697 ignition[769]: PUT error: Put "http://169.254.169.254/v1/token": dial tcp 169.254.169.254:80: connect: network is unreachable Jul 7 00:05:13.026916 ignition[769]: PUT http://169.254.169.254/v1/token: attempt #2 Jul 7 00:05:13.027087 ignition[769]: PUT error: Put "http://169.254.169.254/v1/token": dial tcp 169.254.169.254:80: connect: network is unreachable Jul 7 00:05:13.293331 systemd-networkd[765]: eth0: DHCPv4 address 172.234.31.125/24, gateway 172.234.31.1 acquired from 23.205.167.142 Jul 7 00:05:13.427713 ignition[769]: PUT http://169.254.169.254/v1/token: attempt #3 Jul 7 00:05:13.519316 ignition[769]: PUT result: OK Jul 7 00:05:13.519392 ignition[769]: GET http://169.254.169.254/v1/user-data: attempt #1 Jul 7 00:05:13.631918 ignition[769]: GET result: OK Jul 7 00:05:13.632085 ignition[769]: parsing config with SHA512: 794185a1569c43f9981f60e78de291d07968e5187c2a4e45ef7ca9597f09bff4ef6475b3edfaa89241ae8685e498f874a0df9f7bc251debfd8023f77d7c744c9 Jul 7 00:05:13.636605 unknown[769]: fetched base config from "system" Jul 7 00:05:13.636618 unknown[769]: fetched base config from "system" Jul 7 00:05:13.636870 ignition[769]: fetch: fetch complete Jul 7 00:05:13.636631 unknown[769]: fetched user config from "akamai" Jul 7 00:05:13.636875 ignition[769]: fetch: fetch passed Jul 7 00:05:13.636919 ignition[769]: Ignition finished successfully Jul 7 00:05:13.646736 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Jul 7 00:05:13.652435 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jul 7 00:05:13.670929 ignition[777]: Ignition 2.19.0 Jul 7 00:05:13.670943 ignition[777]: Stage: kargs Jul 7 00:05:13.671096 ignition[777]: no configs at "/usr/lib/ignition/base.d" Jul 7 00:05:13.671109 ignition[777]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jul 7 00:05:13.671824 ignition[777]: kargs: kargs passed Jul 7 00:05:13.673410 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jul 7 00:05:13.671865 ignition[777]: Ignition finished successfully Jul 7 00:05:13.679485 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jul 7 00:05:13.696730 ignition[783]: Ignition 2.19.0 Jul 7 00:05:13.696745 ignition[783]: Stage: disks Jul 7 00:05:13.696889 ignition[783]: no configs at "/usr/lib/ignition/base.d" Jul 7 00:05:13.701079 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jul 7 00:05:13.696903 ignition[783]: no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jul 7 00:05:13.702216 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jul 7 00:05:13.697808 ignition[783]: disks: disks passed Jul 7 00:05:13.703129 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jul 7 00:05:13.697846 ignition[783]: Ignition finished successfully Jul 7 00:05:13.704457 systemd[1]: Reached target local-fs.target - Local File Systems. Jul 7 00:05:13.727770 systemd[1]: Reached target sysinit.target - System Initialization. Jul 7 00:05:13.728999 systemd[1]: Reached target basic.target - Basic System. Jul 7 00:05:13.735461 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jul 7 00:05:13.752318 systemd-fsck[791]: ROOT: clean, 14/553520 files, 52654/553472 blocks Jul 7 00:05:13.754797 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jul 7 00:05:13.760365 systemd[1]: Mounting sysroot.mount - /sysroot... Jul 7 00:05:13.854280 kernel: EXT4-fs (sda9): mounted filesystem c3eefe20-4a42-420d-8034-4d5498275b2f r/w with ordered data mode. Quota mode: none. Jul 7 00:05:13.855431 systemd[1]: Mounted sysroot.mount - /sysroot. Jul 7 00:05:13.856808 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jul 7 00:05:13.874364 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jul 7 00:05:13.877536 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jul 7 00:05:13.879339 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jul 7 00:05:13.879398 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jul 7 00:05:13.879423 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jul 7 00:05:13.891318 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/sda6 scanned by mount (799) Jul 7 00:05:13.894053 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jul 7 00:05:13.902618 kernel: BTRFS info (device sda6): first mount of filesystem 11f56a79-b29d-47db-ad8e-56effe5ac41b Jul 7 00:05:13.902635 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 7 00:05:13.902646 kernel: BTRFS info (device sda6): using free space tree Jul 7 00:05:13.902656 kernel: BTRFS info (device sda6): enabling ssd optimizations Jul 7 00:05:13.902666 kernel: BTRFS info (device sda6): auto enabling async discard Jul 7 00:05:13.904286 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jul 7 00:05:13.910384 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jul 7 00:05:13.957340 initrd-setup-root[823]: cut: /sysroot/etc/passwd: No such file or directory Jul 7 00:05:13.964981 initrd-setup-root[830]: cut: /sysroot/etc/group: No such file or directory Jul 7 00:05:13.970662 initrd-setup-root[837]: cut: /sysroot/etc/shadow: No such file or directory Jul 7 00:05:13.976352 initrd-setup-root[844]: cut: /sysroot/etc/gshadow: No such file or directory Jul 7 00:05:14.080280 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jul 7 00:05:14.085410 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jul 7 00:05:14.088455 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jul 7 00:05:14.099287 kernel: BTRFS info (device sda6): last unmount of filesystem 11f56a79-b29d-47db-ad8e-56effe5ac41b Jul 7 00:05:14.099506 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jul 7 00:05:14.121983 ignition[913]: INFO : Ignition 2.19.0 Jul 7 00:05:14.122592 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jul 7 00:05:14.124869 ignition[913]: INFO : Stage: mount Jul 7 00:05:14.124869 ignition[913]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 7 00:05:14.124869 ignition[913]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jul 7 00:05:14.127264 ignition[913]: INFO : mount: mount passed Jul 7 00:05:14.127264 ignition[913]: INFO : Ignition finished successfully Jul 7 00:05:14.127043 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jul 7 00:05:14.133373 systemd[1]: Starting ignition-files.service - Ignition (files)... Jul 7 00:05:14.559631 systemd-networkd[765]: eth0: Gained IPv6LL Jul 7 00:05:14.862404 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jul 7 00:05:14.876294 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/sda6 scanned by mount (924) Jul 7 00:05:14.876330 kernel: BTRFS info (device sda6): first mount of filesystem 11f56a79-b29d-47db-ad8e-56effe5ac41b Jul 7 00:05:14.880641 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 7 00:05:14.880668 kernel: BTRFS info (device sda6): using free space tree Jul 7 00:05:14.887721 kernel: BTRFS info (device sda6): enabling ssd optimizations Jul 7 00:05:14.887742 kernel: BTRFS info (device sda6): auto enabling async discard Jul 7 00:05:14.891842 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jul 7 00:05:14.917025 ignition[941]: INFO : Ignition 2.19.0 Jul 7 00:05:14.917025 ignition[941]: INFO : Stage: files Jul 7 00:05:14.918454 ignition[941]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 7 00:05:14.918454 ignition[941]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jul 7 00:05:14.918454 ignition[941]: DEBUG : files: compiled without relabeling support, skipping Jul 7 00:05:14.923389 ignition[941]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jul 7 00:05:14.923389 ignition[941]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jul 7 00:05:14.923389 ignition[941]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jul 7 00:05:14.923389 ignition[941]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jul 7 00:05:14.926704 ignition[941]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jul 7 00:05:14.926704 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jul 7 00:05:14.926704 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Jul 7 00:05:14.923527 unknown[941]: wrote ssh authorized keys file for user: core Jul 7 00:05:15.273064 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jul 7 00:05:16.251853 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jul 7 00:05:16.251853 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Jul 7 00:05:16.254863 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-x86-64.raw: attempt #1 Jul 7 00:05:16.793557 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jul 7 00:05:17.091590 ignition[941]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Jul 7 00:05:17.091590 ignition[941]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(d): op(e): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(f): [started] setting preset to enabled for "prepare-helm.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: op(f): [finished] setting preset to enabled for "prepare-helm.service" Jul 7 00:05:17.094666 ignition[941]: INFO : files: createResultFile: createFiles: op(10): [started] writing file "/sysroot/etc/.ignition-result.json" Jul 7 00:05:17.094666 ignition[941]: INFO : files: createResultFile: createFiles: op(10): [finished] writing file "/sysroot/etc/.ignition-result.json" Jul 7 00:05:17.094666 ignition[941]: INFO : files: files passed Jul 7 00:05:17.094666 ignition[941]: INFO : Ignition finished successfully Jul 7 00:05:17.097728 systemd[1]: Finished ignition-files.service - Ignition (files). Jul 7 00:05:17.129504 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jul 7 00:05:17.135401 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jul 7 00:05:17.137118 systemd[1]: ignition-quench.service: Deactivated successfully. Jul 7 00:05:17.137230 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jul 7 00:05:17.152513 initrd-setup-root-after-ignition[969]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 7 00:05:17.152513 initrd-setup-root-after-ignition[969]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jul 7 00:05:17.154766 initrd-setup-root-after-ignition[973]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 7 00:05:17.156175 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jul 7 00:05:17.157016 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jul 7 00:05:17.162406 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jul 7 00:05:17.193961 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jul 7 00:05:17.194109 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jul 7 00:05:17.196138 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jul 7 00:05:17.198097 systemd[1]: Reached target initrd.target - Initrd Default Target. Jul 7 00:05:17.199060 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jul 7 00:05:17.206478 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jul 7 00:05:17.223025 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jul 7 00:05:17.230423 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jul 7 00:05:17.240038 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jul 7 00:05:17.241585 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 7 00:05:17.243100 systemd[1]: Stopped target timers.target - Timer Units. Jul 7 00:05:17.244648 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jul 7 00:05:17.244776 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jul 7 00:05:17.246980 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jul 7 00:05:17.248028 systemd[1]: Stopped target basic.target - Basic System. Jul 7 00:05:17.249222 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jul 7 00:05:17.250370 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jul 7 00:05:17.251695 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jul 7 00:05:17.253415 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jul 7 00:05:17.254878 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jul 7 00:05:17.256166 systemd[1]: Stopped target sysinit.target - System Initialization. Jul 7 00:05:17.257674 systemd[1]: Stopped target local-fs.target - Local File Systems. Jul 7 00:05:17.258930 systemd[1]: Stopped target swap.target - Swaps. Jul 7 00:05:17.260170 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jul 7 00:05:17.260494 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jul 7 00:05:17.261978 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jul 7 00:05:17.262899 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 7 00:05:17.264645 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jul 7 00:05:17.264754 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 7 00:05:17.266137 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jul 7 00:05:17.266236 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jul 7 00:05:17.267818 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jul 7 00:05:17.267927 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jul 7 00:05:17.268881 systemd[1]: ignition-files.service: Deactivated successfully. Jul 7 00:05:17.269050 systemd[1]: Stopped ignition-files.service - Ignition (files). Jul 7 00:05:17.279747 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jul 7 00:05:17.280325 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jul 7 00:05:17.280478 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jul 7 00:05:17.283474 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jul 7 00:05:17.284018 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jul 7 00:05:17.284165 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jul 7 00:05:17.284874 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jul 7 00:05:17.285008 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jul 7 00:05:17.296805 ignition[993]: INFO : Ignition 2.19.0 Jul 7 00:05:17.296805 ignition[993]: INFO : Stage: umount Jul 7 00:05:17.299701 ignition[993]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 7 00:05:17.299701 ignition[993]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/akamai" Jul 7 00:05:17.297711 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jul 7 00:05:17.305030 ignition[993]: INFO : umount: umount passed Jul 7 00:05:17.305030 ignition[993]: INFO : Ignition finished successfully Jul 7 00:05:17.297820 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jul 7 00:05:17.307544 systemd[1]: ignition-mount.service: Deactivated successfully. Jul 7 00:05:17.307665 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jul 7 00:05:17.309544 systemd[1]: ignition-disks.service: Deactivated successfully. Jul 7 00:05:17.309630 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jul 7 00:05:17.313562 systemd[1]: ignition-kargs.service: Deactivated successfully. Jul 7 00:05:17.313613 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jul 7 00:05:17.314479 systemd[1]: ignition-fetch.service: Deactivated successfully. Jul 7 00:05:17.314526 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Jul 7 00:05:17.315052 systemd[1]: Stopped target network.target - Network. Jul 7 00:05:17.315613 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jul 7 00:05:17.315666 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jul 7 00:05:17.317614 systemd[1]: Stopped target paths.target - Path Units. Jul 7 00:05:17.318914 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jul 7 00:05:17.319725 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 7 00:05:17.320324 systemd[1]: Stopped target slices.target - Slice Units. Jul 7 00:05:17.321475 systemd[1]: Stopped target sockets.target - Socket Units. Jul 7 00:05:17.323031 systemd[1]: iscsid.socket: Deactivated successfully. Jul 7 00:05:17.323106 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jul 7 00:05:17.346050 systemd[1]: iscsiuio.socket: Deactivated successfully. Jul 7 00:05:17.346112 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jul 7 00:05:17.347316 systemd[1]: ignition-setup.service: Deactivated successfully. Jul 7 00:05:17.347382 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jul 7 00:05:17.348439 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jul 7 00:05:17.348489 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jul 7 00:05:17.349858 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jul 7 00:05:17.351352 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jul 7 00:05:17.353519 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jul 7 00:05:17.354182 systemd[1]: sysroot-boot.service: Deactivated successfully. Jul 7 00:05:17.354336 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jul 7 00:05:17.355363 systemd-networkd[765]: eth0: DHCPv6 lease lost Jul 7 00:05:17.356031 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jul 7 00:05:17.356121 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jul 7 00:05:17.358221 systemd[1]: systemd-networkd.service: Deactivated successfully. Jul 7 00:05:17.358410 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jul 7 00:05:17.362380 systemd[1]: systemd-resolved.service: Deactivated successfully. Jul 7 00:05:17.362518 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jul 7 00:05:17.366848 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jul 7 00:05:17.366910 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jul 7 00:05:17.373396 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jul 7 00:05:17.374630 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jul 7 00:05:17.374690 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jul 7 00:05:17.376962 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jul 7 00:05:17.377017 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jul 7 00:05:17.378802 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jul 7 00:05:17.378852 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jul 7 00:05:17.379894 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jul 7 00:05:17.379941 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 7 00:05:17.382974 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 7 00:05:17.404991 systemd[1]: systemd-udevd.service: Deactivated successfully. Jul 7 00:05:17.405179 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 7 00:05:17.407008 systemd[1]: network-cleanup.service: Deactivated successfully. Jul 7 00:05:17.407111 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jul 7 00:05:17.409592 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jul 7 00:05:17.409667 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jul 7 00:05:17.410563 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jul 7 00:05:17.410603 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jul 7 00:05:17.411758 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jul 7 00:05:17.411809 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jul 7 00:05:17.413696 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jul 7 00:05:17.413743 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jul 7 00:05:17.414934 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jul 7 00:05:17.414983 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 7 00:05:17.422501 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jul 7 00:05:17.424010 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jul 7 00:05:17.424090 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 7 00:05:17.425493 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 7 00:05:17.425545 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 7 00:05:17.429137 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jul 7 00:05:17.429299 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jul 7 00:05:17.430420 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jul 7 00:05:17.443395 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jul 7 00:05:17.450744 systemd[1]: Switching root. Jul 7 00:05:17.482914 systemd-journald[177]: Journal stopped Jul 7 00:05:18.563417 systemd-journald[177]: Received SIGTERM from PID 1 (systemd). Jul 7 00:05:18.563444 kernel: SELinux: policy capability network_peer_controls=1 Jul 7 00:05:18.563456 kernel: SELinux: policy capability open_perms=1 Jul 7 00:05:18.563465 kernel: SELinux: policy capability extended_socket_class=1 Jul 7 00:05:18.563473 kernel: SELinux: policy capability always_check_network=0 Jul 7 00:05:18.563486 kernel: SELinux: policy capability cgroup_seclabel=1 Jul 7 00:05:18.563495 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jul 7 00:05:18.563504 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jul 7 00:05:18.563512 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jul 7 00:05:18.563521 kernel: audit: type=1403 audit(1751846717.622:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jul 7 00:05:18.563531 systemd[1]: Successfully loaded SELinux policy in 45.694ms. Jul 7 00:05:18.563544 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.238ms. Jul 7 00:05:18.563554 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jul 7 00:05:18.563564 systemd[1]: Detected virtualization kvm. Jul 7 00:05:18.563574 systemd[1]: Detected architecture x86-64. Jul 7 00:05:18.563584 systemd[1]: Detected first boot. Jul 7 00:05:18.563596 systemd[1]: Initializing machine ID from random generator. Jul 7 00:05:18.563606 zram_generator::config[1035]: No configuration found. Jul 7 00:05:18.563616 systemd[1]: Populated /etc with preset unit settings. Jul 7 00:05:18.563625 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jul 7 00:05:18.563635 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jul 7 00:05:18.563645 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jul 7 00:05:18.563656 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jul 7 00:05:18.563669 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jul 7 00:05:18.563678 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jul 7 00:05:18.563688 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jul 7 00:05:18.563698 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jul 7 00:05:18.563708 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jul 7 00:05:18.563718 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jul 7 00:05:18.563728 systemd[1]: Created slice user.slice - User and Session Slice. Jul 7 00:05:18.563740 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 7 00:05:18.563751 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 7 00:05:18.563767 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jul 7 00:05:18.563784 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jul 7 00:05:18.563800 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jul 7 00:05:18.563817 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jul 7 00:05:18.563833 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jul 7 00:05:18.563850 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 7 00:05:18.563869 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jul 7 00:05:18.563879 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jul 7 00:05:18.563892 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jul 7 00:05:18.563902 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jul 7 00:05:18.563912 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 7 00:05:18.563922 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jul 7 00:05:18.563931 systemd[1]: Reached target slices.target - Slice Units. Jul 7 00:05:18.563941 systemd[1]: Reached target swap.target - Swaps. Jul 7 00:05:18.563953 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jul 7 00:05:18.563963 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jul 7 00:05:18.563973 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jul 7 00:05:18.563983 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jul 7 00:05:18.563992 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jul 7 00:05:18.564003 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jul 7 00:05:18.564015 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jul 7 00:05:18.564025 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jul 7 00:05:18.564035 systemd[1]: Mounting media.mount - External Media Directory... Jul 7 00:05:18.564045 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 7 00:05:18.564055 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jul 7 00:05:18.564065 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jul 7 00:05:18.564075 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jul 7 00:05:18.564087 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jul 7 00:05:18.564097 systemd[1]: Reached target machines.target - Containers. Jul 7 00:05:18.564108 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jul 7 00:05:18.564118 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 7 00:05:18.564128 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jul 7 00:05:18.564139 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jul 7 00:05:18.564149 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 7 00:05:18.564161 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jul 7 00:05:18.564180 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 7 00:05:18.564197 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jul 7 00:05:18.564207 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 7 00:05:18.564218 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jul 7 00:05:18.564228 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jul 7 00:05:18.564238 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jul 7 00:05:18.564264 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jul 7 00:05:18.564275 systemd[1]: Stopped systemd-fsck-usr.service. Jul 7 00:05:18.564289 systemd[1]: Starting systemd-journald.service - Journal Service... Jul 7 00:05:18.564299 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jul 7 00:05:18.564309 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jul 7 00:05:18.564319 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jul 7 00:05:18.564329 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jul 7 00:05:18.564339 systemd[1]: verity-setup.service: Deactivated successfully. Jul 7 00:05:18.564351 systemd[1]: Stopped verity-setup.service. Jul 7 00:05:18.564362 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 7 00:05:18.564391 systemd-journald[1125]: Collecting audit messages is disabled. Jul 7 00:05:18.564416 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jul 7 00:05:18.564427 kernel: ACPI: bus type drm_connector registered Jul 7 00:05:18.564436 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jul 7 00:05:18.564446 systemd[1]: Mounted media.mount - External Media Directory. Jul 7 00:05:18.564459 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jul 7 00:05:18.564469 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jul 7 00:05:18.564479 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jul 7 00:05:18.564488 kernel: fuse: init (API version 7.39) Jul 7 00:05:18.564498 systemd-journald[1125]: Journal started Jul 7 00:05:18.564516 systemd-journald[1125]: Runtime Journal (/run/log/journal/0419746ad2204a0484f67306a431666a) is 8.0M, max 78.3M, 70.3M free. Jul 7 00:05:18.564583 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jul 7 00:05:18.193327 systemd[1]: Queued start job for default target multi-user.target. Jul 7 00:05:18.219562 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Jul 7 00:05:18.220287 systemd[1]: systemd-journald.service: Deactivated successfully. Jul 7 00:05:18.569459 systemd[1]: Started systemd-journald.service - Journal Service. Jul 7 00:05:18.570593 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jul 7 00:05:18.571642 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jul 7 00:05:18.571820 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jul 7 00:05:18.572886 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 7 00:05:18.573059 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 7 00:05:18.574271 kernel: loop: module loaded Jul 7 00:05:18.574843 systemd[1]: modprobe@drm.service: Deactivated successfully. Jul 7 00:05:18.575012 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jul 7 00:05:18.576066 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 7 00:05:18.576236 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 7 00:05:18.577149 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jul 7 00:05:18.577680 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jul 7 00:05:18.578586 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 7 00:05:18.578741 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 7 00:05:18.579720 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jul 7 00:05:18.580600 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jul 7 00:05:18.581529 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jul 7 00:05:18.596195 systemd[1]: Reached target network-pre.target - Preparation for Network. Jul 7 00:05:18.626791 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jul 7 00:05:18.631294 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jul 7 00:05:18.632421 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jul 7 00:05:18.633295 systemd[1]: Reached target local-fs.target - Local File Systems. Jul 7 00:05:18.635415 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Jul 7 00:05:18.648410 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jul 7 00:05:18.654345 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jul 7 00:05:18.655032 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 7 00:05:18.657920 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jul 7 00:05:18.661105 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jul 7 00:05:18.662394 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 7 00:05:18.668472 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jul 7 00:05:18.669114 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jul 7 00:05:18.671349 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 7 00:05:18.679570 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jul 7 00:05:18.682804 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jul 7 00:05:18.687039 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jul 7 00:05:18.689995 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jul 7 00:05:18.694984 systemd-journald[1125]: Time spent on flushing to /var/log/journal/0419746ad2204a0484f67306a431666a is 30.610ms for 971 entries. Jul 7 00:05:18.694984 systemd-journald[1125]: System Journal (/var/log/journal/0419746ad2204a0484f67306a431666a) is 8.0M, max 195.6M, 187.6M free. Jul 7 00:05:18.747944 systemd-journald[1125]: Received client request to flush runtime journal. Jul 7 00:05:18.748065 kernel: loop0: detected capacity change from 0 to 142488 Jul 7 00:05:18.705362 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jul 7 00:05:18.706280 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jul 7 00:05:18.722876 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Jul 7 00:05:18.725709 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jul 7 00:05:18.728607 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jul 7 00:05:18.740349 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Jul 7 00:05:18.751728 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jul 7 00:05:18.773841 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jul 7 00:05:18.775288 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 7 00:05:18.786979 udevadm[1163]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Jul 7 00:05:18.790368 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jul 7 00:05:18.793002 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Jul 7 00:05:18.801690 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jul 7 00:05:18.802283 kernel: loop1: detected capacity change from 0 to 8 Jul 7 00:05:18.812984 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jul 7 00:05:18.833395 kernel: loop2: detected capacity change from 0 to 140768 Jul 7 00:05:18.834628 systemd-tmpfiles[1174]: ACLs are not supported, ignoring. Jul 7 00:05:18.835816 systemd-tmpfiles[1174]: ACLs are not supported, ignoring. Jul 7 00:05:18.844716 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 7 00:05:18.883402 kernel: loop3: detected capacity change from 0 to 221472 Jul 7 00:05:18.923341 kernel: loop4: detected capacity change from 0 to 142488 Jul 7 00:05:18.952453 kernel: loop5: detected capacity change from 0 to 8 Jul 7 00:05:18.958241 kernel: loop6: detected capacity change from 0 to 140768 Jul 7 00:05:18.986298 kernel: loop7: detected capacity change from 0 to 221472 Jul 7 00:05:19.006519 (sd-merge)[1180]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-akamai'. Jul 7 00:05:19.009294 (sd-merge)[1180]: Merged extensions into '/usr'. Jul 7 00:05:19.019541 systemd[1]: Reloading requested from client PID 1155 ('systemd-sysext') (unit systemd-sysext.service)... Jul 7 00:05:19.019641 systemd[1]: Reloading... Jul 7 00:05:19.130287 zram_generator::config[1206]: No configuration found. Jul 7 00:05:19.250641 ldconfig[1150]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jul 7 00:05:19.261820 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 7 00:05:19.305696 systemd[1]: Reloading finished in 285 ms. Jul 7 00:05:19.332718 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jul 7 00:05:19.333712 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jul 7 00:05:19.347921 systemd[1]: Starting ensure-sysext.service... Jul 7 00:05:19.352393 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jul 7 00:05:19.360948 systemd[1]: Reloading requested from client PID 1249 ('systemctl') (unit ensure-sysext.service)... Jul 7 00:05:19.360961 systemd[1]: Reloading... Jul 7 00:05:19.386111 systemd-tmpfiles[1250]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jul 7 00:05:19.386717 systemd-tmpfiles[1250]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jul 7 00:05:19.387883 systemd-tmpfiles[1250]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jul 7 00:05:19.388210 systemd-tmpfiles[1250]: ACLs are not supported, ignoring. Jul 7 00:05:19.388403 systemd-tmpfiles[1250]: ACLs are not supported, ignoring. Jul 7 00:05:19.391828 systemd-tmpfiles[1250]: Detected autofs mount point /boot during canonicalization of boot. Jul 7 00:05:19.391908 systemd-tmpfiles[1250]: Skipping /boot Jul 7 00:05:19.407135 systemd-tmpfiles[1250]: Detected autofs mount point /boot during canonicalization of boot. Jul 7 00:05:19.407196 systemd-tmpfiles[1250]: Skipping /boot Jul 7 00:05:19.442345 zram_generator::config[1279]: No configuration found. Jul 7 00:05:19.549218 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 7 00:05:19.589739 systemd[1]: Reloading finished in 228 ms. Jul 7 00:05:19.608069 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jul 7 00:05:19.617115 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 7 00:05:19.628504 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Jul 7 00:05:19.634818 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jul 7 00:05:19.639454 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jul 7 00:05:19.651744 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jul 7 00:05:19.656512 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 7 00:05:19.662471 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jul 7 00:05:19.666386 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 7 00:05:19.666544 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 7 00:05:19.674717 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 7 00:05:19.683396 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 7 00:05:19.692866 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 7 00:05:19.694706 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 7 00:05:19.702325 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jul 7 00:05:19.703102 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 7 00:05:19.704727 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 7 00:05:19.706307 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 7 00:05:19.712463 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 7 00:05:19.713467 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 7 00:05:19.720196 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 7 00:05:19.722386 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 7 00:05:19.722484 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 7 00:05:19.724322 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jul 7 00:05:19.725379 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 7 00:05:19.726079 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 7 00:05:19.727225 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 7 00:05:19.727831 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 7 00:05:19.729589 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 7 00:05:19.729966 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 7 00:05:19.745942 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 7 00:05:19.746177 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 7 00:05:19.748037 systemd-udevd[1332]: Using default interface naming scheme 'v255'. Jul 7 00:05:19.756444 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 7 00:05:19.768621 augenrules[1353]: No rules Jul 7 00:05:19.772136 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jul 7 00:05:19.776129 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 7 00:05:19.786466 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 7 00:05:19.787124 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 7 00:05:19.791512 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jul 7 00:05:19.792081 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 7 00:05:19.793076 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jul 7 00:05:19.795200 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Jul 7 00:05:19.797291 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jul 7 00:05:19.798471 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 7 00:05:19.799790 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 7 00:05:19.800814 systemd[1]: modprobe@drm.service: Deactivated successfully. Jul 7 00:05:19.800978 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jul 7 00:05:19.802546 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 7 00:05:19.803813 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 7 00:05:19.804918 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 7 00:05:19.817964 systemd[1]: Finished ensure-sysext.service. Jul 7 00:05:19.822336 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 7 00:05:19.822698 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 7 00:05:19.835466 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jul 7 00:05:19.836134 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 7 00:05:19.836280 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jul 7 00:05:19.847379 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jul 7 00:05:19.848330 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jul 7 00:05:19.851811 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jul 7 00:05:19.865651 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jul 7 00:05:19.917957 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Jul 7 00:05:19.978533 systemd-networkd[1382]: lo: Link UP Jul 7 00:05:19.980360 systemd-networkd[1382]: lo: Gained carrier Jul 7 00:05:19.985978 systemd-networkd[1382]: Enumeration completed Jul 7 00:05:19.986203 systemd[1]: Started systemd-networkd.service - Network Configuration. Jul 7 00:05:19.986728 systemd-networkd[1382]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 7 00:05:19.988310 systemd-networkd[1382]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jul 7 00:05:19.991712 systemd-networkd[1382]: eth0: Link UP Jul 7 00:05:19.993058 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jul 7 00:05:19.993165 systemd-networkd[1382]: eth0: Gained carrier Jul 7 00:05:19.993204 systemd-networkd[1382]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 7 00:05:19.996155 systemd-resolved[1329]: Positive Trust Anchors: Jul 7 00:05:19.997293 systemd-resolved[1329]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 7 00:05:19.997326 systemd-resolved[1329]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jul 7 00:05:20.005376 systemd-resolved[1329]: Defaulting to hostname 'linux'. Jul 7 00:05:20.007326 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jul 7 00:05:20.008374 systemd[1]: Reached target network.target - Network. Jul 7 00:05:20.009089 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jul 7 00:05:20.013474 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jul 7 00:05:20.014097 systemd[1]: Reached target time-set.target - System Time Set. Jul 7 00:05:20.026489 systemd-networkd[1382]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 7 00:05:20.036282 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 34 scanned by (udev-worker) (1386) Jul 7 00:05:20.048271 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Jul 7 00:05:20.062347 kernel: ACPI: button: Power Button [PWRF] Jul 7 00:05:20.092584 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Jul 7 00:05:20.092851 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Jul 7 00:05:20.093032 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Jul 7 00:05:20.099668 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Jul 7 00:05:20.107629 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jul 7 00:05:20.116346 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Jul 7 00:05:20.128115 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jul 7 00:05:20.134328 kernel: EDAC MC: Ver: 3.0.0 Jul 7 00:05:20.145324 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 7 00:05:20.149122 kernel: mousedev: PS/2 mouse device common for all mice Jul 7 00:05:20.174693 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Jul 7 00:05:20.184087 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Jul 7 00:05:20.211272 lvm[1421]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jul 7 00:05:20.234680 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Jul 7 00:05:20.269821 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 7 00:05:20.271658 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jul 7 00:05:20.272247 systemd[1]: Reached target sysinit.target - System Initialization. Jul 7 00:05:20.272951 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jul 7 00:05:20.273967 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jul 7 00:05:20.274795 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jul 7 00:05:20.275592 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jul 7 00:05:20.276202 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jul 7 00:05:20.276801 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jul 7 00:05:20.276838 systemd[1]: Reached target paths.target - Path Units. Jul 7 00:05:20.277367 systemd[1]: Reached target timers.target - Timer Units. Jul 7 00:05:20.278815 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jul 7 00:05:20.280973 systemd[1]: Starting docker.socket - Docker Socket for the API... Jul 7 00:05:20.286508 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jul 7 00:05:20.288289 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Jul 7 00:05:20.289414 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jul 7 00:05:20.290022 systemd[1]: Reached target sockets.target - Socket Units. Jul 7 00:05:20.290574 systemd[1]: Reached target basic.target - Basic System. Jul 7 00:05:20.291118 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jul 7 00:05:20.291158 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jul 7 00:05:20.292385 systemd[1]: Starting containerd.service - containerd container runtime... Jul 7 00:05:20.298481 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Jul 7 00:05:20.299519 lvm[1428]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jul 7 00:05:20.307410 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jul 7 00:05:20.311801 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jul 7 00:05:20.317383 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jul 7 00:05:20.317944 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jul 7 00:05:20.323411 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jul 7 00:05:20.326818 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jul 7 00:05:20.329441 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jul 7 00:05:20.340406 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jul 7 00:05:20.347496 jq[1432]: false Jul 7 00:05:20.351410 systemd[1]: Starting systemd-logind.service - User Login Management... Jul 7 00:05:20.353446 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jul 7 00:05:20.353884 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jul 7 00:05:20.359703 systemd[1]: Starting update-engine.service - Update Engine... Jul 7 00:05:20.366383 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jul 7 00:05:20.368958 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Jul 7 00:05:20.376649 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jul 7 00:05:20.377377 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jul 7 00:05:20.392002 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jul 7 00:05:20.392238 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jul 7 00:05:20.399890 extend-filesystems[1433]: Found loop4 Jul 7 00:05:20.401235 extend-filesystems[1433]: Found loop5 Jul 7 00:05:20.401235 extend-filesystems[1433]: Found loop6 Jul 7 00:05:20.401235 extend-filesystems[1433]: Found loop7 Jul 7 00:05:20.401235 extend-filesystems[1433]: Found sda Jul 7 00:05:20.401235 extend-filesystems[1433]: Found sda1 Jul 7 00:05:20.401235 extend-filesystems[1433]: Found sda2 Jul 7 00:05:20.401235 extend-filesystems[1433]: Found sda3 Jul 7 00:05:20.401235 extend-filesystems[1433]: Found usr Jul 7 00:05:20.401235 extend-filesystems[1433]: Found sda4 Jul 7 00:05:20.401235 extend-filesystems[1433]: Found sda6 Jul 7 00:05:20.401235 extend-filesystems[1433]: Found sda7 Jul 7 00:05:20.401235 extend-filesystems[1433]: Found sda9 Jul 7 00:05:20.401235 extend-filesystems[1433]: Checking size of /dev/sda9 Jul 7 00:05:20.454715 extend-filesystems[1433]: Resized partition /dev/sda9 Jul 7 00:05:20.458438 coreos-metadata[1430]: Jul 07 00:05:20.450 INFO Putting http://169.254.169.254/v1/token: Attempt #1 Jul 7 00:05:20.434821 dbus-daemon[1431]: [system] SELinux support is enabled Jul 7 00:05:20.434166 (ntainerd)[1456]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jul 7 00:05:20.462955 update_engine[1441]: I20250707 00:05:20.427057 1441 main.cc:92] Flatcar Update Engine starting Jul 7 00:05:20.462955 update_engine[1441]: I20250707 00:05:20.442675 1441 update_check_scheduler.cc:74] Next update check in 5m22s Jul 7 00:05:20.434980 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jul 7 00:05:20.438701 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jul 7 00:05:20.479475 extend-filesystems[1468]: resize2fs 1.47.1 (20-May-2024) Jul 7 00:05:20.481496 jq[1442]: true Jul 7 00:05:20.486345 kernel: EXT4-fs (sda9): resizing filesystem from 553472 to 20360187 blocks Jul 7 00:05:20.438732 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jul 7 00:05:20.486472 tar[1445]: linux-amd64/helm Jul 7 00:05:20.442856 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jul 7 00:05:20.487867 dbus-daemon[1431]: [system] Activating via systemd: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.3' (uid=244 pid=1382 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Jul 7 00:05:20.442873 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jul 7 00:05:20.493382 jq[1459]: true Jul 7 00:05:20.456525 systemd-networkd[1382]: eth0: DHCPv4 address 172.234.31.125/24, gateway 172.234.31.1 acquired from 23.205.167.142 Jul 7 00:05:20.459651 systemd[1]: Started update-engine.service - Update Engine. Jul 7 00:05:20.462359 systemd-timesyncd[1383]: Network configuration changed, trying to establish connection. Jul 7 00:05:20.472451 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jul 7 00:05:20.497536 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Jul 7 00:05:20.516890 systemd[1]: motdgen.service: Deactivated successfully. Jul 7 00:05:20.517790 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jul 7 00:05:20.556364 systemd-logind[1440]: Watching system buttons on /dev/input/event1 (Power Button) Jul 7 00:05:20.556830 systemd-logind[1440]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jul 7 00:05:20.571291 systemd-logind[1440]: New seat seat0. Jul 7 00:05:20.575294 systemd[1]: Started systemd-logind.service - User Login Management. Jul 7 00:05:20.625593 dbus-daemon[1431]: [system] Successfully activated service 'org.freedesktop.hostname1' Jul 7 00:05:20.625931 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Jul 7 00:05:20.632532 dbus-daemon[1431]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.6' (uid=0 pid=1475 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Jul 7 00:05:20.645049 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 34 scanned by (udev-worker) (1369) Jul 7 00:05:20.644170 systemd[1]: Starting polkit.service - Authorization Manager... Jul 7 00:05:20.674029 bash[1492]: Updated "/home/core/.ssh/authorized_keys" Jul 7 00:05:20.674887 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jul 7 00:05:20.691508 systemd[1]: Starting sshkeys.service... Jul 7 00:05:20.715334 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Jul 7 00:05:20.722510 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Jul 7 00:05:21.154172 systemd-resolved[1329]: Clock change detected. Flushing caches. Jul 7 00:05:21.154351 systemd-timesyncd[1383]: Contacted time server 152.67.232.7:123 (0.flatcar.pool.ntp.org). Jul 7 00:05:21.154418 systemd-timesyncd[1383]: Initial clock synchronization to Mon 2025-07-07 00:05:21.154115 UTC. Jul 7 00:05:21.169926 polkitd[1493]: Started polkitd version 121 Jul 7 00:05:21.196757 polkitd[1493]: Loading rules from directory /etc/polkit-1/rules.d Jul 7 00:05:21.196831 polkitd[1493]: Loading rules from directory /usr/share/polkit-1/rules.d Jul 7 00:05:21.207649 polkitd[1493]: Finished loading, compiling and executing 2 rules Jul 7 00:05:21.212927 dbus-daemon[1431]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Jul 7 00:05:21.214086 systemd[1]: Started polkit.service - Authorization Manager. Jul 7 00:05:21.216005 polkitd[1493]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Jul 7 00:05:21.260776 systemd-hostnamed[1475]: Hostname set to <172-234-31-125> (transient) Jul 7 00:05:21.261275 systemd-resolved[1329]: System hostname changed to '172-234-31-125'. Jul 7 00:05:21.268196 coreos-metadata[1501]: Jul 07 00:05:21.267 INFO Putting http://169.254.169.254/v1/token: Attempt #1 Jul 7 00:05:21.332377 kernel: EXT4-fs (sda9): resized filesystem to 20360187 Jul 7 00:05:21.336421 locksmithd[1472]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jul 7 00:05:21.343000 extend-filesystems[1468]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Jul 7 00:05:21.343000 extend-filesystems[1468]: old_desc_blocks = 1, new_desc_blocks = 10 Jul 7 00:05:21.343000 extend-filesystems[1468]: The filesystem on /dev/sda9 is now 20360187 (4k) blocks long. Jul 7 00:05:21.371061 extend-filesystems[1433]: Resized filesystem in /dev/sda9 Jul 7 00:05:21.373290 containerd[1456]: time="2025-07-07T00:05:21.342383006Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Jul 7 00:05:21.343974 systemd[1]: extend-filesystems.service: Deactivated successfully. Jul 7 00:05:21.344195 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jul 7 00:05:21.394009 coreos-metadata[1501]: Jul 07 00:05:21.393 INFO Fetching http://169.254.169.254/v1/ssh-keys: Attempt #1 Jul 7 00:05:21.408136 containerd[1456]: time="2025-07-07T00:05:21.407897733Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Jul 7 00:05:21.415791 containerd[1456]: time="2025-07-07T00:05:21.415520360Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.95-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Jul 7 00:05:21.415791 containerd[1456]: time="2025-07-07T00:05:21.415567330Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Jul 7 00:05:21.415791 containerd[1456]: time="2025-07-07T00:05:21.415601760Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Jul 7 00:05:21.415884 containerd[1456]: time="2025-07-07T00:05:21.415842379Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Jul 7 00:05:21.415884 containerd[1456]: time="2025-07-07T00:05:21.415874739Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Jul 7 00:05:21.416012 containerd[1456]: time="2025-07-07T00:05:21.415978369Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Jul 7 00:05:21.416035 containerd[1456]: time="2025-07-07T00:05:21.416010409Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Jul 7 00:05:21.416313 containerd[1456]: time="2025-07-07T00:05:21.416279549Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jul 7 00:05:21.416345 containerd[1456]: time="2025-07-07T00:05:21.416312529Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Jul 7 00:05:21.416345 containerd[1456]: time="2025-07-07T00:05:21.416336009Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Jul 7 00:05:21.416394 containerd[1456]: time="2025-07-07T00:05:21.416351799Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Jul 7 00:05:21.416715 containerd[1456]: time="2025-07-07T00:05:21.416499219Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Jul 7 00:05:21.416849 containerd[1456]: time="2025-07-07T00:05:21.416817069Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Jul 7 00:05:21.417035 containerd[1456]: time="2025-07-07T00:05:21.417002389Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jul 7 00:05:21.417069 containerd[1456]: time="2025-07-07T00:05:21.417034339Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Jul 7 00:05:21.417295 containerd[1456]: time="2025-07-07T00:05:21.417165259Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Jul 7 00:05:21.417295 containerd[1456]: time="2025-07-07T00:05:21.417248969Z" level=info msg="metadata content store policy set" policy=shared Jul 7 00:05:21.427169 containerd[1456]: time="2025-07-07T00:05:21.427039704Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Jul 7 00:05:21.427169 containerd[1456]: time="2025-07-07T00:05:21.427081734Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Jul 7 00:05:21.427169 containerd[1456]: time="2025-07-07T00:05:21.427105674Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Jul 7 00:05:21.427169 containerd[1456]: time="2025-07-07T00:05:21.427121854Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Jul 7 00:05:21.427169 containerd[1456]: time="2025-07-07T00:05:21.427135884Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Jul 7 00:05:21.427286 containerd[1456]: time="2025-07-07T00:05:21.427261194Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Jul 7 00:05:21.427565 containerd[1456]: time="2025-07-07T00:05:21.427484274Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Jul 7 00:05:21.427802 containerd[1456]: time="2025-07-07T00:05:21.427599544Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Jul 7 00:05:21.427802 containerd[1456]: time="2025-07-07T00:05:21.427619564Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Jul 7 00:05:21.427802 containerd[1456]: time="2025-07-07T00:05:21.427631494Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Jul 7 00:05:21.427802 containerd[1456]: time="2025-07-07T00:05:21.427644113Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Jul 7 00:05:21.427802 containerd[1456]: time="2025-07-07T00:05:21.427664143Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Jul 7 00:05:21.427802 containerd[1456]: time="2025-07-07T00:05:21.427677933Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Jul 7 00:05:21.427802 containerd[1456]: time="2025-07-07T00:05:21.427690933Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Jul 7 00:05:21.427802 containerd[1456]: time="2025-07-07T00:05:21.427704233Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Jul 7 00:05:21.427802 containerd[1456]: time="2025-07-07T00:05:21.427716783Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Jul 7 00:05:21.427802 containerd[1456]: time="2025-07-07T00:05:21.427728163Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Jul 7 00:05:21.427802 containerd[1456]: time="2025-07-07T00:05:21.427738863Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Jul 7 00:05:21.427802 containerd[1456]: time="2025-07-07T00:05:21.427761773Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.427802 containerd[1456]: time="2025-07-07T00:05:21.427773283Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.427802 containerd[1456]: time="2025-07-07T00:05:21.427784683Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.428055 containerd[1456]: time="2025-07-07T00:05:21.427797113Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.428055 containerd[1456]: time="2025-07-07T00:05:21.427808373Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.428055 containerd[1456]: time="2025-07-07T00:05:21.427820053Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.428055 containerd[1456]: time="2025-07-07T00:05:21.427832583Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.428055 containerd[1456]: time="2025-07-07T00:05:21.427849193Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.428055 containerd[1456]: time="2025-07-07T00:05:21.427861473Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.428055 containerd[1456]: time="2025-07-07T00:05:21.427877223Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.428055 containerd[1456]: time="2025-07-07T00:05:21.427889913Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.428055 containerd[1456]: time="2025-07-07T00:05:21.427900793Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.428055 containerd[1456]: time="2025-07-07T00:05:21.427912393Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.428055 containerd[1456]: time="2025-07-07T00:05:21.427926163Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Jul 7 00:05:21.428055 containerd[1456]: time="2025-07-07T00:05:21.427944143Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.428055 containerd[1456]: time="2025-07-07T00:05:21.427955313Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.428055 containerd[1456]: time="2025-07-07T00:05:21.427965173Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Jul 7 00:05:21.428266 containerd[1456]: time="2025-07-07T00:05:21.428027703Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Jul 7 00:05:21.428266 containerd[1456]: time="2025-07-07T00:05:21.428044833Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Jul 7 00:05:21.428266 containerd[1456]: time="2025-07-07T00:05:21.428055753Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Jul 7 00:05:21.428266 containerd[1456]: time="2025-07-07T00:05:21.428067393Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Jul 7 00:05:21.428266 containerd[1456]: time="2025-07-07T00:05:21.428075923Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.428266 containerd[1456]: time="2025-07-07T00:05:21.428087263Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Jul 7 00:05:21.428266 containerd[1456]: time="2025-07-07T00:05:21.428097073Z" level=info msg="NRI interface is disabled by configuration." Jul 7 00:05:21.428266 containerd[1456]: time="2025-07-07T00:05:21.428107073Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Jul 7 00:05:21.431268 containerd[1456]: time="2025-07-07T00:05:21.428352483Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Jul 7 00:05:21.431268 containerd[1456]: time="2025-07-07T00:05:21.430547142Z" level=info msg="Connect containerd service" Jul 7 00:05:21.431268 containerd[1456]: time="2025-07-07T00:05:21.430581462Z" level=info msg="using legacy CRI server" Jul 7 00:05:21.431268 containerd[1456]: time="2025-07-07T00:05:21.430589822Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jul 7 00:05:21.431268 containerd[1456]: time="2025-07-07T00:05:21.430660242Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Jul 7 00:05:21.433788 containerd[1456]: time="2025-07-07T00:05:21.433755810Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jul 7 00:05:21.434127 containerd[1456]: time="2025-07-07T00:05:21.433889220Z" level=info msg="Start subscribing containerd event" Jul 7 00:05:21.434127 containerd[1456]: time="2025-07-07T00:05:21.433941790Z" level=info msg="Start recovering state" Jul 7 00:05:21.434127 containerd[1456]: time="2025-07-07T00:05:21.434000070Z" level=info msg="Start event monitor" Jul 7 00:05:21.434127 containerd[1456]: time="2025-07-07T00:05:21.434023220Z" level=info msg="Start snapshots syncer" Jul 7 00:05:21.434127 containerd[1456]: time="2025-07-07T00:05:21.434032260Z" level=info msg="Start cni network conf syncer for default" Jul 7 00:05:21.434127 containerd[1456]: time="2025-07-07T00:05:21.434039490Z" level=info msg="Start streaming server" Jul 7 00:05:21.436465 containerd[1456]: time="2025-07-07T00:05:21.436438239Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jul 7 00:05:21.436588 containerd[1456]: time="2025-07-07T00:05:21.436554929Z" level=info msg=serving... address=/run/containerd/containerd.sock Jul 7 00:05:21.436781 systemd[1]: Started containerd.service - containerd container runtime. Jul 7 00:05:21.438687 containerd[1456]: time="2025-07-07T00:05:21.438647258Z" level=info msg="containerd successfully booted in 0.103055s" Jul 7 00:05:21.443450 systemd-networkd[1382]: eth0: Gained IPv6LL Jul 7 00:05:21.446221 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jul 7 00:05:21.449563 systemd[1]: Reached target network-online.target - Network is Online. Jul 7 00:05:21.456559 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 7 00:05:21.464686 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jul 7 00:05:21.516296 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jul 7 00:05:21.553019 sshd_keygen[1467]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jul 7 00:05:21.571783 coreos-metadata[1501]: Jul 07 00:05:21.571 INFO Fetch successful Jul 7 00:05:21.598450 update-ssh-keys[1541]: Updated "/home/core/.ssh/authorized_keys" Jul 7 00:05:21.599156 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Jul 7 00:05:21.607052 systemd[1]: Finished sshkeys.service. Jul 7 00:05:21.608222 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jul 7 00:05:21.616573 systemd[1]: Starting issuegen.service - Generate /run/issue... Jul 7 00:05:21.625242 systemd[1]: issuegen.service: Deactivated successfully. Jul 7 00:05:21.625487 systemd[1]: Finished issuegen.service - Generate /run/issue. Jul 7 00:05:21.637444 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jul 7 00:05:21.649858 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jul 7 00:05:21.659832 systemd[1]: Started getty@tty1.service - Getty on tty1. Jul 7 00:05:21.667687 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jul 7 00:05:21.668719 systemd[1]: Reached target getty.target - Login Prompts. Jul 7 00:05:21.762733 tar[1445]: linux-amd64/LICENSE Jul 7 00:05:21.762733 tar[1445]: linux-amd64/README.md Jul 7 00:05:21.775653 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jul 7 00:05:21.881916 coreos-metadata[1430]: Jul 07 00:05:21.881 INFO Putting http://169.254.169.254/v1/token: Attempt #2 Jul 7 00:05:21.972034 coreos-metadata[1430]: Jul 07 00:05:21.971 INFO Fetching http://169.254.169.254/v1/instance: Attempt #1 Jul 7 00:05:22.155118 coreos-metadata[1430]: Jul 07 00:05:22.154 INFO Fetch successful Jul 7 00:05:22.155118 coreos-metadata[1430]: Jul 07 00:05:22.154 INFO Fetching http://169.254.169.254/v1/network: Attempt #1 Jul 7 00:05:22.414477 coreos-metadata[1430]: Jul 07 00:05:22.414 INFO Fetch successful Jul 7 00:05:22.468606 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 7 00:05:22.475897 (kubelet)[1577]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 7 00:05:22.492276 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Jul 7 00:05:22.493847 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jul 7 00:05:22.493968 systemd[1]: Reached target multi-user.target - Multi-User System. Jul 7 00:05:22.496192 systemd[1]: Startup finished in 863ms (kernel) + 7.927s (initrd) + 4.499s (userspace) = 13.290s. Jul 7 00:05:23.017177 kubelet[1577]: E0707 00:05:23.017096 1577 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 7 00:05:23.021959 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 7 00:05:23.022187 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 7 00:05:24.038089 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jul 7 00:05:24.043573 systemd[1]: Started sshd@0-172.234.31.125:22-147.75.109.163:40960.service - OpenSSH per-connection server daemon (147.75.109.163:40960). Jul 7 00:05:24.416061 sshd[1597]: Accepted publickey for core from 147.75.109.163 port 40960 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:05:24.417983 sshd[1597]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:05:24.425909 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jul 7 00:05:24.430705 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jul 7 00:05:24.432502 systemd-logind[1440]: New session 1 of user core. Jul 7 00:05:24.444595 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jul 7 00:05:24.454966 systemd[1]: Starting user@500.service - User Manager for UID 500... Jul 7 00:05:24.469221 (systemd)[1601]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jul 7 00:05:24.567688 systemd[1601]: Queued start job for default target default.target. Jul 7 00:05:24.582644 systemd[1601]: Created slice app.slice - User Application Slice. Jul 7 00:05:24.582671 systemd[1601]: Reached target paths.target - Paths. Jul 7 00:05:24.582684 systemd[1601]: Reached target timers.target - Timers. Jul 7 00:05:24.584192 systemd[1601]: Starting dbus.socket - D-Bus User Message Bus Socket... Jul 7 00:05:24.604183 systemd[1601]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jul 7 00:05:24.604318 systemd[1601]: Reached target sockets.target - Sockets. Jul 7 00:05:24.604333 systemd[1601]: Reached target basic.target - Basic System. Jul 7 00:05:24.604399 systemd[1601]: Reached target default.target - Main User Target. Jul 7 00:05:24.604435 systemd[1601]: Startup finished in 127ms. Jul 7 00:05:24.604558 systemd[1]: Started user@500.service - User Manager for UID 500. Jul 7 00:05:24.612593 systemd[1]: Started session-1.scope - Session 1 of User core. Jul 7 00:05:24.896353 systemd[1]: Started sshd@1-172.234.31.125:22-147.75.109.163:40976.service - OpenSSH per-connection server daemon (147.75.109.163:40976). Jul 7 00:05:25.264689 sshd[1612]: Accepted publickey for core from 147.75.109.163 port 40976 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:05:25.266503 sshd[1612]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:05:25.285519 systemd-logind[1440]: New session 2 of user core. Jul 7 00:05:25.296550 systemd[1]: Started session-2.scope - Session 2 of User core. Jul 7 00:05:25.541170 sshd[1612]: pam_unix(sshd:session): session closed for user core Jul 7 00:05:25.545237 systemd-logind[1440]: Session 2 logged out. Waiting for processes to exit. Jul 7 00:05:25.546083 systemd[1]: sshd@1-172.234.31.125:22-147.75.109.163:40976.service: Deactivated successfully. Jul 7 00:05:25.548268 systemd[1]: session-2.scope: Deactivated successfully. Jul 7 00:05:25.549012 systemd-logind[1440]: Removed session 2. Jul 7 00:05:25.611383 systemd[1]: Started sshd@2-172.234.31.125:22-147.75.109.163:40980.service - OpenSSH per-connection server daemon (147.75.109.163:40980). Jul 7 00:05:25.992980 sshd[1619]: Accepted publickey for core from 147.75.109.163 port 40980 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:05:25.995180 sshd[1619]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:05:26.011780 systemd-logind[1440]: New session 3 of user core. Jul 7 00:05:26.015512 systemd[1]: Started session-3.scope - Session 3 of User core. Jul 7 00:05:26.263567 sshd[1619]: pam_unix(sshd:session): session closed for user core Jul 7 00:05:26.267424 systemd[1]: sshd@2-172.234.31.125:22-147.75.109.163:40980.service: Deactivated successfully. Jul 7 00:05:26.269070 systemd[1]: session-3.scope: Deactivated successfully. Jul 7 00:05:26.269690 systemd-logind[1440]: Session 3 logged out. Waiting for processes to exit. Jul 7 00:05:26.270500 systemd-logind[1440]: Removed session 3. Jul 7 00:05:26.325565 systemd[1]: Started sshd@3-172.234.31.125:22-147.75.109.163:51876.service - OpenSSH per-connection server daemon (147.75.109.163:51876). Jul 7 00:05:26.685104 sshd[1626]: Accepted publickey for core from 147.75.109.163 port 51876 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:05:26.686460 sshd[1626]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:05:26.691620 systemd-logind[1440]: New session 4 of user core. Jul 7 00:05:26.698510 systemd[1]: Started session-4.scope - Session 4 of User core. Jul 7 00:05:26.956166 sshd[1626]: pam_unix(sshd:session): session closed for user core Jul 7 00:05:26.960123 systemd[1]: sshd@3-172.234.31.125:22-147.75.109.163:51876.service: Deactivated successfully. Jul 7 00:05:26.961738 systemd[1]: session-4.scope: Deactivated successfully. Jul 7 00:05:26.962294 systemd-logind[1440]: Session 4 logged out. Waiting for processes to exit. Jul 7 00:05:26.963103 systemd-logind[1440]: Removed session 4. Jul 7 00:05:27.024547 systemd[1]: Started sshd@4-172.234.31.125:22-147.75.109.163:51892.service - OpenSSH per-connection server daemon (147.75.109.163:51892). Jul 7 00:05:27.392998 sshd[1633]: Accepted publickey for core from 147.75.109.163 port 51892 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:05:27.395172 sshd[1633]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:05:27.399455 systemd-logind[1440]: New session 5 of user core. Jul 7 00:05:27.408483 systemd[1]: Started session-5.scope - Session 5 of User core. Jul 7 00:05:27.618640 sudo[1636]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jul 7 00:05:27.618974 sudo[1636]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 7 00:05:27.633600 sudo[1636]: pam_unix(sudo:session): session closed for user root Jul 7 00:05:27.691990 sshd[1633]: pam_unix(sshd:session): session closed for user core Jul 7 00:05:27.696011 systemd[1]: sshd@4-172.234.31.125:22-147.75.109.163:51892.service: Deactivated successfully. Jul 7 00:05:27.697887 systemd[1]: session-5.scope: Deactivated successfully. Jul 7 00:05:27.699734 systemd-logind[1440]: Session 5 logged out. Waiting for processes to exit. Jul 7 00:05:27.701379 systemd-logind[1440]: Removed session 5. Jul 7 00:05:27.764644 systemd[1]: Started sshd@5-172.234.31.125:22-147.75.109.163:51898.service - OpenSSH per-connection server daemon (147.75.109.163:51898). Jul 7 00:05:28.118099 sshd[1641]: Accepted publickey for core from 147.75.109.163 port 51898 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:05:28.119679 sshd[1641]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:05:28.123434 systemd-logind[1440]: New session 6 of user core. Jul 7 00:05:28.133477 systemd[1]: Started session-6.scope - Session 6 of User core. Jul 7 00:05:28.330922 sudo[1645]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jul 7 00:05:28.331304 sudo[1645]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 7 00:05:28.335887 sudo[1645]: pam_unix(sudo:session): session closed for user root Jul 7 00:05:28.344551 sudo[1644]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Jul 7 00:05:28.344880 sudo[1644]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 7 00:05:28.359614 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Jul 7 00:05:28.361409 auditctl[1648]: No rules Jul 7 00:05:28.361720 systemd[1]: audit-rules.service: Deactivated successfully. Jul 7 00:05:28.361957 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Jul 7 00:05:28.365045 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Jul 7 00:05:28.394147 augenrules[1666]: No rules Jul 7 00:05:28.396020 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Jul 7 00:05:28.397782 sudo[1644]: pam_unix(sudo:session): session closed for user root Jul 7 00:05:28.454155 sshd[1641]: pam_unix(sshd:session): session closed for user core Jul 7 00:05:28.457596 systemd[1]: sshd@5-172.234.31.125:22-147.75.109.163:51898.service: Deactivated successfully. Jul 7 00:05:28.459677 systemd[1]: session-6.scope: Deactivated successfully. Jul 7 00:05:28.461155 systemd-logind[1440]: Session 6 logged out. Waiting for processes to exit. Jul 7 00:05:28.462688 systemd-logind[1440]: Removed session 6. Jul 7 00:05:28.518179 systemd[1]: Started sshd@6-172.234.31.125:22-147.75.109.163:51908.service - OpenSSH per-connection server daemon (147.75.109.163:51908). Jul 7 00:05:28.883222 sshd[1674]: Accepted publickey for core from 147.75.109.163 port 51908 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:05:28.884475 sshd[1674]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:05:28.888714 systemd-logind[1440]: New session 7 of user core. Jul 7 00:05:28.896466 systemd[1]: Started session-7.scope - Session 7 of User core. Jul 7 00:05:29.094634 sudo[1677]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jul 7 00:05:29.094958 sudo[1677]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 7 00:05:29.373574 systemd[1]: Starting docker.service - Docker Application Container Engine... Jul 7 00:05:29.375795 (dockerd)[1692]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jul 7 00:05:29.643818 dockerd[1692]: time="2025-07-07T00:05:29.643154285Z" level=info msg="Starting up" Jul 7 00:05:29.739073 dockerd[1692]: time="2025-07-07T00:05:29.738815847Z" level=info msg="Loading containers: start." Jul 7 00:05:29.848649 kernel: Initializing XFRM netlink socket Jul 7 00:05:29.940984 systemd-networkd[1382]: docker0: Link UP Jul 7 00:05:29.956910 dockerd[1692]: time="2025-07-07T00:05:29.956850358Z" level=info msg="Loading containers: done." Jul 7 00:05:29.971930 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck712038658-merged.mount: Deactivated successfully. Jul 7 00:05:29.973185 dockerd[1692]: time="2025-07-07T00:05:29.973122310Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jul 7 00:05:29.973292 dockerd[1692]: time="2025-07-07T00:05:29.973258780Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Jul 7 00:05:29.973426 dockerd[1692]: time="2025-07-07T00:05:29.973403670Z" level=info msg="Daemon has completed initialization" Jul 7 00:05:30.002435 systemd[1]: Started docker.service - Docker Application Container Engine. Jul 7 00:05:30.002575 dockerd[1692]: time="2025-07-07T00:05:30.002415285Z" level=info msg="API listen on /run/docker.sock" Jul 7 00:05:30.581875 containerd[1456]: time="2025-07-07T00:05:30.581830745Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.10\"" Jul 7 00:05:31.130400 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3032539400.mount: Deactivated successfully. Jul 7 00:05:32.607506 containerd[1456]: time="2025-07-07T00:05:32.605559613Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:32.611168 containerd[1456]: time="2025-07-07T00:05:32.608068502Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.10: active requests=0, bytes read=28077750" Jul 7 00:05:32.611447 containerd[1456]: time="2025-07-07T00:05:32.611416580Z" level=info msg="ImageCreate event name:\"sha256:74c5154ea84d9a53c406e6c00e53cf66145cce821fd80e3c74e2e1bf312f3977\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:32.613810 containerd[1456]: time="2025-07-07T00:05:32.613781549Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:083d7d64af31cd090f870eb49fb815e6bb42c175fc602ee9dae2f28f082bd4dc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:32.614823 containerd[1456]: time="2025-07-07T00:05:32.614800329Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.10\" with image id \"sha256:74c5154ea84d9a53c406e6c00e53cf66145cce821fd80e3c74e2e1bf312f3977\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.10\", repo digest \"registry.k8s.io/kube-apiserver@sha256:083d7d64af31cd090f870eb49fb815e6bb42c175fc602ee9dae2f28f082bd4dc\", size \"28074544\" in 2.032929724s" Jul 7 00:05:32.614908 containerd[1456]: time="2025-07-07T00:05:32.614889499Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.10\" returns image reference \"sha256:74c5154ea84d9a53c406e6c00e53cf66145cce821fd80e3c74e2e1bf312f3977\"" Jul 7 00:05:32.616247 containerd[1456]: time="2025-07-07T00:05:32.616228058Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.10\"" Jul 7 00:05:33.135177 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jul 7 00:05:33.144715 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 7 00:05:33.289093 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 7 00:05:33.294188 (kubelet)[1892]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 7 00:05:33.339182 kubelet[1892]: E0707 00:05:33.338301 1892 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 7 00:05:33.345971 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 7 00:05:33.346176 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 7 00:05:33.694892 containerd[1456]: time="2025-07-07T00:05:33.694844158Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:33.695842 containerd[1456]: time="2025-07-07T00:05:33.695808618Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.10: active requests=0, bytes read=24713300" Jul 7 00:05:33.696301 containerd[1456]: time="2025-07-07T00:05:33.696257338Z" level=info msg="ImageCreate event name:\"sha256:c285c4e62c91c434e9928bee7063b361509f43f43faa31641b626d6eff97616d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:33.698750 containerd[1456]: time="2025-07-07T00:05:33.698713786Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:3c67387d023c6114879f1e817669fd641797d30f117230682faf3930ecaaf0fe\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:33.699877 containerd[1456]: time="2025-07-07T00:05:33.699844746Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.10\" with image id \"sha256:c285c4e62c91c434e9928bee7063b361509f43f43faa31641b626d6eff97616d\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.10\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:3c67387d023c6114879f1e817669fd641797d30f117230682faf3930ecaaf0fe\", size \"26315128\" in 1.083589828s" Jul 7 00:05:33.700008 containerd[1456]: time="2025-07-07T00:05:33.699927676Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.10\" returns image reference \"sha256:c285c4e62c91c434e9928bee7063b361509f43f43faa31641b626d6eff97616d\"" Jul 7 00:05:33.700745 containerd[1456]: time="2025-07-07T00:05:33.700713785Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.10\"" Jul 7 00:05:34.643939 containerd[1456]: time="2025-07-07T00:05:34.643884704Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:34.644905 containerd[1456]: time="2025-07-07T00:05:34.644824653Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.10: active requests=0, bytes read=18783677" Jul 7 00:05:34.645444 containerd[1456]: time="2025-07-07T00:05:34.645377923Z" level=info msg="ImageCreate event name:\"sha256:61daeb7d112d9547792027cb16242b1d131f357f511545477381457fff5a69e2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:34.650773 containerd[1456]: time="2025-07-07T00:05:34.649696351Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:284dc2a5cf6afc9b76e39ad4b79c680c23d289488517643b28784a06d0141272\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:34.651130 containerd[1456]: time="2025-07-07T00:05:34.651088760Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.10\" with image id \"sha256:61daeb7d112d9547792027cb16242b1d131f357f511545477381457fff5a69e2\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.10\", repo digest \"registry.k8s.io/kube-scheduler@sha256:284dc2a5cf6afc9b76e39ad4b79c680c23d289488517643b28784a06d0141272\", size \"20385523\" in 950.347185ms" Jul 7 00:05:34.651207 containerd[1456]: time="2025-07-07T00:05:34.651191750Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.10\" returns image reference \"sha256:61daeb7d112d9547792027cb16242b1d131f357f511545477381457fff5a69e2\"" Jul 7 00:05:34.652312 containerd[1456]: time="2025-07-07T00:05:34.652293260Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.10\"" Jul 7 00:05:35.558830 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3635468065.mount: Deactivated successfully. Jul 7 00:05:35.925550 containerd[1456]: time="2025-07-07T00:05:35.923946164Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:35.925550 containerd[1456]: time="2025-07-07T00:05:35.925158703Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.10: active requests=0, bytes read=30383949" Jul 7 00:05:35.925944 containerd[1456]: time="2025-07-07T00:05:35.925770663Z" level=info msg="ImageCreate event name:\"sha256:3ed600862d3e69931e0f9f4dbf5c2b46343af40aa079772434f13de771bdc30c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:35.928719 containerd[1456]: time="2025-07-07T00:05:35.928505821Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:bcbb293812bdf587b28ea98369a8c347ca84884160046296761acdf12b27029d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:35.928719 containerd[1456]: time="2025-07-07T00:05:35.928537431Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.10\" with image id \"sha256:3ed600862d3e69931e0f9f4dbf5c2b46343af40aa079772434f13de771bdc30c\", repo tag \"registry.k8s.io/kube-proxy:v1.31.10\", repo digest \"registry.k8s.io/kube-proxy@sha256:bcbb293812bdf587b28ea98369a8c347ca84884160046296761acdf12b27029d\", size \"30382962\" in 1.276213131s" Jul 7 00:05:35.928719 containerd[1456]: time="2025-07-07T00:05:35.928595281Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.10\" returns image reference \"sha256:3ed600862d3e69931e0f9f4dbf5c2b46343af40aa079772434f13de771bdc30c\"" Jul 7 00:05:35.930516 containerd[1456]: time="2025-07-07T00:05:35.930476780Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Jul 7 00:05:36.524766 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3485793448.mount: Deactivated successfully. Jul 7 00:05:37.338862 containerd[1456]: time="2025-07-07T00:05:37.338776656Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:37.340491 containerd[1456]: time="2025-07-07T00:05:37.340225785Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565247" Jul 7 00:05:37.341167 containerd[1456]: time="2025-07-07T00:05:37.340949625Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:37.344390 containerd[1456]: time="2025-07-07T00:05:37.343551654Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:37.346112 containerd[1456]: time="2025-07-07T00:05:37.344615223Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.414101783s" Jul 7 00:05:37.346112 containerd[1456]: time="2025-07-07T00:05:37.344649523Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Jul 7 00:05:37.347409 containerd[1456]: time="2025-07-07T00:05:37.347262522Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Jul 7 00:05:37.833674 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount206494139.mount: Deactivated successfully. Jul 7 00:05:37.836730 containerd[1456]: time="2025-07-07T00:05:37.836677187Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:37.837535 containerd[1456]: time="2025-07-07T00:05:37.837462877Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321144" Jul 7 00:05:37.838033 containerd[1456]: time="2025-07-07T00:05:37.837980196Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:37.847224 containerd[1456]: time="2025-07-07T00:05:37.845809712Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:37.847224 containerd[1456]: time="2025-07-07T00:05:37.847125852Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 499.83812ms" Jul 7 00:05:37.847224 containerd[1456]: time="2025-07-07T00:05:37.847151972Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Jul 7 00:05:37.848820 containerd[1456]: time="2025-07-07T00:05:37.848789841Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Jul 7 00:05:38.341442 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3884013448.mount: Deactivated successfully. Jul 7 00:05:39.698184 containerd[1456]: time="2025-07-07T00:05:39.698126046Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:39.699379 containerd[1456]: time="2025-07-07T00:05:39.699300195Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=56780019" Jul 7 00:05:39.699894 containerd[1456]: time="2025-07-07T00:05:39.699839115Z" level=info msg="ImageCreate event name:\"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:39.702837 containerd[1456]: time="2025-07-07T00:05:39.702344644Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:39.703447 containerd[1456]: time="2025-07-07T00:05:39.703416673Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"56909194\" in 1.854599522s" Jul 7 00:05:39.703505 containerd[1456]: time="2025-07-07T00:05:39.703448573Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Jul 7 00:05:41.904466 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 7 00:05:41.913589 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 7 00:05:41.937444 systemd[1]: Reloading requested from client PID 2054 ('systemctl') (unit session-7.scope)... Jul 7 00:05:41.937769 systemd[1]: Reloading... Jul 7 00:05:42.092386 zram_generator::config[2095]: No configuration found. Jul 7 00:05:42.204542 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 7 00:05:42.267100 systemd[1]: Reloading finished in 328 ms. Jul 7 00:05:42.319266 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 7 00:05:42.330829 (kubelet)[2139]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jul 7 00:05:42.332403 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jul 7 00:05:42.332668 systemd[1]: kubelet.service: Deactivated successfully. Jul 7 00:05:42.333307 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 7 00:05:42.339853 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 7 00:05:42.512949 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 7 00:05:42.521789 (kubelet)[2151]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jul 7 00:05:42.560696 kubelet[2151]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 7 00:05:42.561001 kubelet[2151]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jul 7 00:05:42.561043 kubelet[2151]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 7 00:05:42.561179 kubelet[2151]: I0707 00:05:42.561147 2151 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jul 7 00:05:42.864634 kubelet[2151]: I0707 00:05:42.864188 2151 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Jul 7 00:05:42.864634 kubelet[2151]: I0707 00:05:42.864226 2151 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jul 7 00:05:42.864912 kubelet[2151]: I0707 00:05:42.864819 2151 server.go:934] "Client rotation is on, will bootstrap in background" Jul 7 00:05:42.884756 kubelet[2151]: E0707 00:05:42.884678 2151 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.234.31.125:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.234.31.125:6443: connect: connection refused" logger="UnhandledError" Jul 7 00:05:42.886088 kubelet[2151]: I0707 00:05:42.885986 2151 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 7 00:05:42.893093 kubelet[2151]: E0707 00:05:42.893059 2151 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Jul 7 00:05:42.893093 kubelet[2151]: I0707 00:05:42.893083 2151 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Jul 7 00:05:42.899795 kubelet[2151]: I0707 00:05:42.899380 2151 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 7 00:05:42.900388 kubelet[2151]: I0707 00:05:42.900311 2151 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jul 7 00:05:42.900625 kubelet[2151]: I0707 00:05:42.900602 2151 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 7 00:05:42.900842 kubelet[2151]: I0707 00:05:42.900666 2151 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"172-234-31-125","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jul 7 00:05:42.900993 kubelet[2151]: I0707 00:05:42.900980 2151 topology_manager.go:138] "Creating topology manager with none policy" Jul 7 00:05:42.901042 kubelet[2151]: I0707 00:05:42.901034 2151 container_manager_linux.go:300] "Creating device plugin manager" Jul 7 00:05:42.901197 kubelet[2151]: I0707 00:05:42.901185 2151 state_mem.go:36] "Initialized new in-memory state store" Jul 7 00:05:42.904293 kubelet[2151]: I0707 00:05:42.904280 2151 kubelet.go:408] "Attempting to sync node with API server" Jul 7 00:05:42.904420 kubelet[2151]: I0707 00:05:42.904355 2151 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Jul 7 00:05:42.904566 kubelet[2151]: I0707 00:05:42.904495 2151 kubelet.go:314] "Adding apiserver pod source" Jul 7 00:05:42.904566 kubelet[2151]: I0707 00:05:42.904517 2151 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 7 00:05:42.910208 kubelet[2151]: W0707 00:05:42.910117 2151 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.234.31.125:6443/api/v1/nodes?fieldSelector=metadata.name%3D172-234-31-125&limit=500&resourceVersion=0": dial tcp 172.234.31.125:6443: connect: connection refused Jul 7 00:05:42.910208 kubelet[2151]: E0707 00:05:42.910196 2151 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.234.31.125:6443/api/v1/nodes?fieldSelector=metadata.name%3D172-234-31-125&limit=500&resourceVersion=0\": dial tcp 172.234.31.125:6443: connect: connection refused" logger="UnhandledError" Jul 7 00:05:42.910329 kubelet[2151]: I0707 00:05:42.910308 2151 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Jul 7 00:05:42.910751 kubelet[2151]: I0707 00:05:42.910720 2151 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jul 7 00:05:42.911653 kubelet[2151]: W0707 00:05:42.911626 2151 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jul 7 00:05:42.913352 kubelet[2151]: W0707 00:05:42.913217 2151 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.234.31.125:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.234.31.125:6443: connect: connection refused Jul 7 00:05:42.913352 kubelet[2151]: E0707 00:05:42.913272 2151 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.234.31.125:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.234.31.125:6443: connect: connection refused" logger="UnhandledError" Jul 7 00:05:42.914327 kubelet[2151]: I0707 00:05:42.914295 2151 server.go:1274] "Started kubelet" Jul 7 00:05:42.914512 kubelet[2151]: I0707 00:05:42.914442 2151 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jul 7 00:05:42.915401 kubelet[2151]: I0707 00:05:42.915386 2151 server.go:449] "Adding debug handlers to kubelet server" Jul 7 00:05:42.917799 kubelet[2151]: I0707 00:05:42.917218 2151 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jul 7 00:05:42.917799 kubelet[2151]: I0707 00:05:42.917519 2151 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jul 7 00:05:42.918863 kubelet[2151]: E0707 00:05:42.917681 2151 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.234.31.125:6443/api/v1/namespaces/default/events\": dial tcp 172.234.31.125:6443: connect: connection refused" event="&Event{ObjectMeta:{172-234-31-125.184fcf54bc5474a6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:172-234-31-125,UID:172-234-31-125,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:172-234-31-125,},FirstTimestamp:2025-07-07 00:05:42.914266278 +0000 UTC m=+0.388529257,LastTimestamp:2025-07-07 00:05:42.914266278 +0000 UTC m=+0.388529257,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172-234-31-125,}" Jul 7 00:05:42.920350 kubelet[2151]: I0707 00:05:42.920332 2151 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jul 7 00:05:42.923189 kubelet[2151]: E0707 00:05:42.922830 2151 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jul 7 00:05:42.923189 kubelet[2151]: I0707 00:05:42.923063 2151 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jul 7 00:05:42.926073 kubelet[2151]: E0707 00:05:42.926017 2151 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"172-234-31-125\" not found" Jul 7 00:05:42.926124 kubelet[2151]: I0707 00:05:42.926081 2151 volume_manager.go:289] "Starting Kubelet Volume Manager" Jul 7 00:05:42.926286 kubelet[2151]: I0707 00:05:42.926258 2151 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Jul 7 00:05:42.926323 kubelet[2151]: I0707 00:05:42.926314 2151 reconciler.go:26] "Reconciler: start to sync state" Jul 7 00:05:42.926941 kubelet[2151]: I0707 00:05:42.926909 2151 factory.go:221] Registration of the systemd container factory successfully Jul 7 00:05:42.927015 kubelet[2151]: I0707 00:05:42.926987 2151 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jul 7 00:05:42.927489 kubelet[2151]: W0707 00:05:42.927385 2151 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.234.31.125:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.234.31.125:6443: connect: connection refused Jul 7 00:05:42.927489 kubelet[2151]: E0707 00:05:42.927422 2151 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.234.31.125:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.234.31.125:6443: connect: connection refused" logger="UnhandledError" Jul 7 00:05:42.927731 kubelet[2151]: E0707 00:05:42.927699 2151 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.234.31.125:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172-234-31-125?timeout=10s\": dial tcp 172.234.31.125:6443: connect: connection refused" interval="200ms" Jul 7 00:05:42.928450 kubelet[2151]: I0707 00:05:42.928297 2151 factory.go:221] Registration of the containerd container factory successfully Jul 7 00:05:42.937714 kubelet[2151]: I0707 00:05:42.937687 2151 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jul 7 00:05:42.938945 kubelet[2151]: I0707 00:05:42.938931 2151 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jul 7 00:05:42.939321 kubelet[2151]: I0707 00:05:42.939003 2151 status_manager.go:217] "Starting to sync pod status with apiserver" Jul 7 00:05:42.939321 kubelet[2151]: I0707 00:05:42.939029 2151 kubelet.go:2321] "Starting kubelet main sync loop" Jul 7 00:05:42.939321 kubelet[2151]: E0707 00:05:42.939070 2151 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 7 00:05:42.950129 kubelet[2151]: W0707 00:05:42.950077 2151 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.234.31.125:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.234.31.125:6443: connect: connection refused Jul 7 00:05:42.950249 kubelet[2151]: E0707 00:05:42.950230 2151 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.234.31.125:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.234.31.125:6443: connect: connection refused" logger="UnhandledError" Jul 7 00:05:42.959281 kubelet[2151]: I0707 00:05:42.959250 2151 cpu_manager.go:214] "Starting CPU manager" policy="none" Jul 7 00:05:42.959281 kubelet[2151]: I0707 00:05:42.959274 2151 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jul 7 00:05:42.959391 kubelet[2151]: I0707 00:05:42.959290 2151 state_mem.go:36] "Initialized new in-memory state store" Jul 7 00:05:42.961010 kubelet[2151]: I0707 00:05:42.960985 2151 policy_none.go:49] "None policy: Start" Jul 7 00:05:42.961722 kubelet[2151]: I0707 00:05:42.961532 2151 memory_manager.go:170] "Starting memorymanager" policy="None" Jul 7 00:05:42.961722 kubelet[2151]: I0707 00:05:42.961674 2151 state_mem.go:35] "Initializing new in-memory state store" Jul 7 00:05:42.967924 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jul 7 00:05:42.984977 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jul 7 00:05:42.996126 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jul 7 00:05:42.997920 kubelet[2151]: I0707 00:05:42.997896 2151 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jul 7 00:05:42.998142 kubelet[2151]: I0707 00:05:42.998109 2151 eviction_manager.go:189] "Eviction manager: starting control loop" Jul 7 00:05:42.998194 kubelet[2151]: I0707 00:05:42.998129 2151 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jul 7 00:05:43.000194 kubelet[2151]: I0707 00:05:42.999952 2151 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jul 7 00:05:43.001097 kubelet[2151]: E0707 00:05:43.001070 2151 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"172-234-31-125\" not found" Jul 7 00:05:43.053945 systemd[1]: Created slice kubepods-burstable-pode0c092ee8be31a6cfdfccfcbf86f82b6.slice - libcontainer container kubepods-burstable-pode0c092ee8be31a6cfdfccfcbf86f82b6.slice. Jul 7 00:05:43.078244 systemd[1]: Created slice kubepods-burstable-pod6a123eb34d53cf31ccd27f0ce9e84c2f.slice - libcontainer container kubepods-burstable-pod6a123eb34d53cf31ccd27f0ce9e84c2f.slice. Jul 7 00:05:43.083203 systemd[1]: Created slice kubepods-burstable-pod2502a970e91bbd0f9e749dbbd719d057.slice - libcontainer container kubepods-burstable-pod2502a970e91bbd0f9e749dbbd719d057.slice. Jul 7 00:05:43.100868 kubelet[2151]: I0707 00:05:43.100742 2151 kubelet_node_status.go:72] "Attempting to register node" node="172-234-31-125" Jul 7 00:05:43.101270 kubelet[2151]: E0707 00:05:43.101230 2151 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.234.31.125:6443/api/v1/nodes\": dial tcp 172.234.31.125:6443: connect: connection refused" node="172-234-31-125" Jul 7 00:05:43.127806 kubelet[2151]: I0707 00:05:43.127084 2151 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e0c092ee8be31a6cfdfccfcbf86f82b6-ca-certs\") pod \"kube-apiserver-172-234-31-125\" (UID: \"e0c092ee8be31a6cfdfccfcbf86f82b6\") " pod="kube-system/kube-apiserver-172-234-31-125" Jul 7 00:05:43.127806 kubelet[2151]: I0707 00:05:43.127128 2151 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e0c092ee8be31a6cfdfccfcbf86f82b6-k8s-certs\") pod \"kube-apiserver-172-234-31-125\" (UID: \"e0c092ee8be31a6cfdfccfcbf86f82b6\") " pod="kube-system/kube-apiserver-172-234-31-125" Jul 7 00:05:43.127806 kubelet[2151]: I0707 00:05:43.127156 2151 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/2502a970e91bbd0f9e749dbbd719d057-ca-certs\") pod \"kube-controller-manager-172-234-31-125\" (UID: \"2502a970e91bbd0f9e749dbbd719d057\") " pod="kube-system/kube-controller-manager-172-234-31-125" Jul 7 00:05:43.127806 kubelet[2151]: I0707 00:05:43.127178 2151 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/2502a970e91bbd0f9e749dbbd719d057-k8s-certs\") pod \"kube-controller-manager-172-234-31-125\" (UID: \"2502a970e91bbd0f9e749dbbd719d057\") " pod="kube-system/kube-controller-manager-172-234-31-125" Jul 7 00:05:43.127806 kubelet[2151]: I0707 00:05:43.127203 2151 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/2502a970e91bbd0f9e749dbbd719d057-usr-share-ca-certificates\") pod \"kube-controller-manager-172-234-31-125\" (UID: \"2502a970e91bbd0f9e749dbbd719d057\") " pod="kube-system/kube-controller-manager-172-234-31-125" Jul 7 00:05:43.127993 kubelet[2151]: I0707 00:05:43.127228 2151 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6a123eb34d53cf31ccd27f0ce9e84c2f-kubeconfig\") pod \"kube-scheduler-172-234-31-125\" (UID: \"6a123eb34d53cf31ccd27f0ce9e84c2f\") " pod="kube-system/kube-scheduler-172-234-31-125" Jul 7 00:05:43.127993 kubelet[2151]: I0707 00:05:43.127257 2151 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e0c092ee8be31a6cfdfccfcbf86f82b6-usr-share-ca-certificates\") pod \"kube-apiserver-172-234-31-125\" (UID: \"e0c092ee8be31a6cfdfccfcbf86f82b6\") " pod="kube-system/kube-apiserver-172-234-31-125" Jul 7 00:05:43.127993 kubelet[2151]: I0707 00:05:43.127278 2151 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/2502a970e91bbd0f9e749dbbd719d057-flexvolume-dir\") pod \"kube-controller-manager-172-234-31-125\" (UID: \"2502a970e91bbd0f9e749dbbd719d057\") " pod="kube-system/kube-controller-manager-172-234-31-125" Jul 7 00:05:43.127993 kubelet[2151]: I0707 00:05:43.127314 2151 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/2502a970e91bbd0f9e749dbbd719d057-kubeconfig\") pod \"kube-controller-manager-172-234-31-125\" (UID: \"2502a970e91bbd0f9e749dbbd719d057\") " pod="kube-system/kube-controller-manager-172-234-31-125" Jul 7 00:05:43.129205 kubelet[2151]: E0707 00:05:43.129007 2151 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.234.31.125:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172-234-31-125?timeout=10s\": dial tcp 172.234.31.125:6443: connect: connection refused" interval="400ms" Jul 7 00:05:43.303518 kubelet[2151]: I0707 00:05:43.303474 2151 kubelet_node_status.go:72] "Attempting to register node" node="172-234-31-125" Jul 7 00:05:43.304229 kubelet[2151]: E0707 00:05:43.304180 2151 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.234.31.125:6443/api/v1/nodes\": dial tcp 172.234.31.125:6443: connect: connection refused" node="172-234-31-125" Jul 7 00:05:43.374844 kubelet[2151]: E0707 00:05:43.374813 2151 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:43.375722 containerd[1456]: time="2025-07-07T00:05:43.375671377Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-172-234-31-125,Uid:e0c092ee8be31a6cfdfccfcbf86f82b6,Namespace:kube-system,Attempt:0,}" Jul 7 00:05:43.382284 kubelet[2151]: E0707 00:05:43.382155 2151 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:43.382643 containerd[1456]: time="2025-07-07T00:05:43.382572963Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-172-234-31-125,Uid:6a123eb34d53cf31ccd27f0ce9e84c2f,Namespace:kube-system,Attempt:0,}" Jul 7 00:05:43.386459 kubelet[2151]: E0707 00:05:43.386422 2151 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:43.387043 containerd[1456]: time="2025-07-07T00:05:43.387006771Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-172-234-31-125,Uid:2502a970e91bbd0f9e749dbbd719d057,Namespace:kube-system,Attempt:0,}" Jul 7 00:05:43.529874 kubelet[2151]: E0707 00:05:43.529821 2151 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.234.31.125:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172-234-31-125?timeout=10s\": dial tcp 172.234.31.125:6443: connect: connection refused" interval="800ms" Jul 7 00:05:43.706323 kubelet[2151]: I0707 00:05:43.706277 2151 kubelet_node_status.go:72] "Attempting to register node" node="172-234-31-125" Jul 7 00:05:43.707314 kubelet[2151]: E0707 00:05:43.706741 2151 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.234.31.125:6443/api/v1/nodes\": dial tcp 172.234.31.125:6443: connect: connection refused" node="172-234-31-125" Jul 7 00:05:43.853263 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1971083042.mount: Deactivated successfully. Jul 7 00:05:43.857989 containerd[1456]: time="2025-07-07T00:05:43.857799766Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 7 00:05:43.858845 containerd[1456]: time="2025-07-07T00:05:43.858785165Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 7 00:05:43.859639 containerd[1456]: time="2025-07-07T00:05:43.859564865Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jul 7 00:05:43.859937 containerd[1456]: time="2025-07-07T00:05:43.859894715Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312062" Jul 7 00:05:43.861409 containerd[1456]: time="2025-07-07T00:05:43.860622114Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 7 00:05:43.861903 containerd[1456]: time="2025-07-07T00:05:43.861863214Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jul 7 00:05:43.862322 containerd[1456]: time="2025-07-07T00:05:43.862298643Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 7 00:05:43.865554 containerd[1456]: time="2025-07-07T00:05:43.865520482Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 7 00:05:43.866126 containerd[1456]: time="2025-07-07T00:05:43.866085362Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 478.991601ms" Jul 7 00:05:43.867065 containerd[1456]: time="2025-07-07T00:05:43.867007461Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 484.383548ms" Jul 7 00:05:43.867968 containerd[1456]: time="2025-07-07T00:05:43.867933011Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 492.172904ms" Jul 7 00:05:43.966888 containerd[1456]: time="2025-07-07T00:05:43.966317151Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:05:43.967189 containerd[1456]: time="2025-07-07T00:05:43.966739331Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:05:43.967309 containerd[1456]: time="2025-07-07T00:05:43.967274961Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:05:43.967552 containerd[1456]: time="2025-07-07T00:05:43.967500141Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:05:43.974513 containerd[1456]: time="2025-07-07T00:05:43.974422527Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:05:43.974513 containerd[1456]: time="2025-07-07T00:05:43.974479887Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:05:43.974513 containerd[1456]: time="2025-07-07T00:05:43.974416577Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:05:43.974657 containerd[1456]: time="2025-07-07T00:05:43.974494617Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:05:43.974657 containerd[1456]: time="2025-07-07T00:05:43.974577737Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:05:43.974878 containerd[1456]: time="2025-07-07T00:05:43.974792987Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:05:43.974878 containerd[1456]: time="2025-07-07T00:05:43.974668647Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:05:43.976063 containerd[1456]: time="2025-07-07T00:05:43.975018417Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:05:44.015530 systemd[1]: Started cri-containerd-162b3b8077387db05acd83dd1feee5ccf969c6573f864a822afc21302ad82ff0.scope - libcontainer container 162b3b8077387db05acd83dd1feee5ccf969c6573f864a822afc21302ad82ff0. Jul 7 00:05:44.017671 systemd[1]: Started cri-containerd-1c29e01053355a3a2253c263966c31260ec38099c383aaaab1484f04d86d56dd.scope - libcontainer container 1c29e01053355a3a2253c263966c31260ec38099c383aaaab1484f04d86d56dd. Jul 7 00:05:44.020847 systemd[1]: Started cri-containerd-818ef1042e6d2517a9d3fb3f6cd5b31885e82aba5a0eeb87a4d1da91f53a1422.scope - libcontainer container 818ef1042e6d2517a9d3fb3f6cd5b31885e82aba5a0eeb87a4d1da91f53a1422. Jul 7 00:05:44.049174 kubelet[2151]: W0707 00:05:44.049088 2151 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.234.31.125:6443/api/v1/nodes?fieldSelector=metadata.name%3D172-234-31-125&limit=500&resourceVersion=0": dial tcp 172.234.31.125:6443: connect: connection refused Jul 7 00:05:44.049256 kubelet[2151]: E0707 00:05:44.049196 2151 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.234.31.125:6443/api/v1/nodes?fieldSelector=metadata.name%3D172-234-31-125&limit=500&resourceVersion=0\": dial tcp 172.234.31.125:6443: connect: connection refused" logger="UnhandledError" Jul 7 00:05:44.054157 kubelet[2151]: W0707 00:05:44.054111 2151 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.234.31.125:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.234.31.125:6443: connect: connection refused Jul 7 00:05:44.054157 kubelet[2151]: E0707 00:05:44.054156 2151 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.234.31.125:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.234.31.125:6443: connect: connection refused" logger="UnhandledError" Jul 7 00:05:44.076699 containerd[1456]: time="2025-07-07T00:05:44.076628676Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-172-234-31-125,Uid:6a123eb34d53cf31ccd27f0ce9e84c2f,Namespace:kube-system,Attempt:0,} returns sandbox id \"818ef1042e6d2517a9d3fb3f6cd5b31885e82aba5a0eeb87a4d1da91f53a1422\"" Jul 7 00:05:44.078754 kubelet[2151]: W0707 00:05:44.078563 2151 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.234.31.125:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.234.31.125:6443: connect: connection refused Jul 7 00:05:44.078754 kubelet[2151]: E0707 00:05:44.078649 2151 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.234.31.125:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.234.31.125:6443: connect: connection refused" logger="UnhandledError" Jul 7 00:05:44.080462 kubelet[2151]: E0707 00:05:44.080431 2151 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:44.084843 containerd[1456]: time="2025-07-07T00:05:44.084797742Z" level=info msg="CreateContainer within sandbox \"818ef1042e6d2517a9d3fb3f6cd5b31885e82aba5a0eeb87a4d1da91f53a1422\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jul 7 00:05:44.099144 containerd[1456]: time="2025-07-07T00:05:44.099004445Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-172-234-31-125,Uid:e0c092ee8be31a6cfdfccfcbf86f82b6,Namespace:kube-system,Attempt:0,} returns sandbox id \"162b3b8077387db05acd83dd1feee5ccf969c6573f864a822afc21302ad82ff0\"" Jul 7 00:05:44.100073 containerd[1456]: time="2025-07-07T00:05:44.100036265Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-172-234-31-125,Uid:2502a970e91bbd0f9e749dbbd719d057,Namespace:kube-system,Attempt:0,} returns sandbox id \"1c29e01053355a3a2253c263966c31260ec38099c383aaaab1484f04d86d56dd\"" Jul 7 00:05:44.101396 kubelet[2151]: E0707 00:05:44.101273 2151 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:44.102874 kubelet[2151]: E0707 00:05:44.102846 2151 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:44.105160 containerd[1456]: time="2025-07-07T00:05:44.105123942Z" level=info msg="CreateContainer within sandbox \"162b3b8077387db05acd83dd1feee5ccf969c6573f864a822afc21302ad82ff0\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jul 7 00:05:44.107983 containerd[1456]: time="2025-07-07T00:05:44.107883021Z" level=info msg="CreateContainer within sandbox \"1c29e01053355a3a2253c263966c31260ec38099c383aaaab1484f04d86d56dd\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jul 7 00:05:44.115030 containerd[1456]: time="2025-07-07T00:05:44.114989777Z" level=info msg="CreateContainer within sandbox \"818ef1042e6d2517a9d3fb3f6cd5b31885e82aba5a0eeb87a4d1da91f53a1422\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"47292bd697409583011a0c2fbf3c90e5ab6d2fe866cf20fa07f172ae1b1729af\"" Jul 7 00:05:44.116387 containerd[1456]: time="2025-07-07T00:05:44.116002967Z" level=info msg="StartContainer for \"47292bd697409583011a0c2fbf3c90e5ab6d2fe866cf20fa07f172ae1b1729af\"" Jul 7 00:05:44.119615 containerd[1456]: time="2025-07-07T00:05:44.119564405Z" level=info msg="CreateContainer within sandbox \"162b3b8077387db05acd83dd1feee5ccf969c6573f864a822afc21302ad82ff0\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"0e894cebf32bf4ac00d8cd0d55b9f6e9c7589c44aac0b127eb8bc8901461889c\"" Jul 7 00:05:44.120699 containerd[1456]: time="2025-07-07T00:05:44.120665994Z" level=info msg="StartContainer for \"0e894cebf32bf4ac00d8cd0d55b9f6e9c7589c44aac0b127eb8bc8901461889c\"" Jul 7 00:05:44.127326 containerd[1456]: time="2025-07-07T00:05:44.127282381Z" level=info msg="CreateContainer within sandbox \"1c29e01053355a3a2253c263966c31260ec38099c383aaaab1484f04d86d56dd\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"abe574ff273171a3e8aaa9e04d377bcdccdeb7003f0c10cb45bd35c29b67416f\"" Jul 7 00:05:44.129387 containerd[1456]: time="2025-07-07T00:05:44.127990991Z" level=info msg="StartContainer for \"abe574ff273171a3e8aaa9e04d377bcdccdeb7003f0c10cb45bd35c29b67416f\"" Jul 7 00:05:44.154561 systemd[1]: Started cri-containerd-0e894cebf32bf4ac00d8cd0d55b9f6e9c7589c44aac0b127eb8bc8901461889c.scope - libcontainer container 0e894cebf32bf4ac00d8cd0d55b9f6e9c7589c44aac0b127eb8bc8901461889c. Jul 7 00:05:44.166857 systemd[1]: Started cri-containerd-47292bd697409583011a0c2fbf3c90e5ab6d2fe866cf20fa07f172ae1b1729af.scope - libcontainer container 47292bd697409583011a0c2fbf3c90e5ab6d2fe866cf20fa07f172ae1b1729af. Jul 7 00:05:44.175499 systemd[1]: Started cri-containerd-abe574ff273171a3e8aaa9e04d377bcdccdeb7003f0c10cb45bd35c29b67416f.scope - libcontainer container abe574ff273171a3e8aaa9e04d377bcdccdeb7003f0c10cb45bd35c29b67416f. Jul 7 00:05:44.222143 containerd[1456]: time="2025-07-07T00:05:44.221852634Z" level=info msg="StartContainer for \"47292bd697409583011a0c2fbf3c90e5ab6d2fe866cf20fa07f172ae1b1729af\" returns successfully" Jul 7 00:05:44.224331 containerd[1456]: time="2025-07-07T00:05:44.224289682Z" level=info msg="StartContainer for \"0e894cebf32bf4ac00d8cd0d55b9f6e9c7589c44aac0b127eb8bc8901461889c\" returns successfully" Jul 7 00:05:44.249370 containerd[1456]: time="2025-07-07T00:05:44.249290530Z" level=info msg="StartContainer for \"abe574ff273171a3e8aaa9e04d377bcdccdeb7003f0c10cb45bd35c29b67416f\" returns successfully" Jul 7 00:05:44.510149 kubelet[2151]: I0707 00:05:44.510019 2151 kubelet_node_status.go:72] "Attempting to register node" node="172-234-31-125" Jul 7 00:05:44.963511 kubelet[2151]: E0707 00:05:44.961456 2151 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:44.965221 kubelet[2151]: E0707 00:05:44.964754 2151 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:44.974829 kubelet[2151]: E0707 00:05:44.974641 2151 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:45.462167 kubelet[2151]: E0707 00:05:45.462124 2151 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"172-234-31-125\" not found" node="172-234-31-125" Jul 7 00:05:45.506896 kubelet[2151]: I0707 00:05:45.506856 2151 kubelet_node_status.go:75] "Successfully registered node" node="172-234-31-125" Jul 7 00:05:45.506896 kubelet[2151]: E0707 00:05:45.506888 2151 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"172-234-31-125\": node \"172-234-31-125\" not found" Jul 7 00:05:45.515603 kubelet[2151]: E0707 00:05:45.515535 2151 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"172-234-31-125\" not found" Jul 7 00:05:45.616547 kubelet[2151]: E0707 00:05:45.616493 2151 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"172-234-31-125\" not found" Jul 7 00:05:45.718482 kubelet[2151]: E0707 00:05:45.717917 2151 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"172-234-31-125\" not found" Jul 7 00:05:45.818609 kubelet[2151]: E0707 00:05:45.818565 2151 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"172-234-31-125\" not found" Jul 7 00:05:45.918981 kubelet[2151]: E0707 00:05:45.918929 2151 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"172-234-31-125\" not found" Jul 7 00:05:45.968228 kubelet[2151]: E0707 00:05:45.968202 2151 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:46.472469 kubelet[2151]: E0707 00:05:46.472389 2151 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:46.915392 kubelet[2151]: I0707 00:05:46.915315 2151 apiserver.go:52] "Watching apiserver" Jul 7 00:05:46.927186 kubelet[2151]: I0707 00:05:46.927162 2151 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Jul 7 00:05:46.968168 kubelet[2151]: E0707 00:05:46.968103 2151 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:46.974372 kubelet[2151]: E0707 00:05:46.974302 2151 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:47.470756 systemd[1]: Reloading requested from client PID 2424 ('systemctl') (unit session-7.scope)... Jul 7 00:05:47.470776 systemd[1]: Reloading... Jul 7 00:05:47.589436 zram_generator::config[2467]: No configuration found. Jul 7 00:05:47.715553 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 7 00:05:47.790968 systemd[1]: Reloading finished in 319 ms. Jul 7 00:05:47.836977 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jul 7 00:05:47.861086 systemd[1]: kubelet.service: Deactivated successfully. Jul 7 00:05:47.861481 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 7 00:05:47.868588 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 7 00:05:48.026532 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 7 00:05:48.026753 (kubelet)[2515]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jul 7 00:05:48.074535 kubelet[2515]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 7 00:05:48.074535 kubelet[2515]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jul 7 00:05:48.074535 kubelet[2515]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 7 00:05:48.074535 kubelet[2515]: I0707 00:05:48.074240 2515 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jul 7 00:05:48.081458 kubelet[2515]: I0707 00:05:48.081422 2515 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Jul 7 00:05:48.081458 kubelet[2515]: I0707 00:05:48.081445 2515 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jul 7 00:05:48.081821 kubelet[2515]: I0707 00:05:48.081794 2515 server.go:934] "Client rotation is on, will bootstrap in background" Jul 7 00:05:48.084869 kubelet[2515]: I0707 00:05:48.084850 2515 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jul 7 00:05:48.088070 kubelet[2515]: I0707 00:05:48.087662 2515 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 7 00:05:48.097471 kubelet[2515]: E0707 00:05:48.097434 2515 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Jul 7 00:05:48.097522 kubelet[2515]: I0707 00:05:48.097471 2515 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Jul 7 00:05:48.101723 kubelet[2515]: I0707 00:05:48.101700 2515 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 7 00:05:48.101829 kubelet[2515]: I0707 00:05:48.101813 2515 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jul 7 00:05:48.101994 kubelet[2515]: I0707 00:05:48.101953 2515 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 7 00:05:48.102174 kubelet[2515]: I0707 00:05:48.101986 2515 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"172-234-31-125","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jul 7 00:05:48.102243 kubelet[2515]: I0707 00:05:48.102189 2515 topology_manager.go:138] "Creating topology manager with none policy" Jul 7 00:05:48.102243 kubelet[2515]: I0707 00:05:48.102201 2515 container_manager_linux.go:300] "Creating device plugin manager" Jul 7 00:05:48.102243 kubelet[2515]: I0707 00:05:48.102230 2515 state_mem.go:36] "Initialized new in-memory state store" Jul 7 00:05:48.102389 kubelet[2515]: I0707 00:05:48.102346 2515 kubelet.go:408] "Attempting to sync node with API server" Jul 7 00:05:48.102389 kubelet[2515]: I0707 00:05:48.102387 2515 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Jul 7 00:05:48.102443 kubelet[2515]: I0707 00:05:48.102426 2515 kubelet.go:314] "Adding apiserver pod source" Jul 7 00:05:48.102467 kubelet[2515]: I0707 00:05:48.102444 2515 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 7 00:05:48.105640 kubelet[2515]: I0707 00:05:48.105486 2515 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Jul 7 00:05:48.105914 kubelet[2515]: I0707 00:05:48.105889 2515 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jul 7 00:05:48.106464 kubelet[2515]: I0707 00:05:48.106323 2515 server.go:1274] "Started kubelet" Jul 7 00:05:48.110030 kubelet[2515]: I0707 00:05:48.110006 2515 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jul 7 00:05:48.113113 kubelet[2515]: I0707 00:05:48.113073 2515 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jul 7 00:05:48.117380 kubelet[2515]: I0707 00:05:48.116649 2515 server.go:449] "Adding debug handlers to kubelet server" Jul 7 00:05:48.117605 kubelet[2515]: I0707 00:05:48.117550 2515 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jul 7 00:05:48.118015 kubelet[2515]: I0707 00:05:48.117986 2515 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jul 7 00:05:48.118911 kubelet[2515]: I0707 00:05:48.118881 2515 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jul 7 00:05:48.122452 kubelet[2515]: I0707 00:05:48.122281 2515 volume_manager.go:289] "Starting Kubelet Volume Manager" Jul 7 00:05:48.122898 kubelet[2515]: E0707 00:05:48.122869 2515 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"172-234-31-125\" not found" Jul 7 00:05:48.125319 kubelet[2515]: I0707 00:05:48.125279 2515 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Jul 7 00:05:48.125466 kubelet[2515]: I0707 00:05:48.125440 2515 reconciler.go:26] "Reconciler: start to sync state" Jul 7 00:05:48.130386 kubelet[2515]: E0707 00:05:48.130351 2515 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jul 7 00:05:48.131102 kubelet[2515]: I0707 00:05:48.131080 2515 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jul 7 00:05:48.135817 kubelet[2515]: I0707 00:05:48.135801 2515 factory.go:221] Registration of the containerd container factory successfully Jul 7 00:05:48.135871 kubelet[2515]: I0707 00:05:48.135863 2515 factory.go:221] Registration of the systemd container factory successfully Jul 7 00:05:48.142043 kubelet[2515]: I0707 00:05:48.142009 2515 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jul 7 00:05:48.143194 kubelet[2515]: I0707 00:05:48.143162 2515 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jul 7 00:05:48.143194 kubelet[2515]: I0707 00:05:48.143188 2515 status_manager.go:217] "Starting to sync pod status with apiserver" Jul 7 00:05:48.143259 kubelet[2515]: I0707 00:05:48.143217 2515 kubelet.go:2321] "Starting kubelet main sync loop" Jul 7 00:05:48.143303 kubelet[2515]: E0707 00:05:48.143271 2515 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 7 00:05:48.195499 kubelet[2515]: I0707 00:05:48.195465 2515 cpu_manager.go:214] "Starting CPU manager" policy="none" Jul 7 00:05:48.195499 kubelet[2515]: I0707 00:05:48.195491 2515 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jul 7 00:05:48.196598 kubelet[2515]: I0707 00:05:48.195514 2515 state_mem.go:36] "Initialized new in-memory state store" Jul 7 00:05:48.196598 kubelet[2515]: I0707 00:05:48.195666 2515 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jul 7 00:05:48.196598 kubelet[2515]: I0707 00:05:48.195678 2515 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jul 7 00:05:48.196598 kubelet[2515]: I0707 00:05:48.195699 2515 policy_none.go:49] "None policy: Start" Jul 7 00:05:48.198161 kubelet[2515]: I0707 00:05:48.197337 2515 memory_manager.go:170] "Starting memorymanager" policy="None" Jul 7 00:05:48.198161 kubelet[2515]: I0707 00:05:48.197376 2515 state_mem.go:35] "Initializing new in-memory state store" Jul 7 00:05:48.198161 kubelet[2515]: I0707 00:05:48.197497 2515 state_mem.go:75] "Updated machine memory state" Jul 7 00:05:48.203818 kubelet[2515]: I0707 00:05:48.203801 2515 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jul 7 00:05:48.204460 kubelet[2515]: I0707 00:05:48.204447 2515 eviction_manager.go:189] "Eviction manager: starting control loop" Jul 7 00:05:48.204534 kubelet[2515]: I0707 00:05:48.204507 2515 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jul 7 00:05:48.205014 kubelet[2515]: I0707 00:05:48.204926 2515 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jul 7 00:05:48.253345 kubelet[2515]: E0707 00:05:48.253275 2515 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-172-234-31-125\" already exists" pod="kube-system/kube-controller-manager-172-234-31-125" Jul 7 00:05:48.257631 kubelet[2515]: E0707 00:05:48.257455 2515 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-172-234-31-125\" already exists" pod="kube-system/kube-apiserver-172-234-31-125" Jul 7 00:05:48.309389 kubelet[2515]: I0707 00:05:48.309321 2515 kubelet_node_status.go:72] "Attempting to register node" node="172-234-31-125" Jul 7 00:05:48.315675 kubelet[2515]: I0707 00:05:48.315466 2515 kubelet_node_status.go:111] "Node was previously registered" node="172-234-31-125" Jul 7 00:05:48.315675 kubelet[2515]: I0707 00:05:48.315544 2515 kubelet_node_status.go:75] "Successfully registered node" node="172-234-31-125" Jul 7 00:05:48.327557 kubelet[2515]: I0707 00:05:48.326883 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e0c092ee8be31a6cfdfccfcbf86f82b6-ca-certs\") pod \"kube-apiserver-172-234-31-125\" (UID: \"e0c092ee8be31a6cfdfccfcbf86f82b6\") " pod="kube-system/kube-apiserver-172-234-31-125" Jul 7 00:05:48.327557 kubelet[2515]: I0707 00:05:48.326910 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/2502a970e91bbd0f9e749dbbd719d057-k8s-certs\") pod \"kube-controller-manager-172-234-31-125\" (UID: \"2502a970e91bbd0f9e749dbbd719d057\") " pod="kube-system/kube-controller-manager-172-234-31-125" Jul 7 00:05:48.327557 kubelet[2515]: I0707 00:05:48.326928 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/2502a970e91bbd0f9e749dbbd719d057-kubeconfig\") pod \"kube-controller-manager-172-234-31-125\" (UID: \"2502a970e91bbd0f9e749dbbd719d057\") " pod="kube-system/kube-controller-manager-172-234-31-125" Jul 7 00:05:48.327557 kubelet[2515]: I0707 00:05:48.326941 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6a123eb34d53cf31ccd27f0ce9e84c2f-kubeconfig\") pod \"kube-scheduler-172-234-31-125\" (UID: \"6a123eb34d53cf31ccd27f0ce9e84c2f\") " pod="kube-system/kube-scheduler-172-234-31-125" Jul 7 00:05:48.327557 kubelet[2515]: I0707 00:05:48.326955 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e0c092ee8be31a6cfdfccfcbf86f82b6-k8s-certs\") pod \"kube-apiserver-172-234-31-125\" (UID: \"e0c092ee8be31a6cfdfccfcbf86f82b6\") " pod="kube-system/kube-apiserver-172-234-31-125" Jul 7 00:05:48.327698 kubelet[2515]: I0707 00:05:48.326970 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e0c092ee8be31a6cfdfccfcbf86f82b6-usr-share-ca-certificates\") pod \"kube-apiserver-172-234-31-125\" (UID: \"e0c092ee8be31a6cfdfccfcbf86f82b6\") " pod="kube-system/kube-apiserver-172-234-31-125" Jul 7 00:05:48.327698 kubelet[2515]: I0707 00:05:48.326985 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/2502a970e91bbd0f9e749dbbd719d057-ca-certs\") pod \"kube-controller-manager-172-234-31-125\" (UID: \"2502a970e91bbd0f9e749dbbd719d057\") " pod="kube-system/kube-controller-manager-172-234-31-125" Jul 7 00:05:48.327698 kubelet[2515]: I0707 00:05:48.326998 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/2502a970e91bbd0f9e749dbbd719d057-flexvolume-dir\") pod \"kube-controller-manager-172-234-31-125\" (UID: \"2502a970e91bbd0f9e749dbbd719d057\") " pod="kube-system/kube-controller-manager-172-234-31-125" Jul 7 00:05:48.327698 kubelet[2515]: I0707 00:05:48.327014 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/2502a970e91bbd0f9e749dbbd719d057-usr-share-ca-certificates\") pod \"kube-controller-manager-172-234-31-125\" (UID: \"2502a970e91bbd0f9e749dbbd719d057\") " pod="kube-system/kube-controller-manager-172-234-31-125" Jul 7 00:05:48.554407 kubelet[2515]: E0707 00:05:48.554334 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:48.557032 kubelet[2515]: E0707 00:05:48.556527 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:48.559849 kubelet[2515]: E0707 00:05:48.558753 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:49.104101 kubelet[2515]: I0707 00:05:49.103818 2515 apiserver.go:52] "Watching apiserver" Jul 7 00:05:49.126171 kubelet[2515]: I0707 00:05:49.126109 2515 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Jul 7 00:05:49.171430 kubelet[2515]: E0707 00:05:49.169230 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:49.179071 kubelet[2515]: E0707 00:05:49.178571 2515 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-172-234-31-125\" already exists" pod="kube-system/kube-controller-manager-172-234-31-125" Jul 7 00:05:49.179071 kubelet[2515]: E0707 00:05:49.178736 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:49.182975 kubelet[2515]: E0707 00:05:49.182939 2515 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-172-234-31-125\" already exists" pod="kube-system/kube-apiserver-172-234-31-125" Jul 7 00:05:49.183275 kubelet[2515]: E0707 00:05:49.183262 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:49.209252 kubelet[2515]: I0707 00:05:49.209146 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-172-234-31-125" podStartSLOduration=3.209129819 podStartE2EDuration="3.209129819s" podCreationTimestamp="2025-07-07 00:05:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-07 00:05:49.203192322 +0000 UTC m=+1.171062645" watchObservedRunningTime="2025-07-07 00:05:49.209129819 +0000 UTC m=+1.177000142" Jul 7 00:05:49.215134 kubelet[2515]: I0707 00:05:49.215101 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-172-234-31-125" podStartSLOduration=3.215073006 podStartE2EDuration="3.215073006s" podCreationTimestamp="2025-07-07 00:05:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-07 00:05:49.209857899 +0000 UTC m=+1.177728212" watchObservedRunningTime="2025-07-07 00:05:49.215073006 +0000 UTC m=+1.182943319" Jul 7 00:05:50.171388 kubelet[2515]: E0707 00:05:50.170446 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:50.171388 kubelet[2515]: E0707 00:05:50.170844 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:51.291749 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Jul 7 00:05:52.663068 kubelet[2515]: E0707 00:05:52.663027 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:52.773484 kubelet[2515]: I0707 00:05:52.773430 2515 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jul 7 00:05:52.773951 containerd[1456]: time="2025-07-07T00:05:52.773872287Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jul 7 00:05:52.774916 kubelet[2515]: I0707 00:05:52.774135 2515 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jul 7 00:05:53.815023 kubelet[2515]: I0707 00:05:53.814164 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-172-234-31-125" podStartSLOduration=5.814146566 podStartE2EDuration="5.814146566s" podCreationTimestamp="2025-07-07 00:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-07 00:05:49.215472686 +0000 UTC m=+1.183343009" watchObservedRunningTime="2025-07-07 00:05:53.814146566 +0000 UTC m=+5.782016879" Jul 7 00:05:53.828752 systemd[1]: Created slice kubepods-besteffort-pod71019118_161b_4fc0_91ac_f9448a9b6f7f.slice - libcontainer container kubepods-besteffort-pod71019118_161b_4fc0_91ac_f9448a9b6f7f.slice. Jul 7 00:05:53.864102 kubelet[2515]: I0707 00:05:53.864044 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/71019118-161b-4fc0-91ac-f9448a9b6f7f-lib-modules\") pod \"kube-proxy-nf25v\" (UID: \"71019118-161b-4fc0-91ac-f9448a9b6f7f\") " pod="kube-system/kube-proxy-nf25v" Jul 7 00:05:53.864284 kubelet[2515]: I0707 00:05:53.864103 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psvds\" (UniqueName: \"kubernetes.io/projected/71019118-161b-4fc0-91ac-f9448a9b6f7f-kube-api-access-psvds\") pod \"kube-proxy-nf25v\" (UID: \"71019118-161b-4fc0-91ac-f9448a9b6f7f\") " pod="kube-system/kube-proxy-nf25v" Jul 7 00:05:53.864284 kubelet[2515]: I0707 00:05:53.864132 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/71019118-161b-4fc0-91ac-f9448a9b6f7f-xtables-lock\") pod \"kube-proxy-nf25v\" (UID: \"71019118-161b-4fc0-91ac-f9448a9b6f7f\") " pod="kube-system/kube-proxy-nf25v" Jul 7 00:05:53.864284 kubelet[2515]: I0707 00:05:53.864158 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/71019118-161b-4fc0-91ac-f9448a9b6f7f-kube-proxy\") pod \"kube-proxy-nf25v\" (UID: \"71019118-161b-4fc0-91ac-f9448a9b6f7f\") " pod="kube-system/kube-proxy-nf25v" Jul 7 00:05:53.947598 systemd[1]: Created slice kubepods-besteffort-pod7e6f832c_fed3_4868_af56_9633aa365a65.slice - libcontainer container kubepods-besteffort-pod7e6f832c_fed3_4868_af56_9633aa365a65.slice. Jul 7 00:05:53.964868 kubelet[2515]: I0707 00:05:53.964782 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/7e6f832c-fed3-4868-af56-9633aa365a65-var-lib-calico\") pod \"tigera-operator-5bf8dfcb4-92txt\" (UID: \"7e6f832c-fed3-4868-af56-9633aa365a65\") " pod="tigera-operator/tigera-operator-5bf8dfcb4-92txt" Jul 7 00:05:53.964950 kubelet[2515]: I0707 00:05:53.964902 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js4jf\" (UniqueName: \"kubernetes.io/projected/7e6f832c-fed3-4868-af56-9633aa365a65-kube-api-access-js4jf\") pod \"tigera-operator-5bf8dfcb4-92txt\" (UID: \"7e6f832c-fed3-4868-af56-9633aa365a65\") " pod="tigera-operator/tigera-operator-5bf8dfcb4-92txt" Jul 7 00:05:54.015821 kubelet[2515]: E0707 00:05:54.015713 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:54.135641 kubelet[2515]: E0707 00:05:54.135609 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:54.137041 containerd[1456]: time="2025-07-07T00:05:54.136983871Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-nf25v,Uid:71019118-161b-4fc0-91ac-f9448a9b6f7f,Namespace:kube-system,Attempt:0,}" Jul 7 00:05:54.164651 containerd[1456]: time="2025-07-07T00:05:54.164168768Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:05:54.164651 containerd[1456]: time="2025-07-07T00:05:54.164230687Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:05:54.164651 containerd[1456]: time="2025-07-07T00:05:54.164246336Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:05:54.164651 containerd[1456]: time="2025-07-07T00:05:54.164395310Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:05:54.176879 kubelet[2515]: E0707 00:05:54.176858 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:54.197636 systemd[1]: Started cri-containerd-833a16dff42c0a3d901609b8cb52c2c9bcc10724219272fa3e65d298214d92d7.scope - libcontainer container 833a16dff42c0a3d901609b8cb52c2c9bcc10724219272fa3e65d298214d92d7. Jul 7 00:05:54.229043 containerd[1456]: time="2025-07-07T00:05:54.229005629Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-nf25v,Uid:71019118-161b-4fc0-91ac-f9448a9b6f7f,Namespace:kube-system,Attempt:0,} returns sandbox id \"833a16dff42c0a3d901609b8cb52c2c9bcc10724219272fa3e65d298214d92d7\"" Jul 7 00:05:54.229785 kubelet[2515]: E0707 00:05:54.229744 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:54.232415 containerd[1456]: time="2025-07-07T00:05:54.232330391Z" level=info msg="CreateContainer within sandbox \"833a16dff42c0a3d901609b8cb52c2c9bcc10724219272fa3e65d298214d92d7\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jul 7 00:05:54.246160 containerd[1456]: time="2025-07-07T00:05:54.246119398Z" level=info msg="CreateContainer within sandbox \"833a16dff42c0a3d901609b8cb52c2c9bcc10724219272fa3e65d298214d92d7\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"0b8d1b50d08fd645f7ccf5b54e6b128c16df618e9ef2a6f816fe173b2ef5c9ae\"" Jul 7 00:05:54.250110 containerd[1456]: time="2025-07-07T00:05:54.247629155Z" level=info msg="StartContainer for \"0b8d1b50d08fd645f7ccf5b54e6b128c16df618e9ef2a6f816fe173b2ef5c9ae\"" Jul 7 00:05:54.254548 containerd[1456]: time="2025-07-07T00:05:54.254525524Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-5bf8dfcb4-92txt,Uid:7e6f832c-fed3-4868-af56-9633aa365a65,Namespace:tigera-operator,Attempt:0,}" Jul 7 00:05:54.278717 containerd[1456]: time="2025-07-07T00:05:54.277908779Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:05:54.278717 containerd[1456]: time="2025-07-07T00:05:54.278055852Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:05:54.278717 containerd[1456]: time="2025-07-07T00:05:54.278071251Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:05:54.280061 containerd[1456]: time="2025-07-07T00:05:54.278773251Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:05:54.284762 systemd[1]: Started cri-containerd-0b8d1b50d08fd645f7ccf5b54e6b128c16df618e9ef2a6f816fe173b2ef5c9ae.scope - libcontainer container 0b8d1b50d08fd645f7ccf5b54e6b128c16df618e9ef2a6f816fe173b2ef5c9ae. Jul 7 00:05:54.303638 systemd[1]: Started cri-containerd-8db1c8c0a4724e7b28c25ca89571b421792939132865ccbe9966697e779ec1a5.scope - libcontainer container 8db1c8c0a4724e7b28c25ca89571b421792939132865ccbe9966697e779ec1a5. Jul 7 00:05:54.330055 containerd[1456]: time="2025-07-07T00:05:54.330013676Z" level=info msg="StartContainer for \"0b8d1b50d08fd645f7ccf5b54e6b128c16df618e9ef2a6f816fe173b2ef5c9ae\" returns successfully" Jul 7 00:05:54.368037 containerd[1456]: time="2025-07-07T00:05:54.366424842Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-5bf8dfcb4-92txt,Uid:7e6f832c-fed3-4868-af56-9633aa365a65,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"8db1c8c0a4724e7b28c25ca89571b421792939132865ccbe9966697e779ec1a5\"" Jul 7 00:05:54.369192 containerd[1456]: time="2025-07-07T00:05:54.369154893Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\"" Jul 7 00:05:55.179576 kubelet[2515]: E0707 00:05:55.179548 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:55.258244 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1694330900.mount: Deactivated successfully. Jul 7 00:05:55.821270 containerd[1456]: time="2025-07-07T00:05:55.821167071Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:55.822292 containerd[1456]: time="2025-07-07T00:05:55.822211684Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.3: active requests=0, bytes read=25056543" Jul 7 00:05:55.823049 containerd[1456]: time="2025-07-07T00:05:55.822759105Z" level=info msg="ImageCreate event name:\"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:55.825105 containerd[1456]: time="2025-07-07T00:05:55.824875558Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:05:55.825820 containerd[1456]: time="2025-07-07T00:05:55.825782981Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.3\" with image id \"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\", repo tag \"quay.io/tigera/operator:v1.38.3\", repo digest \"quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121\", size \"25052538\" in 1.456574585s" Jul 7 00:05:55.825865 containerd[1456]: time="2025-07-07T00:05:55.825818962Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\" returns image reference \"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\"" Jul 7 00:05:55.829011 containerd[1456]: time="2025-07-07T00:05:55.828885963Z" level=info msg="CreateContainer within sandbox \"8db1c8c0a4724e7b28c25ca89571b421792939132865ccbe9966697e779ec1a5\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Jul 7 00:05:55.851516 containerd[1456]: time="2025-07-07T00:05:55.851480861Z" level=info msg="CreateContainer within sandbox \"8db1c8c0a4724e7b28c25ca89571b421792939132865ccbe9966697e779ec1a5\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"967e4261d3b2caced3532e781daa18e1c5476e53e5c905988a400d2ccad66c4c\"" Jul 7 00:05:55.852090 containerd[1456]: time="2025-07-07T00:05:55.851991571Z" level=info msg="StartContainer for \"967e4261d3b2caced3532e781daa18e1c5476e53e5c905988a400d2ccad66c4c\"" Jul 7 00:05:55.878481 systemd[1]: Started cri-containerd-967e4261d3b2caced3532e781daa18e1c5476e53e5c905988a400d2ccad66c4c.scope - libcontainer container 967e4261d3b2caced3532e781daa18e1c5476e53e5c905988a400d2ccad66c4c. Jul 7 00:05:55.906718 containerd[1456]: time="2025-07-07T00:05:55.906617347Z" level=info msg="StartContainer for \"967e4261d3b2caced3532e781daa18e1c5476e53e5c905988a400d2ccad66c4c\" returns successfully" Jul 7 00:05:55.992171 kubelet[2515]: E0707 00:05:55.991543 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:56.022341 kubelet[2515]: I0707 00:05:56.022281 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-nf25v" podStartSLOduration=3.022264752 podStartE2EDuration="3.022264752s" podCreationTimestamp="2025-07-07 00:05:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-07 00:05:55.195276547 +0000 UTC m=+7.163146890" watchObservedRunningTime="2025-07-07 00:05:56.022264752 +0000 UTC m=+7.990135075" Jul 7 00:05:56.185847 kubelet[2515]: E0707 00:05:56.185734 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:05:56.217326 kubelet[2515]: I0707 00:05:56.216987 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-5bf8dfcb4-92txt" podStartSLOduration=1.757997563 podStartE2EDuration="3.216969184s" podCreationTimestamp="2025-07-07 00:05:53 +0000 UTC" firstStartedPulling="2025-07-07 00:05:54.368227972 +0000 UTC m=+6.336098295" lastFinishedPulling="2025-07-07 00:05:55.827199593 +0000 UTC m=+7.795069916" observedRunningTime="2025-07-07 00:05:56.207117741 +0000 UTC m=+8.174988064" watchObservedRunningTime="2025-07-07 00:05:56.216969184 +0000 UTC m=+8.184839507" Jul 7 00:06:01.354603 sudo[1677]: pam_unix(sudo:session): session closed for user root Jul 7 00:06:01.414540 sshd[1674]: pam_unix(sshd:session): session closed for user core Jul 7 00:06:01.418792 systemd[1]: sshd@6-172.234.31.125:22-147.75.109.163:51908.service: Deactivated successfully. Jul 7 00:06:01.422120 systemd[1]: session-7.scope: Deactivated successfully. Jul 7 00:06:01.422958 systemd[1]: session-7.scope: Consumed 4.139s CPU time, 158.1M memory peak, 0B memory swap peak. Jul 7 00:06:01.424700 systemd-logind[1440]: Session 7 logged out. Waiting for processes to exit. Jul 7 00:06:01.426845 systemd-logind[1440]: Removed session 7. Jul 7 00:06:02.667835 kubelet[2515]: E0707 00:06:02.667466 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:04.410910 systemd[1]: Created slice kubepods-besteffort-pod8d3e2623_7ebf_4f41_a90b_931d1b3d12c3.slice - libcontainer container kubepods-besteffort-pod8d3e2623_7ebf_4f41_a90b_931d1b3d12c3.slice. Jul 7 00:06:04.530491 kubelet[2515]: I0707 00:06:04.530448 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d3e2623-7ebf-4f41-a90b-931d1b3d12c3-tigera-ca-bundle\") pod \"calico-typha-5b487879c5-hx4jl\" (UID: \"8d3e2623-7ebf-4f41-a90b-931d1b3d12c3\") " pod="calico-system/calico-typha-5b487879c5-hx4jl" Jul 7 00:06:04.530491 kubelet[2515]: I0707 00:06:04.530493 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8znd\" (UniqueName: \"kubernetes.io/projected/8d3e2623-7ebf-4f41-a90b-931d1b3d12c3-kube-api-access-f8znd\") pod \"calico-typha-5b487879c5-hx4jl\" (UID: \"8d3e2623-7ebf-4f41-a90b-931d1b3d12c3\") " pod="calico-system/calico-typha-5b487879c5-hx4jl" Jul 7 00:06:04.530946 kubelet[2515]: I0707 00:06:04.530519 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/8d3e2623-7ebf-4f41-a90b-931d1b3d12c3-typha-certs\") pod \"calico-typha-5b487879c5-hx4jl\" (UID: \"8d3e2623-7ebf-4f41-a90b-931d1b3d12c3\") " pod="calico-system/calico-typha-5b487879c5-hx4jl" Jul 7 00:06:04.717184 kubelet[2515]: E0707 00:06:04.714825 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:04.717273 containerd[1456]: time="2025-07-07T00:06:04.716146001Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5b487879c5-hx4jl,Uid:8d3e2623-7ebf-4f41-a90b-931d1b3d12c3,Namespace:calico-system,Attempt:0,}" Jul 7 00:06:04.745187 containerd[1456]: time="2025-07-07T00:06:04.745091845Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:06:04.749405 containerd[1456]: time="2025-07-07T00:06:04.747437616Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:06:04.749405 containerd[1456]: time="2025-07-07T00:06:04.747491483Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:04.749405 containerd[1456]: time="2025-07-07T00:06:04.747584134Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:04.775837 systemd[1]: Started cri-containerd-2b4c82b9225cf6292971cddf2b87381dfbe5ba03a00d07c8e495896676cfe46f.scope - libcontainer container 2b4c82b9225cf6292971cddf2b87381dfbe5ba03a00d07c8e495896676cfe46f. Jul 7 00:06:04.824880 containerd[1456]: time="2025-07-07T00:06:04.824822416Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-5b487879c5-hx4jl,Uid:8d3e2623-7ebf-4f41-a90b-931d1b3d12c3,Namespace:calico-system,Attempt:0,} returns sandbox id \"2b4c82b9225cf6292971cddf2b87381dfbe5ba03a00d07c8e495896676cfe46f\"" Jul 7 00:06:04.825717 kubelet[2515]: E0707 00:06:04.825667 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:04.829964 containerd[1456]: time="2025-07-07T00:06:04.829646670Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\"" Jul 7 00:06:04.847280 systemd[1]: Created slice kubepods-besteffort-pod89ba0f01_8ed5_42a4_b954_6ab2fd634387.slice - libcontainer container kubepods-besteffort-pod89ba0f01_8ed5_42a4_b954_6ab2fd634387.slice. Jul 7 00:06:04.932410 kubelet[2515]: I0707 00:06:04.932359 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/89ba0f01-8ed5-42a4-b954-6ab2fd634387-node-certs\") pod \"calico-node-cvl5d\" (UID: \"89ba0f01-8ed5-42a4-b954-6ab2fd634387\") " pod="calico-system/calico-node-cvl5d" Jul 7 00:06:04.932410 kubelet[2515]: I0707 00:06:04.932413 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/89ba0f01-8ed5-42a4-b954-6ab2fd634387-lib-modules\") pod \"calico-node-cvl5d\" (UID: \"89ba0f01-8ed5-42a4-b954-6ab2fd634387\") " pod="calico-system/calico-node-cvl5d" Jul 7 00:06:04.932724 kubelet[2515]: I0707 00:06:04.932431 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j47nf\" (UniqueName: \"kubernetes.io/projected/89ba0f01-8ed5-42a4-b954-6ab2fd634387-kube-api-access-j47nf\") pod \"calico-node-cvl5d\" (UID: \"89ba0f01-8ed5-42a4-b954-6ab2fd634387\") " pod="calico-system/calico-node-cvl5d" Jul 7 00:06:04.932724 kubelet[2515]: I0707 00:06:04.932448 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/89ba0f01-8ed5-42a4-b954-6ab2fd634387-policysync\") pod \"calico-node-cvl5d\" (UID: \"89ba0f01-8ed5-42a4-b954-6ab2fd634387\") " pod="calico-system/calico-node-cvl5d" Jul 7 00:06:04.932724 kubelet[2515]: I0707 00:06:04.932465 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89ba0f01-8ed5-42a4-b954-6ab2fd634387-tigera-ca-bundle\") pod \"calico-node-cvl5d\" (UID: \"89ba0f01-8ed5-42a4-b954-6ab2fd634387\") " pod="calico-system/calico-node-cvl5d" Jul 7 00:06:04.932724 kubelet[2515]: I0707 00:06:04.932480 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/89ba0f01-8ed5-42a4-b954-6ab2fd634387-cni-bin-dir\") pod \"calico-node-cvl5d\" (UID: \"89ba0f01-8ed5-42a4-b954-6ab2fd634387\") " pod="calico-system/calico-node-cvl5d" Jul 7 00:06:04.932724 kubelet[2515]: I0707 00:06:04.932497 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/89ba0f01-8ed5-42a4-b954-6ab2fd634387-var-lib-calico\") pod \"calico-node-cvl5d\" (UID: \"89ba0f01-8ed5-42a4-b954-6ab2fd634387\") " pod="calico-system/calico-node-cvl5d" Jul 7 00:06:04.932850 kubelet[2515]: I0707 00:06:04.932513 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/89ba0f01-8ed5-42a4-b954-6ab2fd634387-xtables-lock\") pod \"calico-node-cvl5d\" (UID: \"89ba0f01-8ed5-42a4-b954-6ab2fd634387\") " pod="calico-system/calico-node-cvl5d" Jul 7 00:06:04.932850 kubelet[2515]: I0707 00:06:04.932530 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/89ba0f01-8ed5-42a4-b954-6ab2fd634387-cni-net-dir\") pod \"calico-node-cvl5d\" (UID: \"89ba0f01-8ed5-42a4-b954-6ab2fd634387\") " pod="calico-system/calico-node-cvl5d" Jul 7 00:06:04.932850 kubelet[2515]: I0707 00:06:04.932546 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/89ba0f01-8ed5-42a4-b954-6ab2fd634387-flexvol-driver-host\") pod \"calico-node-cvl5d\" (UID: \"89ba0f01-8ed5-42a4-b954-6ab2fd634387\") " pod="calico-system/calico-node-cvl5d" Jul 7 00:06:04.932850 kubelet[2515]: I0707 00:06:04.932561 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/89ba0f01-8ed5-42a4-b954-6ab2fd634387-cni-log-dir\") pod \"calico-node-cvl5d\" (UID: \"89ba0f01-8ed5-42a4-b954-6ab2fd634387\") " pod="calico-system/calico-node-cvl5d" Jul 7 00:06:04.932850 kubelet[2515]: I0707 00:06:04.932576 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/89ba0f01-8ed5-42a4-b954-6ab2fd634387-var-run-calico\") pod \"calico-node-cvl5d\" (UID: \"89ba0f01-8ed5-42a4-b954-6ab2fd634387\") " pod="calico-system/calico-node-cvl5d" Jul 7 00:06:05.041167 kubelet[2515]: E0707 00:06:05.040351 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.041167 kubelet[2515]: W0707 00:06:05.040398 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.041167 kubelet[2515]: E0707 00:06:05.040420 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.048264 kubelet[2515]: E0707 00:06:05.048248 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.048356 kubelet[2515]: W0707 00:06:05.048343 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.048661 kubelet[2515]: E0707 00:06:05.048625 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.122048 kubelet[2515]: E0707 00:06:05.122012 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-9knjc" podUID="2001a2e7-8165-484d-b2ef-2ba8a539af9a" Jul 7 00:06:05.134387 kubelet[2515]: E0707 00:06:05.134338 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.134387 kubelet[2515]: W0707 00:06:05.134375 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.134493 kubelet[2515]: E0707 00:06:05.134395 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.134751 kubelet[2515]: E0707 00:06:05.134730 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.134751 kubelet[2515]: W0707 00:06:05.134745 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.135092 kubelet[2515]: E0707 00:06:05.134864 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.135656 kubelet[2515]: E0707 00:06:05.135632 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.135656 kubelet[2515]: W0707 00:06:05.135647 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.135726 kubelet[2515]: E0707 00:06:05.135657 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.136358 kubelet[2515]: E0707 00:06:05.136335 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.136358 kubelet[2515]: W0707 00:06:05.136350 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.136436 kubelet[2515]: E0707 00:06:05.136375 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.137641 kubelet[2515]: E0707 00:06:05.137613 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.138135 kubelet[2515]: W0707 00:06:05.137632 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.138135 kubelet[2515]: E0707 00:06:05.138132 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.138482 kubelet[2515]: E0707 00:06:05.138450 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.138482 kubelet[2515]: W0707 00:06:05.138467 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.138482 kubelet[2515]: E0707 00:06:05.138476 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.138808 kubelet[2515]: E0707 00:06:05.138770 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.138862 kubelet[2515]: W0707 00:06:05.138800 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.138862 kubelet[2515]: E0707 00:06:05.138855 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.140258 kubelet[2515]: E0707 00:06:05.140230 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.140258 kubelet[2515]: W0707 00:06:05.140247 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.140322 kubelet[2515]: E0707 00:06:05.140268 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.140542 kubelet[2515]: E0707 00:06:05.140510 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.140542 kubelet[2515]: W0707 00:06:05.140526 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.140542 kubelet[2515]: E0707 00:06:05.140535 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.140883 kubelet[2515]: E0707 00:06:05.140845 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.140883 kubelet[2515]: W0707 00:06:05.140861 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.140883 kubelet[2515]: E0707 00:06:05.140869 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.141147 kubelet[2515]: E0707 00:06:05.141127 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.141147 kubelet[2515]: W0707 00:06:05.141141 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.141209 kubelet[2515]: E0707 00:06:05.141149 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.141581 kubelet[2515]: E0707 00:06:05.141435 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.141581 kubelet[2515]: W0707 00:06:05.141452 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.141581 kubelet[2515]: E0707 00:06:05.141470 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.141812 kubelet[2515]: E0707 00:06:05.141801 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.141864 kubelet[2515]: W0707 00:06:05.141854 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.142242 kubelet[2515]: E0707 00:06:05.141908 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.143361 kubelet[2515]: E0707 00:06:05.142779 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.143361 kubelet[2515]: W0707 00:06:05.142791 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.143361 kubelet[2515]: E0707 00:06:05.142814 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.143623 kubelet[2515]: E0707 00:06:05.143612 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.143844 kubelet[2515]: W0707 00:06:05.143667 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.143844 kubelet[2515]: E0707 00:06:05.143680 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.144135 kubelet[2515]: E0707 00:06:05.144045 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.144135 kubelet[2515]: W0707 00:06:05.144056 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.144135 kubelet[2515]: E0707 00:06:05.144066 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.144422 kubelet[2515]: E0707 00:06:05.144331 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.144422 kubelet[2515]: W0707 00:06:05.144341 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.144422 kubelet[2515]: E0707 00:06:05.144349 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.144821 kubelet[2515]: E0707 00:06:05.144739 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.144821 kubelet[2515]: W0707 00:06:05.144751 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.144821 kubelet[2515]: E0707 00:06:05.144760 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.145514 kubelet[2515]: E0707 00:06:05.145439 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.145514 kubelet[2515]: W0707 00:06:05.145452 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.145514 kubelet[2515]: E0707 00:06:05.145462 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.146093 kubelet[2515]: E0707 00:06:05.145951 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.146093 kubelet[2515]: W0707 00:06:05.145962 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.146093 kubelet[2515]: E0707 00:06:05.145995 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.153821 containerd[1456]: time="2025-07-07T00:06:05.153460031Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-cvl5d,Uid:89ba0f01-8ed5-42a4-b954-6ab2fd634387,Namespace:calico-system,Attempt:0,}" Jul 7 00:06:05.176594 containerd[1456]: time="2025-07-07T00:06:05.176512406Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:06:05.176890 containerd[1456]: time="2025-07-07T00:06:05.176840528Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:06:05.177249 containerd[1456]: time="2025-07-07T00:06:05.177188955Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:05.177633 containerd[1456]: time="2025-07-07T00:06:05.177554398Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:05.196531 systemd[1]: Started cri-containerd-6261b6a8cda854340ae6bb4e54c2ab4e67e0c6461ca9da4594865c0bfe9bb534.scope - libcontainer container 6261b6a8cda854340ae6bb4e54c2ab4e67e0c6461ca9da4594865c0bfe9bb534. Jul 7 00:06:05.232090 containerd[1456]: time="2025-07-07T00:06:05.231815900Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-cvl5d,Uid:89ba0f01-8ed5-42a4-b954-6ab2fd634387,Namespace:calico-system,Attempt:0,} returns sandbox id \"6261b6a8cda854340ae6bb4e54c2ab4e67e0c6461ca9da4594865c0bfe9bb534\"" Jul 7 00:06:05.234088 kubelet[2515]: E0707 00:06:05.234012 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.234088 kubelet[2515]: W0707 00:06:05.234028 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.234088 kubelet[2515]: E0707 00:06:05.234044 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.234474 kubelet[2515]: I0707 00:06:05.234243 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnl92\" (UniqueName: \"kubernetes.io/projected/2001a2e7-8165-484d-b2ef-2ba8a539af9a-kube-api-access-nnl92\") pod \"csi-node-driver-9knjc\" (UID: \"2001a2e7-8165-484d-b2ef-2ba8a539af9a\") " pod="calico-system/csi-node-driver-9knjc" Jul 7 00:06:05.234783 kubelet[2515]: E0707 00:06:05.234743 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.234783 kubelet[2515]: W0707 00:06:05.234754 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.234982 kubelet[2515]: E0707 00:06:05.234957 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.235123 kubelet[2515]: E0707 00:06:05.235099 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.235123 kubelet[2515]: W0707 00:06:05.235110 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.235274 kubelet[2515]: E0707 00:06:05.235199 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.235274 kubelet[2515]: I0707 00:06:05.235217 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2001a2e7-8165-484d-b2ef-2ba8a539af9a-socket-dir\") pod \"csi-node-driver-9knjc\" (UID: \"2001a2e7-8165-484d-b2ef-2ba8a539af9a\") " pod="calico-system/csi-node-driver-9knjc" Jul 7 00:06:05.235640 kubelet[2515]: E0707 00:06:05.235538 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.235640 kubelet[2515]: W0707 00:06:05.235550 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.235640 kubelet[2515]: E0707 00:06:05.235559 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.236116 kubelet[2515]: E0707 00:06:05.236017 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.236116 kubelet[2515]: W0707 00:06:05.236028 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.236116 kubelet[2515]: E0707 00:06:05.236041 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.236443 kubelet[2515]: E0707 00:06:05.236309 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.236443 kubelet[2515]: W0707 00:06:05.236320 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.236443 kubelet[2515]: E0707 00:06:05.236333 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.236603 kubelet[2515]: E0707 00:06:05.236592 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.236724 kubelet[2515]: W0707 00:06:05.236637 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.236724 kubelet[2515]: E0707 00:06:05.236648 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.236724 kubelet[2515]: I0707 00:06:05.236675 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/2001a2e7-8165-484d-b2ef-2ba8a539af9a-varrun\") pod \"csi-node-driver-9knjc\" (UID: \"2001a2e7-8165-484d-b2ef-2ba8a539af9a\") " pod="calico-system/csi-node-driver-9knjc" Jul 7 00:06:05.237065 kubelet[2515]: E0707 00:06:05.236978 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.237065 kubelet[2515]: W0707 00:06:05.236988 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.237065 kubelet[2515]: E0707 00:06:05.237010 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.237065 kubelet[2515]: I0707 00:06:05.237024 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2001a2e7-8165-484d-b2ef-2ba8a539af9a-registration-dir\") pod \"csi-node-driver-9knjc\" (UID: \"2001a2e7-8165-484d-b2ef-2ba8a539af9a\") " pod="calico-system/csi-node-driver-9knjc" Jul 7 00:06:05.237503 kubelet[2515]: E0707 00:06:05.237413 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.237503 kubelet[2515]: W0707 00:06:05.237426 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.237503 kubelet[2515]: E0707 00:06:05.237447 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.237503 kubelet[2515]: I0707 00:06:05.237467 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2001a2e7-8165-484d-b2ef-2ba8a539af9a-kubelet-dir\") pod \"csi-node-driver-9knjc\" (UID: \"2001a2e7-8165-484d-b2ef-2ba8a539af9a\") " pod="calico-system/csi-node-driver-9knjc" Jul 7 00:06:05.237981 kubelet[2515]: E0707 00:06:05.237871 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.237981 kubelet[2515]: W0707 00:06:05.237882 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.237981 kubelet[2515]: E0707 00:06:05.237895 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.238182 kubelet[2515]: E0707 00:06:05.238152 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.238182 kubelet[2515]: W0707 00:06:05.238162 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.238420 kubelet[2515]: E0707 00:06:05.238300 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.238550 kubelet[2515]: E0707 00:06:05.238539 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.238643 kubelet[2515]: W0707 00:06:05.238584 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.238773 kubelet[2515]: E0707 00:06:05.238701 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.238901 kubelet[2515]: E0707 00:06:05.238890 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.238998 kubelet[2515]: W0707 00:06:05.238938 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.238998 kubelet[2515]: E0707 00:06:05.238963 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.239281 kubelet[2515]: E0707 00:06:05.239252 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.239281 kubelet[2515]: W0707 00:06:05.239261 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.239281 kubelet[2515]: E0707 00:06:05.239270 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.239651 kubelet[2515]: E0707 00:06:05.239603 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.239651 kubelet[2515]: W0707 00:06:05.239613 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.239651 kubelet[2515]: E0707 00:06:05.239623 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.338549 kubelet[2515]: E0707 00:06:05.338428 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.338549 kubelet[2515]: W0707 00:06:05.338450 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.338549 kubelet[2515]: E0707 00:06:05.338470 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.340825 kubelet[2515]: E0707 00:06:05.340083 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.340825 kubelet[2515]: W0707 00:06:05.340099 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.340825 kubelet[2515]: E0707 00:06:05.340120 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.340825 kubelet[2515]: E0707 00:06:05.340562 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.340825 kubelet[2515]: W0707 00:06:05.340570 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.340825 kubelet[2515]: E0707 00:06:05.340592 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.340825 kubelet[2515]: E0707 00:06:05.340830 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.341001 kubelet[2515]: W0707 00:06:05.340839 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.341001 kubelet[2515]: E0707 00:06:05.340860 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.341131 kubelet[2515]: E0707 00:06:05.341093 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.341131 kubelet[2515]: W0707 00:06:05.341109 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.341509 kubelet[2515]: E0707 00:06:05.341130 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.341705 kubelet[2515]: E0707 00:06:05.341591 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.341705 kubelet[2515]: W0707 00:06:05.341604 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.341705 kubelet[2515]: E0707 00:06:05.341615 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.341927 kubelet[2515]: E0707 00:06:05.341823 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.341927 kubelet[2515]: W0707 00:06:05.341835 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.341979 kubelet[2515]: E0707 00:06:05.341932 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.342149 kubelet[2515]: E0707 00:06:05.342058 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.342149 kubelet[2515]: W0707 00:06:05.342070 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.342149 kubelet[2515]: E0707 00:06:05.342148 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.342398 kubelet[2515]: E0707 00:06:05.342286 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.342398 kubelet[2515]: W0707 00:06:05.342299 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.342398 kubelet[2515]: E0707 00:06:05.342392 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.342637 kubelet[2515]: E0707 00:06:05.342513 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.342637 kubelet[2515]: W0707 00:06:05.342525 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.342637 kubelet[2515]: E0707 00:06:05.342619 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.342833 kubelet[2515]: E0707 00:06:05.342731 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.342833 kubelet[2515]: W0707 00:06:05.342743 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.342833 kubelet[2515]: E0707 00:06:05.342813 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.343040 kubelet[2515]: E0707 00:06:05.342927 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.343040 kubelet[2515]: W0707 00:06:05.342938 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.343040 kubelet[2515]: E0707 00:06:05.342948 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.343306 kubelet[2515]: E0707 00:06:05.343192 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.343306 kubelet[2515]: W0707 00:06:05.343204 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.343306 kubelet[2515]: E0707 00:06:05.343223 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.343604 kubelet[2515]: E0707 00:06:05.343481 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.343604 kubelet[2515]: W0707 00:06:05.343496 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.343604 kubelet[2515]: E0707 00:06:05.343519 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.343767 kubelet[2515]: E0707 00:06:05.343708 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.343767 kubelet[2515]: W0707 00:06:05.343716 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.343767 kubelet[2515]: E0707 00:06:05.343788 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.344104 kubelet[2515]: E0707 00:06:05.344088 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.344104 kubelet[2515]: W0707 00:06:05.344100 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.344225 kubelet[2515]: E0707 00:06:05.344204 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.344525 kubelet[2515]: E0707 00:06:05.344509 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.344525 kubelet[2515]: W0707 00:06:05.344522 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.344602 kubelet[2515]: E0707 00:06:05.344587 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.344788 kubelet[2515]: E0707 00:06:05.344773 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.344788 kubelet[2515]: W0707 00:06:05.344784 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.344992 kubelet[2515]: E0707 00:06:05.344905 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.345035 kubelet[2515]: E0707 00:06:05.345018 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.345035 kubelet[2515]: W0707 00:06:05.345033 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.345508 kubelet[2515]: E0707 00:06:05.345301 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.346350 kubelet[2515]: E0707 00:06:05.346316 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.346350 kubelet[2515]: W0707 00:06:05.346333 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.346483 kubelet[2515]: E0707 00:06:05.346440 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.346796 kubelet[2515]: E0707 00:06:05.346773 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.346796 kubelet[2515]: W0707 00:06:05.346788 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.346962 kubelet[2515]: E0707 00:06:05.346912 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.347384 kubelet[2515]: E0707 00:06:05.347267 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.347384 kubelet[2515]: W0707 00:06:05.347283 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.347442 kubelet[2515]: E0707 00:06:05.347400 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.348288 kubelet[2515]: E0707 00:06:05.348111 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.348288 kubelet[2515]: W0707 00:06:05.348125 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.348288 kubelet[2515]: E0707 00:06:05.348144 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.348610 kubelet[2515]: E0707 00:06:05.348474 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.348610 kubelet[2515]: W0707 00:06:05.348499 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.348610 kubelet[2515]: E0707 00:06:05.348520 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.348791 kubelet[2515]: E0707 00:06:05.348732 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.348791 kubelet[2515]: W0707 00:06:05.348740 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.348791 kubelet[2515]: E0707 00:06:05.348748 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.355308 kubelet[2515]: E0707 00:06:05.355291 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:05.355308 kubelet[2515]: W0707 00:06:05.355304 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:05.355393 kubelet[2515]: E0707 00:06:05.355315 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:05.741759 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3173801737.mount: Deactivated successfully. Jul 7 00:06:06.207781 update_engine[1441]: I20250707 00:06:06.207376 1441 update_attempter.cc:509] Updating boot flags... Jul 7 00:06:06.284800 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 34 scanned by (udev-worker) (3089) Jul 7 00:06:06.416463 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 34 scanned by (udev-worker) (3093) Jul 7 00:06:06.715610 containerd[1456]: time="2025-07-07T00:06:06.715567368Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:06.716588 containerd[1456]: time="2025-07-07T00:06:06.716375471Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.2: active requests=0, bytes read=35233364" Jul 7 00:06:06.717388 containerd[1456]: time="2025-07-07T00:06:06.717063730Z" level=info msg="ImageCreate event name:\"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:06.718825 containerd[1456]: time="2025-07-07T00:06:06.718775684Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:06.719244 containerd[1456]: time="2025-07-07T00:06:06.719205358Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.2\" with image id \"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af\", size \"35233218\" in 1.889518324s" Jul 7 00:06:06.719244 containerd[1456]: time="2025-07-07T00:06:06.719241448Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\" returns image reference \"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\"" Jul 7 00:06:06.720410 containerd[1456]: time="2025-07-07T00:06:06.720382737Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\"" Jul 7 00:06:06.733170 containerd[1456]: time="2025-07-07T00:06:06.731690921Z" level=info msg="CreateContainer within sandbox \"2b4c82b9225cf6292971cddf2b87381dfbe5ba03a00d07c8e495896676cfe46f\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Jul 7 00:06:06.755441 containerd[1456]: time="2025-07-07T00:06:06.755414146Z" level=info msg="CreateContainer within sandbox \"2b4c82b9225cf6292971cddf2b87381dfbe5ba03a00d07c8e495896676cfe46f\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"c232a2676962e5d25a8bb36519ed101d47aa24361d3d4e22a0c71bd5abd49833\"" Jul 7 00:06:06.756233 containerd[1456]: time="2025-07-07T00:06:06.756189270Z" level=info msg="StartContainer for \"c232a2676962e5d25a8bb36519ed101d47aa24361d3d4e22a0c71bd5abd49833\"" Jul 7 00:06:06.795623 systemd[1]: Started cri-containerd-c232a2676962e5d25a8bb36519ed101d47aa24361d3d4e22a0c71bd5abd49833.scope - libcontainer container c232a2676962e5d25a8bb36519ed101d47aa24361d3d4e22a0c71bd5abd49833. Jul 7 00:06:06.838089 containerd[1456]: time="2025-07-07T00:06:06.837656207Z" level=info msg="StartContainer for \"c232a2676962e5d25a8bb36519ed101d47aa24361d3d4e22a0c71bd5abd49833\" returns successfully" Jul 7 00:06:07.144149 kubelet[2515]: E0707 00:06:07.144117 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-9knjc" podUID="2001a2e7-8165-484d-b2ef-2ba8a539af9a" Jul 7 00:06:07.218063 kubelet[2515]: E0707 00:06:07.218019 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:07.259618 kubelet[2515]: E0707 00:06:07.259592 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.259685 kubelet[2515]: W0707 00:06:07.259617 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.259685 kubelet[2515]: E0707 00:06:07.259640 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.259967 kubelet[2515]: E0707 00:06:07.259951 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.259967 kubelet[2515]: W0707 00:06:07.259965 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.260063 kubelet[2515]: E0707 00:06:07.259978 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.260295 kubelet[2515]: E0707 00:06:07.260264 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.260295 kubelet[2515]: W0707 00:06:07.260282 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.260295 kubelet[2515]: E0707 00:06:07.260302 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.260791 kubelet[2515]: E0707 00:06:07.260774 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.260869 kubelet[2515]: W0707 00:06:07.260842 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.260869 kubelet[2515]: E0707 00:06:07.260860 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.261111 kubelet[2515]: E0707 00:06:07.261090 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.261111 kubelet[2515]: W0707 00:06:07.261104 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.261111 kubelet[2515]: E0707 00:06:07.261114 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.261387 kubelet[2515]: E0707 00:06:07.261350 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.261387 kubelet[2515]: W0707 00:06:07.261383 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.261455 kubelet[2515]: E0707 00:06:07.261393 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.261816 kubelet[2515]: E0707 00:06:07.261796 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.261877 kubelet[2515]: W0707 00:06:07.261809 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.261877 kubelet[2515]: E0707 00:06:07.261866 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.262175 kubelet[2515]: E0707 00:06:07.262141 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.262175 kubelet[2515]: W0707 00:06:07.262153 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.262257 kubelet[2515]: E0707 00:06:07.262194 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.262708 kubelet[2515]: E0707 00:06:07.262686 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.262708 kubelet[2515]: W0707 00:06:07.262706 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.262792 kubelet[2515]: E0707 00:06:07.262721 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.262996 kubelet[2515]: E0707 00:06:07.262963 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.262996 kubelet[2515]: W0707 00:06:07.262978 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.263084 kubelet[2515]: E0707 00:06:07.262994 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.263283 kubelet[2515]: E0707 00:06:07.263265 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.263453 kubelet[2515]: W0707 00:06:07.263314 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.263453 kubelet[2515]: E0707 00:06:07.263400 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.263967 kubelet[2515]: E0707 00:06:07.263850 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.263967 kubelet[2515]: W0707 00:06:07.263864 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.263967 kubelet[2515]: E0707 00:06:07.263874 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.264267 kubelet[2515]: E0707 00:06:07.264248 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.264303 kubelet[2515]: W0707 00:06:07.264266 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.264303 kubelet[2515]: E0707 00:06:07.264280 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.264683 kubelet[2515]: E0707 00:06:07.264667 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.264683 kubelet[2515]: W0707 00:06:07.264680 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.264683 kubelet[2515]: E0707 00:06:07.264691 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.264922 kubelet[2515]: E0707 00:06:07.264902 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.264922 kubelet[2515]: W0707 00:06:07.264917 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.265009 kubelet[2515]: E0707 00:06:07.264928 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.340749 containerd[1456]: time="2025-07-07T00:06:07.340686036Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:07.341315 containerd[1456]: time="2025-07-07T00:06:07.341280467Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2: active requests=0, bytes read=4446956" Jul 7 00:06:07.341888 containerd[1456]: time="2025-07-07T00:06:07.341655879Z" level=info msg="ImageCreate event name:\"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:07.343462 containerd[1456]: time="2025-07-07T00:06:07.343436801Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:07.343908 containerd[1456]: time="2025-07-07T00:06:07.343865797Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" with image id \"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b\", size \"5939619\" in 623.454892ms" Jul 7 00:06:07.343908 containerd[1456]: time="2025-07-07T00:06:07.343900246Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" returns image reference \"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\"" Jul 7 00:06:07.346530 containerd[1456]: time="2025-07-07T00:06:07.346428930Z" level=info msg="CreateContainer within sandbox \"6261b6a8cda854340ae6bb4e54c2ab4e67e0c6461ca9da4594865c0bfe9bb534\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Jul 7 00:06:07.358247 containerd[1456]: time="2025-07-07T00:06:07.358217479Z" level=info msg="CreateContainer within sandbox \"6261b6a8cda854340ae6bb4e54c2ab4e67e0c6461ca9da4594865c0bfe9bb534\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"00172b9dd5396d046495f0783e1fb61d3d775c7d06034111304c14b63e48c4e4\"" Jul 7 00:06:07.359233 containerd[1456]: time="2025-07-07T00:06:07.359192612Z" level=info msg="StartContainer for \"00172b9dd5396d046495f0783e1fb61d3d775c7d06034111304c14b63e48c4e4\"" Jul 7 00:06:07.360837 kubelet[2515]: E0707 00:06:07.360506 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.360837 kubelet[2515]: W0707 00:06:07.360522 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.360837 kubelet[2515]: E0707 00:06:07.360744 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.361222 kubelet[2515]: E0707 00:06:07.361184 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.361475 kubelet[2515]: W0707 00:06:07.361461 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.361572 kubelet[2515]: E0707 00:06:07.361560 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.361915 kubelet[2515]: E0707 00:06:07.361888 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.362067 kubelet[2515]: W0707 00:06:07.361964 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.362067 kubelet[2515]: E0707 00:06:07.361978 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.362316 kubelet[2515]: E0707 00:06:07.362286 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.362455 kubelet[2515]: W0707 00:06:07.362379 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.362455 kubelet[2515]: E0707 00:06:07.362401 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.362836 kubelet[2515]: E0707 00:06:07.362730 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.362836 kubelet[2515]: W0707 00:06:07.362741 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.362989 kubelet[2515]: E0707 00:06:07.362923 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.363202 kubelet[2515]: E0707 00:06:07.363123 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.363202 kubelet[2515]: W0707 00:06:07.363133 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.363432 kubelet[2515]: E0707 00:06:07.363287 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.363562 kubelet[2515]: E0707 00:06:07.363551 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.363680 kubelet[2515]: W0707 00:06:07.363636 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.363934 kubelet[2515]: E0707 00:06:07.363809 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.364097 kubelet[2515]: E0707 00:06:07.364086 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.364260 kubelet[2515]: W0707 00:06:07.364168 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.364260 kubelet[2515]: E0707 00:06:07.364211 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.364809 kubelet[2515]: E0707 00:06:07.364692 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.364809 kubelet[2515]: W0707 00:06:07.364703 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.364809 kubelet[2515]: E0707 00:06:07.364733 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.365235 kubelet[2515]: E0707 00:06:07.365173 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.365323 kubelet[2515]: W0707 00:06:07.365185 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.365491 kubelet[2515]: E0707 00:06:07.365409 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.365993 kubelet[2515]: E0707 00:06:07.365954 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.365993 kubelet[2515]: W0707 00:06:07.365972 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.366226 kubelet[2515]: E0707 00:06:07.366138 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.366424 kubelet[2515]: E0707 00:06:07.366412 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.366721 kubelet[2515]: W0707 00:06:07.366472 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.366782 kubelet[2515]: E0707 00:06:07.366769 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.367142 kubelet[2515]: E0707 00:06:07.367130 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.367200 kubelet[2515]: W0707 00:06:07.367189 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.367412 kubelet[2515]: E0707 00:06:07.367400 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.367734 kubelet[2515]: E0707 00:06:07.367488 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.367795 kubelet[2515]: W0707 00:06:07.367779 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.368042 kubelet[2515]: E0707 00:06:07.367905 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.368235 kubelet[2515]: E0707 00:06:07.368223 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.368288 kubelet[2515]: W0707 00:06:07.368278 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.368450 kubelet[2515]: E0707 00:06:07.368325 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.368807 kubelet[2515]: E0707 00:06:07.368795 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.368864 kubelet[2515]: W0707 00:06:07.368853 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.368987 kubelet[2515]: E0707 00:06:07.368922 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.369607 kubelet[2515]: E0707 00:06:07.369484 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.369607 kubelet[2515]: W0707 00:06:07.369495 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.369607 kubelet[2515]: E0707 00:06:07.369507 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.369994 kubelet[2515]: E0707 00:06:07.369929 2515 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 7 00:06:07.369994 kubelet[2515]: W0707 00:06:07.369940 2515 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 7 00:06:07.369994 kubelet[2515]: E0707 00:06:07.369975 2515 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 7 00:06:07.393686 systemd[1]: Started cri-containerd-00172b9dd5396d046495f0783e1fb61d3d775c7d06034111304c14b63e48c4e4.scope - libcontainer container 00172b9dd5396d046495f0783e1fb61d3d775c7d06034111304c14b63e48c4e4. Jul 7 00:06:07.428427 containerd[1456]: time="2025-07-07T00:06:07.427649549Z" level=info msg="StartContainer for \"00172b9dd5396d046495f0783e1fb61d3d775c7d06034111304c14b63e48c4e4\" returns successfully" Jul 7 00:06:07.443775 systemd[1]: cri-containerd-00172b9dd5396d046495f0783e1fb61d3d775c7d06034111304c14b63e48c4e4.scope: Deactivated successfully. Jul 7 00:06:07.534672 containerd[1456]: time="2025-07-07T00:06:07.534441965Z" level=info msg="shim disconnected" id=00172b9dd5396d046495f0783e1fb61d3d775c7d06034111304c14b63e48c4e4 namespace=k8s.io Jul 7 00:06:07.534672 containerd[1456]: time="2025-07-07T00:06:07.534510823Z" level=warning msg="cleaning up after shim disconnected" id=00172b9dd5396d046495f0783e1fb61d3d775c7d06034111304c14b63e48c4e4 namespace=k8s.io Jul 7 00:06:07.534672 containerd[1456]: time="2025-07-07T00:06:07.534525107Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 7 00:06:08.221526 kubelet[2515]: I0707 00:06:08.221428 2515 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 7 00:06:08.222056 kubelet[2515]: E0707 00:06:08.221704 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:08.225338 containerd[1456]: time="2025-07-07T00:06:08.224995508Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\"" Jul 7 00:06:08.234771 kubelet[2515]: I0707 00:06:08.234712 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-5b487879c5-hx4jl" podStartSLOduration=2.341511321 podStartE2EDuration="4.234703269s" podCreationTimestamp="2025-07-07 00:06:04 +0000 UTC" firstStartedPulling="2025-07-07 00:06:04.827097213 +0000 UTC m=+16.794967536" lastFinishedPulling="2025-07-07 00:06:06.720289171 +0000 UTC m=+18.688159484" observedRunningTime="2025-07-07 00:06:07.23045644 +0000 UTC m=+19.198326753" watchObservedRunningTime="2025-07-07 00:06:08.234703269 +0000 UTC m=+20.202573582" Jul 7 00:06:09.145357 kubelet[2515]: E0707 00:06:09.145274 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-9knjc" podUID="2001a2e7-8165-484d-b2ef-2ba8a539af9a" Jul 7 00:06:10.172654 containerd[1456]: time="2025-07-07T00:06:10.172595675Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:10.174015 containerd[1456]: time="2025-07-07T00:06:10.173969181Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.2: active requests=0, bytes read=70436221" Jul 7 00:06:10.176125 containerd[1456]: time="2025-07-07T00:06:10.174514372Z" level=info msg="ImageCreate event name:\"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:10.177417 containerd[1456]: time="2025-07-07T00:06:10.177229427Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:10.178239 containerd[1456]: time="2025-07-07T00:06:10.177825360Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.2\" with image id \"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d\", size \"71928924\" in 1.952764096s" Jul 7 00:06:10.178239 containerd[1456]: time="2025-07-07T00:06:10.177867780Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\" returns image reference \"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\"" Jul 7 00:06:10.181494 containerd[1456]: time="2025-07-07T00:06:10.181436925Z" level=info msg="CreateContainer within sandbox \"6261b6a8cda854340ae6bb4e54c2ab4e67e0c6461ca9da4594865c0bfe9bb534\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jul 7 00:06:10.206876 containerd[1456]: time="2025-07-07T00:06:10.206817859Z" level=info msg="CreateContainer within sandbox \"6261b6a8cda854340ae6bb4e54c2ab4e67e0c6461ca9da4594865c0bfe9bb534\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"7b59576cc9673b660894ced89b268fde31c03073836b39f4fb6a6ab860b75cca\"" Jul 7 00:06:10.210190 containerd[1456]: time="2025-07-07T00:06:10.208811764Z" level=info msg="StartContainer for \"7b59576cc9673b660894ced89b268fde31c03073836b39f4fb6a6ab860b75cca\"" Jul 7 00:06:10.209555 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2633693056.mount: Deactivated successfully. Jul 7 00:06:10.257514 systemd[1]: Started cri-containerd-7b59576cc9673b660894ced89b268fde31c03073836b39f4fb6a6ab860b75cca.scope - libcontainer container 7b59576cc9673b660894ced89b268fde31c03073836b39f4fb6a6ab860b75cca. Jul 7 00:06:10.299476 containerd[1456]: time="2025-07-07T00:06:10.299100230Z" level=info msg="StartContainer for \"7b59576cc9673b660894ced89b268fde31c03073836b39f4fb6a6ab860b75cca\" returns successfully" Jul 7 00:06:10.805324 containerd[1456]: time="2025-07-07T00:06:10.805253059Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jul 7 00:06:10.808646 systemd[1]: cri-containerd-7b59576cc9673b660894ced89b268fde31c03073836b39f4fb6a6ab860b75cca.scope: Deactivated successfully. Jul 7 00:06:10.830125 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7b59576cc9673b660894ced89b268fde31c03073836b39f4fb6a6ab860b75cca-rootfs.mount: Deactivated successfully. Jul 7 00:06:10.850861 kubelet[2515]: I0707 00:06:10.850731 2515 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Jul 7 00:06:10.881850 containerd[1456]: time="2025-07-07T00:06:10.881715095Z" level=info msg="shim disconnected" id=7b59576cc9673b660894ced89b268fde31c03073836b39f4fb6a6ab860b75cca namespace=k8s.io Jul 7 00:06:10.882041 containerd[1456]: time="2025-07-07T00:06:10.881906128Z" level=warning msg="cleaning up after shim disconnected" id=7b59576cc9673b660894ced89b268fde31c03073836b39f4fb6a6ab860b75cca namespace=k8s.io Jul 7 00:06:10.882041 containerd[1456]: time="2025-07-07T00:06:10.881919471Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 7 00:06:10.908878 systemd[1]: Created slice kubepods-burstable-pod6b2ab80d_76dc_4520_bb95_f08a370b3d13.slice - libcontainer container kubepods-burstable-pod6b2ab80d_76dc_4520_bb95_f08a370b3d13.slice. Jul 7 00:06:10.928858 systemd[1]: Created slice kubepods-besteffort-pod4741c1ac_aedc_48ec_8809_16303ada3c56.slice - libcontainer container kubepods-besteffort-pod4741c1ac_aedc_48ec_8809_16303ada3c56.slice. Jul 7 00:06:10.938479 systemd[1]: Created slice kubepods-burstable-pod178fc278_1010_4c58_8d67_4a6ee0c1491b.slice - libcontainer container kubepods-burstable-pod178fc278_1010_4c58_8d67_4a6ee0c1491b.slice. Jul 7 00:06:10.946923 systemd[1]: Created slice kubepods-besteffort-pod7f03664a_b5c1_41d9_b19a_954bfd2a5ea7.slice - libcontainer container kubepods-besteffort-pod7f03664a_b5c1_41d9_b19a_954bfd2a5ea7.slice. Jul 7 00:06:10.954506 systemd[1]: Created slice kubepods-besteffort-podcd6ca0df_602a_496f_a023_657b3d62c9da.slice - libcontainer container kubepods-besteffort-podcd6ca0df_602a_496f_a023_657b3d62c9da.slice. Jul 7 00:06:10.965168 systemd[1]: Created slice kubepods-besteffort-pod06260b4f_49a9_4902_b2df_d9734428c436.slice - libcontainer container kubepods-besteffort-pod06260b4f_49a9_4902_b2df_d9734428c436.slice. Jul 7 00:06:10.974588 systemd[1]: Created slice kubepods-besteffort-pod15e90dc0_720a_4ae9_aac2_01720a32d5a8.slice - libcontainer container kubepods-besteffort-pod15e90dc0_720a_4ae9_aac2_01720a32d5a8.slice. Jul 7 00:06:11.092329 kubelet[2515]: I0707 00:06:11.091180 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7f03664a-b5c1-41d9-b19a-954bfd2a5ea7-tigera-ca-bundle\") pod \"calico-kube-controllers-94599b4db-wbn2z\" (UID: \"7f03664a-b5c1-41d9-b19a-954bfd2a5ea7\") " pod="calico-system/calico-kube-controllers-94599b4db-wbn2z" Jul 7 00:06:11.092329 kubelet[2515]: I0707 00:06:11.091299 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw8hh\" (UniqueName: \"kubernetes.io/projected/15e90dc0-720a-4ae9-aac2-01720a32d5a8-kube-api-access-qw8hh\") pod \"calico-apiserver-85b7446c9f-78mbn\" (UID: \"15e90dc0-720a-4ae9-aac2-01720a32d5a8\") " pod="calico-apiserver/calico-apiserver-85b7446c9f-78mbn" Jul 7 00:06:11.092329 kubelet[2515]: I0707 00:06:11.091339 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06260b4f-49a9-4902-b2df-d9734428c436-config\") pod \"goldmane-58fd7646b9-d799z\" (UID: \"06260b4f-49a9-4902-b2df-d9734428c436\") " pod="calico-system/goldmane-58fd7646b9-d799z" Jul 7 00:06:11.092329 kubelet[2515]: I0707 00:06:11.091411 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/178fc278-1010-4c58-8d67-4a6ee0c1491b-config-volume\") pod \"coredns-7c65d6cfc9-x9bb2\" (UID: \"178fc278-1010-4c58-8d67-4a6ee0c1491b\") " pod="kube-system/coredns-7c65d6cfc9-x9bb2" Jul 7 00:06:11.092329 kubelet[2515]: I0707 00:06:11.091455 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06260b4f-49a9-4902-b2df-d9734428c436-goldmane-ca-bundle\") pod \"goldmane-58fd7646b9-d799z\" (UID: \"06260b4f-49a9-4902-b2df-d9734428c436\") " pod="calico-system/goldmane-58fd7646b9-d799z" Jul 7 00:06:11.092721 kubelet[2515]: I0707 00:06:11.091487 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x59qv\" (UniqueName: \"kubernetes.io/projected/7f03664a-b5c1-41d9-b19a-954bfd2a5ea7-kube-api-access-x59qv\") pod \"calico-kube-controllers-94599b4db-wbn2z\" (UID: \"7f03664a-b5c1-41d9-b19a-954bfd2a5ea7\") " pod="calico-system/calico-kube-controllers-94599b4db-wbn2z" Jul 7 00:06:11.092721 kubelet[2515]: I0707 00:06:11.091517 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/06260b4f-49a9-4902-b2df-d9734428c436-goldmane-key-pair\") pod \"goldmane-58fd7646b9-d799z\" (UID: \"06260b4f-49a9-4902-b2df-d9734428c436\") " pod="calico-system/goldmane-58fd7646b9-d799z" Jul 7 00:06:11.092721 kubelet[2515]: I0707 00:06:11.091550 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlzqc\" (UniqueName: \"kubernetes.io/projected/06260b4f-49a9-4902-b2df-d9734428c436-kube-api-access-xlzqc\") pod \"goldmane-58fd7646b9-d799z\" (UID: \"06260b4f-49a9-4902-b2df-d9734428c436\") " pod="calico-system/goldmane-58fd7646b9-d799z" Jul 7 00:06:11.092721 kubelet[2515]: I0707 00:06:11.091581 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd6ca0df-602a-496f-a023-657b3d62c9da-whisker-ca-bundle\") pod \"whisker-59c44fdcd5-rlcnr\" (UID: \"cd6ca0df-602a-496f-a023-657b3d62c9da\") " pod="calico-system/whisker-59c44fdcd5-rlcnr" Jul 7 00:06:11.092721 kubelet[2515]: I0707 00:06:11.091609 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnmd8\" (UniqueName: \"kubernetes.io/projected/cd6ca0df-602a-496f-a023-657b3d62c9da-kube-api-access-bnmd8\") pod \"whisker-59c44fdcd5-rlcnr\" (UID: \"cd6ca0df-602a-496f-a023-657b3d62c9da\") " pod="calico-system/whisker-59c44fdcd5-rlcnr" Jul 7 00:06:11.092926 kubelet[2515]: I0707 00:06:11.091706 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwzkd\" (UniqueName: \"kubernetes.io/projected/4741c1ac-aedc-48ec-8809-16303ada3c56-kube-api-access-vwzkd\") pod \"calico-apiserver-85b7446c9f-8xc2r\" (UID: \"4741c1ac-aedc-48ec-8809-16303ada3c56\") " pod="calico-apiserver/calico-apiserver-85b7446c9f-8xc2r" Jul 7 00:06:11.092926 kubelet[2515]: I0707 00:06:11.091758 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg2mg\" (UniqueName: \"kubernetes.io/projected/178fc278-1010-4c58-8d67-4a6ee0c1491b-kube-api-access-sg2mg\") pod \"coredns-7c65d6cfc9-x9bb2\" (UID: \"178fc278-1010-4c58-8d67-4a6ee0c1491b\") " pod="kube-system/coredns-7c65d6cfc9-x9bb2" Jul 7 00:06:11.092926 kubelet[2515]: I0707 00:06:11.091779 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4qzz\" (UniqueName: \"kubernetes.io/projected/6b2ab80d-76dc-4520-bb95-f08a370b3d13-kube-api-access-p4qzz\") pod \"coredns-7c65d6cfc9-pdhcs\" (UID: \"6b2ab80d-76dc-4520-bb95-f08a370b3d13\") " pod="kube-system/coredns-7c65d6cfc9-pdhcs" Jul 7 00:06:11.092926 kubelet[2515]: I0707 00:06:11.091803 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/4741c1ac-aedc-48ec-8809-16303ada3c56-calico-apiserver-certs\") pod \"calico-apiserver-85b7446c9f-8xc2r\" (UID: \"4741c1ac-aedc-48ec-8809-16303ada3c56\") " pod="calico-apiserver/calico-apiserver-85b7446c9f-8xc2r" Jul 7 00:06:11.092926 kubelet[2515]: I0707 00:06:11.091829 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/15e90dc0-720a-4ae9-aac2-01720a32d5a8-calico-apiserver-certs\") pod \"calico-apiserver-85b7446c9f-78mbn\" (UID: \"15e90dc0-720a-4ae9-aac2-01720a32d5a8\") " pod="calico-apiserver/calico-apiserver-85b7446c9f-78mbn" Jul 7 00:06:11.093125 kubelet[2515]: I0707 00:06:11.091852 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/cd6ca0df-602a-496f-a023-657b3d62c9da-whisker-backend-key-pair\") pod \"whisker-59c44fdcd5-rlcnr\" (UID: \"cd6ca0df-602a-496f-a023-657b3d62c9da\") " pod="calico-system/whisker-59c44fdcd5-rlcnr" Jul 7 00:06:11.093125 kubelet[2515]: I0707 00:06:11.091871 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b2ab80d-76dc-4520-bb95-f08a370b3d13-config-volume\") pod \"coredns-7c65d6cfc9-pdhcs\" (UID: \"6b2ab80d-76dc-4520-bb95-f08a370b3d13\") " pod="kube-system/coredns-7c65d6cfc9-pdhcs" Jul 7 00:06:11.153111 systemd[1]: Created slice kubepods-besteffort-pod2001a2e7_8165_484d_b2ef_2ba8a539af9a.slice - libcontainer container kubepods-besteffort-pod2001a2e7_8165_484d_b2ef_2ba8a539af9a.slice. Jul 7 00:06:11.156716 containerd[1456]: time="2025-07-07T00:06:11.156643661Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-9knjc,Uid:2001a2e7-8165-484d-b2ef-2ba8a539af9a,Namespace:calico-system,Attempt:0,}" Jul 7 00:06:11.239828 containerd[1456]: time="2025-07-07T00:06:11.239716721Z" level=error msg="Failed to destroy network for sandbox \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.240254 containerd[1456]: time="2025-07-07T00:06:11.240118405Z" level=error msg="encountered an error cleaning up failed sandbox \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.240254 containerd[1456]: time="2025-07-07T00:06:11.240170246Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-9knjc,Uid:2001a2e7-8165-484d-b2ef-2ba8a539af9a,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.247703 kubelet[2515]: E0707 00:06:11.241858 2515 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.247703 kubelet[2515]: E0707 00:06:11.241923 2515 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-9knjc" Jul 7 00:06:11.247703 kubelet[2515]: E0707 00:06:11.241980 2515 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-9knjc" Jul 7 00:06:11.244062 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7-shm.mount: Deactivated successfully. Jul 7 00:06:11.248221 kubelet[2515]: E0707 00:06:11.242018 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-9knjc_calico-system(2001a2e7-8165-484d-b2ef-2ba8a539af9a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-9knjc_calico-system(2001a2e7-8165-484d-b2ef-2ba8a539af9a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-9knjc" podUID="2001a2e7-8165-484d-b2ef-2ba8a539af9a" Jul 7 00:06:11.280538 containerd[1456]: time="2025-07-07T00:06:11.280491396Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85b7446c9f-78mbn,Uid:15e90dc0-720a-4ae9-aac2-01720a32d5a8,Namespace:calico-apiserver,Attempt:0,}" Jul 7 00:06:11.282214 containerd[1456]: time="2025-07-07T00:06:11.282012415Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\"" Jul 7 00:06:11.365925 containerd[1456]: time="2025-07-07T00:06:11.365009988Z" level=error msg="Failed to destroy network for sandbox \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.366075 containerd[1456]: time="2025-07-07T00:06:11.366025920Z" level=error msg="encountered an error cleaning up failed sandbox \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.366148 containerd[1456]: time="2025-07-07T00:06:11.366111648Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85b7446c9f-78mbn,Uid:15e90dc0-720a-4ae9-aac2-01720a32d5a8,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.366498 kubelet[2515]: E0707 00:06:11.366424 2515 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.366588 kubelet[2515]: E0707 00:06:11.366507 2515 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-85b7446c9f-78mbn" Jul 7 00:06:11.366588 kubelet[2515]: E0707 00:06:11.366532 2515 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-85b7446c9f-78mbn" Jul 7 00:06:11.366910 kubelet[2515]: E0707 00:06:11.366634 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-85b7446c9f-78mbn_calico-apiserver(15e90dc0-720a-4ae9-aac2-01720a32d5a8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-85b7446c9f-78mbn_calico-apiserver(15e90dc0-720a-4ae9-aac2-01720a32d5a8)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-85b7446c9f-78mbn" podUID="15e90dc0-720a-4ae9-aac2-01720a32d5a8" Jul 7 00:06:11.515888 kubelet[2515]: E0707 00:06:11.515807 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:11.516556 containerd[1456]: time="2025-07-07T00:06:11.516506047Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-pdhcs,Uid:6b2ab80d-76dc-4520-bb95-f08a370b3d13,Namespace:kube-system,Attempt:0,}" Jul 7 00:06:11.544753 kubelet[2515]: E0707 00:06:11.544709 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:11.545896 containerd[1456]: time="2025-07-07T00:06:11.545821820Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-x9bb2,Uid:178fc278-1010-4c58-8d67-4a6ee0c1491b,Namespace:kube-system,Attempt:0,}" Jul 7 00:06:11.547384 containerd[1456]: time="2025-07-07T00:06:11.546381326Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85b7446c9f-8xc2r,Uid:4741c1ac-aedc-48ec-8809-16303ada3c56,Namespace:calico-apiserver,Attempt:0,}" Jul 7 00:06:11.552516 containerd[1456]: time="2025-07-07T00:06:11.552483780Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-94599b4db-wbn2z,Uid:7f03664a-b5c1-41d9-b19a-954bfd2a5ea7,Namespace:calico-system,Attempt:0,}" Jul 7 00:06:11.561195 containerd[1456]: time="2025-07-07T00:06:11.561159152Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-59c44fdcd5-rlcnr,Uid:cd6ca0df-602a-496f-a023-657b3d62c9da,Namespace:calico-system,Attempt:0,}" Jul 7 00:06:11.574037 containerd[1456]: time="2025-07-07T00:06:11.573979340Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-d799z,Uid:06260b4f-49a9-4902-b2df-d9734428c436,Namespace:calico-system,Attempt:0,}" Jul 7 00:06:11.649755 containerd[1456]: time="2025-07-07T00:06:11.649541861Z" level=error msg="Failed to destroy network for sandbox \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.650743 containerd[1456]: time="2025-07-07T00:06:11.650560513Z" level=error msg="encountered an error cleaning up failed sandbox \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.651104 containerd[1456]: time="2025-07-07T00:06:11.651072660Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-pdhcs,Uid:6b2ab80d-76dc-4520-bb95-f08a370b3d13,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.651661 kubelet[2515]: E0707 00:06:11.651614 2515 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.651707 kubelet[2515]: E0707 00:06:11.651676 2515 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-pdhcs" Jul 7 00:06:11.651707 kubelet[2515]: E0707 00:06:11.651700 2515 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-pdhcs" Jul 7 00:06:11.651897 kubelet[2515]: E0707 00:06:11.651743 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-pdhcs_kube-system(6b2ab80d-76dc-4520-bb95-f08a370b3d13)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-pdhcs_kube-system(6b2ab80d-76dc-4520-bb95-f08a370b3d13)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-pdhcs" podUID="6b2ab80d-76dc-4520-bb95-f08a370b3d13" Jul 7 00:06:11.657328 containerd[1456]: time="2025-07-07T00:06:11.657141898Z" level=error msg="Failed to destroy network for sandbox \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.657919 containerd[1456]: time="2025-07-07T00:06:11.657870180Z" level=error msg="encountered an error cleaning up failed sandbox \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.657967 containerd[1456]: time="2025-07-07T00:06:11.657914580Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-94599b4db-wbn2z,Uid:7f03664a-b5c1-41d9-b19a-954bfd2a5ea7,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.658414 kubelet[2515]: E0707 00:06:11.658337 2515 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.658414 kubelet[2515]: E0707 00:06:11.658389 2515 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-94599b4db-wbn2z" Jul 7 00:06:11.658414 kubelet[2515]: E0707 00:06:11.658406 2515 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-94599b4db-wbn2z" Jul 7 00:06:11.658511 kubelet[2515]: E0707 00:06:11.658432 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-94599b4db-wbn2z_calico-system(7f03664a-b5c1-41d9-b19a-954bfd2a5ea7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-94599b4db-wbn2z_calico-system(7f03664a-b5c1-41d9-b19a-954bfd2a5ea7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-94599b4db-wbn2z" podUID="7f03664a-b5c1-41d9-b19a-954bfd2a5ea7" Jul 7 00:06:11.724888 containerd[1456]: time="2025-07-07T00:06:11.724314707Z" level=error msg="Failed to destroy network for sandbox \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.725504 containerd[1456]: time="2025-07-07T00:06:11.725467347Z" level=error msg="encountered an error cleaning up failed sandbox \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.725570 containerd[1456]: time="2025-07-07T00:06:11.725520949Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85b7446c9f-8xc2r,Uid:4741c1ac-aedc-48ec-8809-16303ada3c56,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.725767 kubelet[2515]: E0707 00:06:11.725715 2515 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.725815 kubelet[2515]: E0707 00:06:11.725773 2515 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-85b7446c9f-8xc2r" Jul 7 00:06:11.725815 kubelet[2515]: E0707 00:06:11.725794 2515 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-85b7446c9f-8xc2r" Jul 7 00:06:11.725878 kubelet[2515]: E0707 00:06:11.725834 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-85b7446c9f-8xc2r_calico-apiserver(4741c1ac-aedc-48ec-8809-16303ada3c56)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-85b7446c9f-8xc2r_calico-apiserver(4741c1ac-aedc-48ec-8809-16303ada3c56)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-85b7446c9f-8xc2r" podUID="4741c1ac-aedc-48ec-8809-16303ada3c56" Jul 7 00:06:11.735850 containerd[1456]: time="2025-07-07T00:06:11.735814349Z" level=error msg="Failed to destroy network for sandbox \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.737538 containerd[1456]: time="2025-07-07T00:06:11.737416203Z" level=error msg="encountered an error cleaning up failed sandbox \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.737778 containerd[1456]: time="2025-07-07T00:06:11.737654023Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-x9bb2,Uid:178fc278-1010-4c58-8d67-4a6ee0c1491b,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.738083 kubelet[2515]: E0707 00:06:11.737984 2515 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.738083 kubelet[2515]: E0707 00:06:11.738025 2515 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-x9bb2" Jul 7 00:06:11.738083 kubelet[2515]: E0707 00:06:11.738041 2515 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-x9bb2" Jul 7 00:06:11.738180 kubelet[2515]: E0707 00:06:11.738067 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-x9bb2_kube-system(178fc278-1010-4c58-8d67-4a6ee0c1491b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-x9bb2_kube-system(178fc278-1010-4c58-8d67-4a6ee0c1491b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-x9bb2" podUID="178fc278-1010-4c58-8d67-4a6ee0c1491b" Jul 7 00:06:11.752238 containerd[1456]: time="2025-07-07T00:06:11.751841096Z" level=error msg="Failed to destroy network for sandbox \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.752238 containerd[1456]: time="2025-07-07T00:06:11.752130736Z" level=error msg="encountered an error cleaning up failed sandbox \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.752238 containerd[1456]: time="2025-07-07T00:06:11.752167484Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-d799z,Uid:06260b4f-49a9-4902-b2df-d9734428c436,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.752484 kubelet[2515]: E0707 00:06:11.752354 2515 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.752484 kubelet[2515]: E0707 00:06:11.752436 2515 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-58fd7646b9-d799z" Jul 7 00:06:11.752484 kubelet[2515]: E0707 00:06:11.752476 2515 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-58fd7646b9-d799z" Jul 7 00:06:11.752567 kubelet[2515]: E0707 00:06:11.752511 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-58fd7646b9-d799z_calico-system(06260b4f-49a9-4902-b2df-d9734428c436)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-58fd7646b9-d799z_calico-system(06260b4f-49a9-4902-b2df-d9734428c436)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-58fd7646b9-d799z" podUID="06260b4f-49a9-4902-b2df-d9734428c436" Jul 7 00:06:11.752947 containerd[1456]: time="2025-07-07T00:06:11.752914079Z" level=error msg="Failed to destroy network for sandbox \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.753429 containerd[1456]: time="2025-07-07T00:06:11.753284587Z" level=error msg="encountered an error cleaning up failed sandbox \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.753429 containerd[1456]: time="2025-07-07T00:06:11.753318674Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-59c44fdcd5-rlcnr,Uid:cd6ca0df-602a-496f-a023-657b3d62c9da,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.753580 kubelet[2515]: E0707 00:06:11.753558 2515 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:11.753630 kubelet[2515]: E0707 00:06:11.753586 2515 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-59c44fdcd5-rlcnr" Jul 7 00:06:11.753630 kubelet[2515]: E0707 00:06:11.753603 2515 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-59c44fdcd5-rlcnr" Jul 7 00:06:11.753694 kubelet[2515]: E0707 00:06:11.753672 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-59c44fdcd5-rlcnr_calico-system(cd6ca0df-602a-496f-a023-657b3d62c9da)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-59c44fdcd5-rlcnr_calico-system(cd6ca0df-602a-496f-a023-657b3d62c9da)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-59c44fdcd5-rlcnr" podUID="cd6ca0df-602a-496f-a023-657b3d62c9da" Jul 7 00:06:12.288653 kubelet[2515]: I0707 00:06:12.288619 2515 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Jul 7 00:06:12.291096 containerd[1456]: time="2025-07-07T00:06:12.290668989Z" level=info msg="StopPodSandbox for \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\"" Jul 7 00:06:12.291096 containerd[1456]: time="2025-07-07T00:06:12.290837422Z" level=info msg="Ensure that sandbox 3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0 in task-service has been cleanup successfully" Jul 7 00:06:12.291096 containerd[1456]: time="2025-07-07T00:06:12.294059013Z" level=info msg="StopPodSandbox for \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\"" Jul 7 00:06:12.291096 containerd[1456]: time="2025-07-07T00:06:12.294167664Z" level=info msg="Ensure that sandbox 455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7 in task-service has been cleanup successfully" Jul 7 00:06:12.297003 kubelet[2515]: I0707 00:06:12.293656 2515 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Jul 7 00:06:12.297003 kubelet[2515]: I0707 00:06:12.296720 2515 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Jul 7 00:06:12.300306 containerd[1456]: time="2025-07-07T00:06:12.297108079Z" level=info msg="StopPodSandbox for \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\"" Jul 7 00:06:12.300306 containerd[1456]: time="2025-07-07T00:06:12.297276673Z" level=info msg="Ensure that sandbox 09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8 in task-service has been cleanup successfully" Jul 7 00:06:12.310783 kubelet[2515]: I0707 00:06:12.309546 2515 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Jul 7 00:06:12.311886 containerd[1456]: time="2025-07-07T00:06:12.311646766Z" level=info msg="StopPodSandbox for \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\"" Jul 7 00:06:12.311886 containerd[1456]: time="2025-07-07T00:06:12.311790544Z" level=info msg="Ensure that sandbox 71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980 in task-service has been cleanup successfully" Jul 7 00:06:12.320719 kubelet[2515]: I0707 00:06:12.320129 2515 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Jul 7 00:06:12.322846 containerd[1456]: time="2025-07-07T00:06:12.322787746Z" level=info msg="StopPodSandbox for \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\"" Jul 7 00:06:12.324442 kubelet[2515]: I0707 00:06:12.324424 2515 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Jul 7 00:06:12.325463 containerd[1456]: time="2025-07-07T00:06:12.325428004Z" level=info msg="StopPodSandbox for \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\"" Jul 7 00:06:12.325778 containerd[1456]: time="2025-07-07T00:06:12.325761419Z" level=info msg="Ensure that sandbox 3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f in task-service has been cleanup successfully" Jul 7 00:06:12.326683 containerd[1456]: time="2025-07-07T00:06:12.325447147Z" level=info msg="Ensure that sandbox 4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1 in task-service has been cleanup successfully" Jul 7 00:06:12.328709 kubelet[2515]: I0707 00:06:12.328670 2515 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Jul 7 00:06:12.330735 containerd[1456]: time="2025-07-07T00:06:12.330716478Z" level=info msg="StopPodSandbox for \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\"" Jul 7 00:06:12.331557 containerd[1456]: time="2025-07-07T00:06:12.331520346Z" level=info msg="Ensure that sandbox b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd in task-service has been cleanup successfully" Jul 7 00:06:12.339897 kubelet[2515]: I0707 00:06:12.339863 2515 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Jul 7 00:06:12.340694 containerd[1456]: time="2025-07-07T00:06:12.340649373Z" level=info msg="StopPodSandbox for \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\"" Jul 7 00:06:12.342089 containerd[1456]: time="2025-07-07T00:06:12.341837706Z" level=info msg="Ensure that sandbox 071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93 in task-service has been cleanup successfully" Jul 7 00:06:12.375388 containerd[1456]: time="2025-07-07T00:06:12.374772693Z" level=error msg="StopPodSandbox for \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\" failed" error="failed to destroy network for sandbox \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:12.375463 kubelet[2515]: E0707 00:06:12.374939 2515 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Jul 7 00:06:12.375463 kubelet[2515]: E0707 00:06:12.374978 2515 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8"} Jul 7 00:06:12.375463 kubelet[2515]: E0707 00:06:12.375069 2515 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"15e90dc0-720a-4ae9-aac2-01720a32d5a8\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jul 7 00:06:12.375463 kubelet[2515]: E0707 00:06:12.375093 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"15e90dc0-720a-4ae9-aac2-01720a32d5a8\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-85b7446c9f-78mbn" podUID="15e90dc0-720a-4ae9-aac2-01720a32d5a8" Jul 7 00:06:12.405522 containerd[1456]: time="2025-07-07T00:06:12.405481155Z" level=error msg="StopPodSandbox for \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\" failed" error="failed to destroy network for sandbox \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:12.405889 kubelet[2515]: E0707 00:06:12.405820 2515 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Jul 7 00:06:12.405889 kubelet[2515]: E0707 00:06:12.405882 2515 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0"} Jul 7 00:06:12.405993 kubelet[2515]: E0707 00:06:12.405913 2515 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"4741c1ac-aedc-48ec-8809-16303ada3c56\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jul 7 00:06:12.405993 kubelet[2515]: E0707 00:06:12.405935 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"4741c1ac-aedc-48ec-8809-16303ada3c56\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-85b7446c9f-8xc2r" podUID="4741c1ac-aedc-48ec-8809-16303ada3c56" Jul 7 00:06:12.416810 containerd[1456]: time="2025-07-07T00:06:12.416762363Z" level=error msg="StopPodSandbox for \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\" failed" error="failed to destroy network for sandbox \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:12.416984 kubelet[2515]: E0707 00:06:12.416947 2515 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Jul 7 00:06:12.417023 kubelet[2515]: E0707 00:06:12.416985 2515 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93"} Jul 7 00:06:12.417110 kubelet[2515]: E0707 00:06:12.417016 2515 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"178fc278-1010-4c58-8d67-4a6ee0c1491b\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jul 7 00:06:12.417110 kubelet[2515]: E0707 00:06:12.417041 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"178fc278-1010-4c58-8d67-4a6ee0c1491b\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-x9bb2" podUID="178fc278-1010-4c58-8d67-4a6ee0c1491b" Jul 7 00:06:12.422144 containerd[1456]: time="2025-07-07T00:06:12.422105780Z" level=error msg="StopPodSandbox for \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\" failed" error="failed to destroy network for sandbox \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:12.422388 kubelet[2515]: E0707 00:06:12.422337 2515 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Jul 7 00:06:12.422971 kubelet[2515]: E0707 00:06:12.422933 2515 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7"} Jul 7 00:06:12.422971 kubelet[2515]: E0707 00:06:12.422966 2515 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"2001a2e7-8165-484d-b2ef-2ba8a539af9a\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jul 7 00:06:12.423130 kubelet[2515]: E0707 00:06:12.422985 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"2001a2e7-8165-484d-b2ef-2ba8a539af9a\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-9knjc" podUID="2001a2e7-8165-484d-b2ef-2ba8a539af9a" Jul 7 00:06:12.425427 containerd[1456]: time="2025-07-07T00:06:12.425390783Z" level=error msg="StopPodSandbox for \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\" failed" error="failed to destroy network for sandbox \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:12.425583 kubelet[2515]: E0707 00:06:12.425518 2515 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Jul 7 00:06:12.425583 kubelet[2515]: E0707 00:06:12.425543 2515 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd"} Jul 7 00:06:12.425583 kubelet[2515]: E0707 00:06:12.425563 2515 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"6b2ab80d-76dc-4520-bb95-f08a370b3d13\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jul 7 00:06:12.425717 kubelet[2515]: E0707 00:06:12.425581 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"6b2ab80d-76dc-4520-bb95-f08a370b3d13\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-pdhcs" podUID="6b2ab80d-76dc-4520-bb95-f08a370b3d13" Jul 7 00:06:12.436822 containerd[1456]: time="2025-07-07T00:06:12.436740344Z" level=error msg="StopPodSandbox for \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\" failed" error="failed to destroy network for sandbox \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:12.436948 kubelet[2515]: E0707 00:06:12.436900 2515 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Jul 7 00:06:12.436948 kubelet[2515]: E0707 00:06:12.436933 2515 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980"} Jul 7 00:06:12.437028 kubelet[2515]: E0707 00:06:12.436959 2515 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"cd6ca0df-602a-496f-a023-657b3d62c9da\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jul 7 00:06:12.437028 kubelet[2515]: E0707 00:06:12.436977 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"cd6ca0df-602a-496f-a023-657b3d62c9da\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-59c44fdcd5-rlcnr" podUID="cd6ca0df-602a-496f-a023-657b3d62c9da" Jul 7 00:06:12.441519 containerd[1456]: time="2025-07-07T00:06:12.441467090Z" level=error msg="StopPodSandbox for \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\" failed" error="failed to destroy network for sandbox \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:12.441637 kubelet[2515]: E0707 00:06:12.441590 2515 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Jul 7 00:06:12.441669 kubelet[2515]: E0707 00:06:12.441639 2515 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f"} Jul 7 00:06:12.441669 kubelet[2515]: E0707 00:06:12.441661 2515 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"06260b4f-49a9-4902-b2df-d9734428c436\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jul 7 00:06:12.441747 kubelet[2515]: E0707 00:06:12.441681 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"06260b4f-49a9-4902-b2df-d9734428c436\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-58fd7646b9-d799z" podUID="06260b4f-49a9-4902-b2df-d9734428c436" Jul 7 00:06:12.443492 containerd[1456]: time="2025-07-07T00:06:12.443467741Z" level=error msg="StopPodSandbox for \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\" failed" error="failed to destroy network for sandbox \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 7 00:06:12.443688 kubelet[2515]: E0707 00:06:12.443657 2515 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Jul 7 00:06:12.443688 kubelet[2515]: E0707 00:06:12.443687 2515 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1"} Jul 7 00:06:12.443783 kubelet[2515]: E0707 00:06:12.443706 2515 kuberuntime_manager.go:1079] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"7f03664a-b5c1-41d9-b19a-954bfd2a5ea7\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jul 7 00:06:12.443783 kubelet[2515]: E0707 00:06:12.443723 2515 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"7f03664a-b5c1-41d9-b19a-954bfd2a5ea7\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-94599b4db-wbn2z" podUID="7f03664a-b5c1-41d9-b19a-954bfd2a5ea7" Jul 7 00:06:14.768447 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount974419815.mount: Deactivated successfully. Jul 7 00:06:14.799476 containerd[1456]: time="2025-07-07T00:06:14.799421871Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:14.800204 containerd[1456]: time="2025-07-07T00:06:14.800168210Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.2: active requests=0, bytes read=158500163" Jul 7 00:06:14.800850 containerd[1456]: time="2025-07-07T00:06:14.800802628Z" level=info msg="ImageCreate event name:\"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:14.803116 containerd[1456]: time="2025-07-07T00:06:14.802503280Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:14.803754 containerd[1456]: time="2025-07-07T00:06:14.803594909Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.2\" with image id \"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760\", size \"158500025\" in 3.519495778s" Jul 7 00:06:14.803754 containerd[1456]: time="2025-07-07T00:06:14.803626014Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\" returns image reference \"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\"" Jul 7 00:06:14.827227 containerd[1456]: time="2025-07-07T00:06:14.827180655Z" level=info msg="CreateContainer within sandbox \"6261b6a8cda854340ae6bb4e54c2ab4e67e0c6461ca9da4594865c0bfe9bb534\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Jul 7 00:06:14.842709 containerd[1456]: time="2025-07-07T00:06:14.842656277Z" level=info msg="CreateContainer within sandbox \"6261b6a8cda854340ae6bb4e54c2ab4e67e0c6461ca9da4594865c0bfe9bb534\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"4cf0bc139daeb67691822e7adc2074362d46fec62277a2028a3b713ddffc59b6\"" Jul 7 00:06:14.843799 containerd[1456]: time="2025-07-07T00:06:14.843767258Z" level=info msg="StartContainer for \"4cf0bc139daeb67691822e7adc2074362d46fec62277a2028a3b713ddffc59b6\"" Jul 7 00:06:14.878496 systemd[1]: Started cri-containerd-4cf0bc139daeb67691822e7adc2074362d46fec62277a2028a3b713ddffc59b6.scope - libcontainer container 4cf0bc139daeb67691822e7adc2074362d46fec62277a2028a3b713ddffc59b6. Jul 7 00:06:14.908630 containerd[1456]: time="2025-07-07T00:06:14.908551181Z" level=info msg="StartContainer for \"4cf0bc139daeb67691822e7adc2074362d46fec62277a2028a3b713ddffc59b6\" returns successfully" Jul 7 00:06:14.992817 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Jul 7 00:06:14.992885 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Jul 7 00:06:15.071608 containerd[1456]: time="2025-07-07T00:06:15.070373310Z" level=info msg="StopPodSandbox for \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\"" Jul 7 00:06:15.183138 containerd[1456]: 2025-07-07 00:06:15.141 [INFO][3727] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Jul 7 00:06:15.183138 containerd[1456]: 2025-07-07 00:06:15.142 [INFO][3727] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" iface="eth0" netns="/var/run/netns/cni-60712b42-f92d-7e15-d084-2fa13b88b78b" Jul 7 00:06:15.183138 containerd[1456]: 2025-07-07 00:06:15.143 [INFO][3727] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" iface="eth0" netns="/var/run/netns/cni-60712b42-f92d-7e15-d084-2fa13b88b78b" Jul 7 00:06:15.183138 containerd[1456]: 2025-07-07 00:06:15.145 [INFO][3727] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" iface="eth0" netns="/var/run/netns/cni-60712b42-f92d-7e15-d084-2fa13b88b78b" Jul 7 00:06:15.183138 containerd[1456]: 2025-07-07 00:06:15.145 [INFO][3727] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Jul 7 00:06:15.183138 containerd[1456]: 2025-07-07 00:06:15.145 [INFO][3727] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Jul 7 00:06:15.183138 containerd[1456]: 2025-07-07 00:06:15.168 [INFO][3740] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" HandleID="k8s-pod-network.71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Workload="172--234--31--125-k8s-whisker--59c44fdcd5--rlcnr-eth0" Jul 7 00:06:15.183138 containerd[1456]: 2025-07-07 00:06:15.168 [INFO][3740] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:15.183138 containerd[1456]: 2025-07-07 00:06:15.168 [INFO][3740] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:15.183138 containerd[1456]: 2025-07-07 00:06:15.175 [WARNING][3740] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" HandleID="k8s-pod-network.71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Workload="172--234--31--125-k8s-whisker--59c44fdcd5--rlcnr-eth0" Jul 7 00:06:15.183138 containerd[1456]: 2025-07-07 00:06:15.175 [INFO][3740] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" HandleID="k8s-pod-network.71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Workload="172--234--31--125-k8s-whisker--59c44fdcd5--rlcnr-eth0" Jul 7 00:06:15.183138 containerd[1456]: 2025-07-07 00:06:15.177 [INFO][3740] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:15.183138 containerd[1456]: 2025-07-07 00:06:15.180 [INFO][3727] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Jul 7 00:06:15.184221 containerd[1456]: time="2025-07-07T00:06:15.183334282Z" level=info msg="TearDown network for sandbox \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\" successfully" Jul 7 00:06:15.184221 containerd[1456]: time="2025-07-07T00:06:15.183390702Z" level=info msg="StopPodSandbox for \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\" returns successfully" Jul 7 00:06:15.229685 kubelet[2515]: I0707 00:06:15.228757 2515 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd6ca0df-602a-496f-a023-657b3d62c9da-whisker-ca-bundle\") pod \"cd6ca0df-602a-496f-a023-657b3d62c9da\" (UID: \"cd6ca0df-602a-496f-a023-657b3d62c9da\") " Jul 7 00:06:15.229685 kubelet[2515]: I0707 00:06:15.228791 2515 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/cd6ca0df-602a-496f-a023-657b3d62c9da-whisker-backend-key-pair\") pod \"cd6ca0df-602a-496f-a023-657b3d62c9da\" (UID: \"cd6ca0df-602a-496f-a023-657b3d62c9da\") " Jul 7 00:06:15.229685 kubelet[2515]: I0707 00:06:15.228829 2515 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnmd8\" (UniqueName: \"kubernetes.io/projected/cd6ca0df-602a-496f-a023-657b3d62c9da-kube-api-access-bnmd8\") pod \"cd6ca0df-602a-496f-a023-657b3d62c9da\" (UID: \"cd6ca0df-602a-496f-a023-657b3d62c9da\") " Jul 7 00:06:15.231321 kubelet[2515]: I0707 00:06:15.230579 2515 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd6ca0df-602a-496f-a023-657b3d62c9da-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "cd6ca0df-602a-496f-a023-657b3d62c9da" (UID: "cd6ca0df-602a-496f-a023-657b3d62c9da"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jul 7 00:06:15.233844 kubelet[2515]: I0707 00:06:15.233778 2515 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd6ca0df-602a-496f-a023-657b3d62c9da-kube-api-access-bnmd8" (OuterVolumeSpecName: "kube-api-access-bnmd8") pod "cd6ca0df-602a-496f-a023-657b3d62c9da" (UID: "cd6ca0df-602a-496f-a023-657b3d62c9da"). InnerVolumeSpecName "kube-api-access-bnmd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jul 7 00:06:15.236969 kubelet[2515]: I0707 00:06:15.236593 2515 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd6ca0df-602a-496f-a023-657b3d62c9da-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "cd6ca0df-602a-496f-a023-657b3d62c9da" (UID: "cd6ca0df-602a-496f-a023-657b3d62c9da"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGidValue "" Jul 7 00:06:15.330006 kubelet[2515]: I0707 00:06:15.329478 2515 reconciler_common.go:293] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cd6ca0df-602a-496f-a023-657b3d62c9da-whisker-ca-bundle\") on node \"172-234-31-125\" DevicePath \"\"" Jul 7 00:06:15.330006 kubelet[2515]: I0707 00:06:15.329526 2515 reconciler_common.go:293] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/cd6ca0df-602a-496f-a023-657b3d62c9da-whisker-backend-key-pair\") on node \"172-234-31-125\" DevicePath \"\"" Jul 7 00:06:15.330006 kubelet[2515]: I0707 00:06:15.329552 2515 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnmd8\" (UniqueName: \"kubernetes.io/projected/cd6ca0df-602a-496f-a023-657b3d62c9da-kube-api-access-bnmd8\") on node \"172-234-31-125\" DevicePath \"\"" Jul 7 00:06:15.355349 systemd[1]: Removed slice kubepods-besteffort-podcd6ca0df_602a_496f_a023_657b3d62c9da.slice - libcontainer container kubepods-besteffort-podcd6ca0df_602a_496f_a023_657b3d62c9da.slice. Jul 7 00:06:15.365188 kubelet[2515]: I0707 00:06:15.364626 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-cvl5d" podStartSLOduration=1.793311814 podStartE2EDuration="11.364611798s" podCreationTimestamp="2025-07-07 00:06:04 +0000 UTC" firstStartedPulling="2025-07-07 00:06:05.233445862 +0000 UTC m=+17.201316175" lastFinishedPulling="2025-07-07 00:06:14.804745846 +0000 UTC m=+26.772616159" observedRunningTime="2025-07-07 00:06:15.362918445 +0000 UTC m=+27.330788758" watchObservedRunningTime="2025-07-07 00:06:15.364611798 +0000 UTC m=+27.332482111" Jul 7 00:06:15.416154 systemd[1]: Created slice kubepods-besteffort-pod3c73d839_0227_44af_9a41_a794da30c843.slice - libcontainer container kubepods-besteffort-pod3c73d839_0227_44af_9a41_a794da30c843.slice. Jul 7 00:06:15.429714 kubelet[2515]: I0707 00:06:15.429677 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5rcg\" (UniqueName: \"kubernetes.io/projected/3c73d839-0227-44af-9a41-a794da30c843-kube-api-access-w5rcg\") pod \"whisker-94bd567f5-f7gkf\" (UID: \"3c73d839-0227-44af-9a41-a794da30c843\") " pod="calico-system/whisker-94bd567f5-f7gkf" Jul 7 00:06:15.429768 kubelet[2515]: I0707 00:06:15.429752 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/3c73d839-0227-44af-9a41-a794da30c843-whisker-backend-key-pair\") pod \"whisker-94bd567f5-f7gkf\" (UID: \"3c73d839-0227-44af-9a41-a794da30c843\") " pod="calico-system/whisker-94bd567f5-f7gkf" Jul 7 00:06:15.429768 kubelet[2515]: I0707 00:06:15.429769 2515 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3c73d839-0227-44af-9a41-a794da30c843-whisker-ca-bundle\") pod \"whisker-94bd567f5-f7gkf\" (UID: \"3c73d839-0227-44af-9a41-a794da30c843\") " pod="calico-system/whisker-94bd567f5-f7gkf" Jul 7 00:06:15.720483 containerd[1456]: time="2025-07-07T00:06:15.720406919Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-94bd567f5-f7gkf,Uid:3c73d839-0227-44af-9a41-a794da30c843,Namespace:calico-system,Attempt:0,}" Jul 7 00:06:15.775409 systemd[1]: run-netns-cni\x2d60712b42\x2df92d\x2d7e15\x2dd084\x2d2fa13b88b78b.mount: Deactivated successfully. Jul 7 00:06:15.775609 systemd[1]: var-lib-kubelet-pods-cd6ca0df\x2d602a\x2d496f\x2da023\x2d657b3d62c9da-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dbnmd8.mount: Deactivated successfully. Jul 7 00:06:15.775888 systemd[1]: var-lib-kubelet-pods-cd6ca0df\x2d602a\x2d496f\x2da023\x2d657b3d62c9da-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Jul 7 00:06:15.839612 systemd-networkd[1382]: calib899fbd8f5b: Link UP Jul 7 00:06:15.841058 systemd-networkd[1382]: calib899fbd8f5b: Gained carrier Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.748 [INFO][3764] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.759 [INFO][3764] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--234--31--125-k8s-whisker--94bd567f5--f7gkf-eth0 whisker-94bd567f5- calico-system 3c73d839-0227-44af-9a41-a794da30c843 912 0 2025-07-07 00:06:15 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:94bd567f5 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s 172-234-31-125 whisker-94bd567f5-f7gkf eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calib899fbd8f5b [] [] }} ContainerID="8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" Namespace="calico-system" Pod="whisker-94bd567f5-f7gkf" WorkloadEndpoint="172--234--31--125-k8s-whisker--94bd567f5--f7gkf-" Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.759 [INFO][3764] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" Namespace="calico-system" Pod="whisker-94bd567f5-f7gkf" WorkloadEndpoint="172--234--31--125-k8s-whisker--94bd567f5--f7gkf-eth0" Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.797 [INFO][3776] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" HandleID="k8s-pod-network.8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" Workload="172--234--31--125-k8s-whisker--94bd567f5--f7gkf-eth0" Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.797 [INFO][3776] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" HandleID="k8s-pod-network.8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" Workload="172--234--31--125-k8s-whisker--94bd567f5--f7gkf-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f930), Attrs:map[string]string{"namespace":"calico-system", "node":"172-234-31-125", "pod":"whisker-94bd567f5-f7gkf", "timestamp":"2025-07-07 00:06:15.797802648 +0000 UTC"}, Hostname:"172-234-31-125", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.798 [INFO][3776] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.798 [INFO][3776] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.798 [INFO][3776] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-234-31-125' Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.804 [INFO][3776] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" host="172-234-31-125" Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.808 [INFO][3776] ipam/ipam.go 394: Looking up existing affinities for host host="172-234-31-125" Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.811 [INFO][3776] ipam/ipam.go 511: Trying affinity for 192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.812 [INFO][3776] ipam/ipam.go 158: Attempting to load block cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.814 [INFO][3776] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.814 [INFO][3776] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.38.192/26 handle="k8s-pod-network.8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" host="172-234-31-125" Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.815 [INFO][3776] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.818 [INFO][3776] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.38.192/26 handle="k8s-pod-network.8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" host="172-234-31-125" Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.824 [INFO][3776] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.38.193/26] block=192.168.38.192/26 handle="k8s-pod-network.8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" host="172-234-31-125" Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.824 [INFO][3776] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.38.193/26] handle="k8s-pod-network.8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" host="172-234-31-125" Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.824 [INFO][3776] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:15.857130 containerd[1456]: 2025-07-07 00:06:15.824 [INFO][3776] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.38.193/26] IPv6=[] ContainerID="8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" HandleID="k8s-pod-network.8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" Workload="172--234--31--125-k8s-whisker--94bd567f5--f7gkf-eth0" Jul 7 00:06:15.857964 containerd[1456]: 2025-07-07 00:06:15.827 [INFO][3764] cni-plugin/k8s.go 418: Populated endpoint ContainerID="8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" Namespace="calico-system" Pod="whisker-94bd567f5-f7gkf" WorkloadEndpoint="172--234--31--125-k8s-whisker--94bd567f5--f7gkf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-whisker--94bd567f5--f7gkf-eth0", GenerateName:"whisker-94bd567f5-", Namespace:"calico-system", SelfLink:"", UID:"3c73d839-0227-44af-9a41-a794da30c843", ResourceVersion:"912", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 15, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"94bd567f5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"", Pod:"whisker-94bd567f5-f7gkf", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.38.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calib899fbd8f5b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:15.857964 containerd[1456]: 2025-07-07 00:06:15.827 [INFO][3764] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.38.193/32] ContainerID="8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" Namespace="calico-system" Pod="whisker-94bd567f5-f7gkf" WorkloadEndpoint="172--234--31--125-k8s-whisker--94bd567f5--f7gkf-eth0" Jul 7 00:06:15.857964 containerd[1456]: 2025-07-07 00:06:15.827 [INFO][3764] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calib899fbd8f5b ContainerID="8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" Namespace="calico-system" Pod="whisker-94bd567f5-f7gkf" WorkloadEndpoint="172--234--31--125-k8s-whisker--94bd567f5--f7gkf-eth0" Jul 7 00:06:15.857964 containerd[1456]: 2025-07-07 00:06:15.838 [INFO][3764] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" Namespace="calico-system" Pod="whisker-94bd567f5-f7gkf" WorkloadEndpoint="172--234--31--125-k8s-whisker--94bd567f5--f7gkf-eth0" Jul 7 00:06:15.857964 containerd[1456]: 2025-07-07 00:06:15.839 [INFO][3764] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" Namespace="calico-system" Pod="whisker-94bd567f5-f7gkf" WorkloadEndpoint="172--234--31--125-k8s-whisker--94bd567f5--f7gkf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-whisker--94bd567f5--f7gkf-eth0", GenerateName:"whisker-94bd567f5-", Namespace:"calico-system", SelfLink:"", UID:"3c73d839-0227-44af-9a41-a794da30c843", ResourceVersion:"912", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 15, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"94bd567f5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be", Pod:"whisker-94bd567f5-f7gkf", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.38.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calib899fbd8f5b", MAC:"5a:28:39:bc:24:a3", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:15.857964 containerd[1456]: 2025-07-07 00:06:15.853 [INFO][3764] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be" Namespace="calico-system" Pod="whisker-94bd567f5-f7gkf" WorkloadEndpoint="172--234--31--125-k8s-whisker--94bd567f5--f7gkf-eth0" Jul 7 00:06:15.875256 containerd[1456]: time="2025-07-07T00:06:15.875155578Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:06:15.875256 containerd[1456]: time="2025-07-07T00:06:15.875203186Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:06:15.875448 containerd[1456]: time="2025-07-07T00:06:15.875214218Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:15.876056 containerd[1456]: time="2025-07-07T00:06:15.875680933Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:15.898530 systemd[1]: Started cri-containerd-8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be.scope - libcontainer container 8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be. Jul 7 00:06:15.950769 containerd[1456]: time="2025-07-07T00:06:15.950732723Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-94bd567f5-f7gkf,Uid:3c73d839-0227-44af-9a41-a794da30c843,Namespace:calico-system,Attempt:0,} returns sandbox id \"8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be\"" Jul 7 00:06:15.954398 containerd[1456]: time="2025-07-07T00:06:15.953651684Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\"" Jul 7 00:06:16.151425 kubelet[2515]: I0707 00:06:16.151235 2515 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd6ca0df-602a-496f-a023-657b3d62c9da" path="/var/lib/kubelet/pods/cd6ca0df-602a-496f-a023-657b3d62c9da/volumes" Jul 7 00:06:16.354930 kubelet[2515]: I0707 00:06:16.354641 2515 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 7 00:06:16.846133 containerd[1456]: time="2025-07-07T00:06:16.846071423Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:16.846912 containerd[1456]: time="2025-07-07T00:06:16.846880456Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.2: active requests=0, bytes read=4661207" Jul 7 00:06:16.847846 containerd[1456]: time="2025-07-07T00:06:16.847823769Z" level=info msg="ImageCreate event name:\"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:16.852381 containerd[1456]: time="2025-07-07T00:06:16.851270390Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:16.852974 containerd[1456]: time="2025-07-07T00:06:16.852952684Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.2\" with image id \"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678\", size \"6153902\" in 899.273846ms" Jul 7 00:06:16.853058 containerd[1456]: time="2025-07-07T00:06:16.853037377Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\" returns image reference \"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\"" Jul 7 00:06:16.856261 containerd[1456]: time="2025-07-07T00:06:16.856233229Z" level=info msg="CreateContainer within sandbox \"8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Jul 7 00:06:16.867953 containerd[1456]: time="2025-07-07T00:06:16.867921606Z" level=info msg="CreateContainer within sandbox \"8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"1a9f5d07a3224dc18065b17a561a43fe812cfbbe97276921c1feae95ec25b17c\"" Jul 7 00:06:16.870445 containerd[1456]: time="2025-07-07T00:06:16.868393837Z" level=info msg="StartContainer for \"1a9f5d07a3224dc18065b17a561a43fe812cfbbe97276921c1feae95ec25b17c\"" Jul 7 00:06:16.906536 systemd[1]: Started cri-containerd-1a9f5d07a3224dc18065b17a561a43fe812cfbbe97276921c1feae95ec25b17c.scope - libcontainer container 1a9f5d07a3224dc18065b17a561a43fe812cfbbe97276921c1feae95ec25b17c. Jul 7 00:06:16.948394 containerd[1456]: time="2025-07-07T00:06:16.947909254Z" level=info msg="StartContainer for \"1a9f5d07a3224dc18065b17a561a43fe812cfbbe97276921c1feae95ec25b17c\" returns successfully" Jul 7 00:06:16.950033 containerd[1456]: time="2025-07-07T00:06:16.949656298Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\"" Jul 7 00:06:17.890559 systemd-networkd[1382]: calib899fbd8f5b: Gained IPv6LL Jul 7 00:06:18.190064 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1807442897.mount: Deactivated successfully. Jul 7 00:06:18.200644 containerd[1456]: time="2025-07-07T00:06:18.200601494Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:18.201492 containerd[1456]: time="2025-07-07T00:06:18.201333282Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.2: active requests=0, bytes read=33083477" Jul 7 00:06:18.203025 containerd[1456]: time="2025-07-07T00:06:18.201992109Z" level=info msg="ImageCreate event name:\"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:18.204240 containerd[1456]: time="2025-07-07T00:06:18.203520993Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:18.204240 containerd[1456]: time="2025-07-07T00:06:18.204143484Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" with image id \"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5\", size \"33083307\" in 1.254111382s" Jul 7 00:06:18.204240 containerd[1456]: time="2025-07-07T00:06:18.204167718Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" returns image reference \"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\"" Jul 7 00:06:18.207072 containerd[1456]: time="2025-07-07T00:06:18.207049661Z" level=info msg="CreateContainer within sandbox \"8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Jul 7 00:06:18.219674 containerd[1456]: time="2025-07-07T00:06:18.219614249Z" level=info msg="CreateContainer within sandbox \"8ac992d35a184fa693743884d34409dfb6db33f2d7434de21e0b3ee09689b6be\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"aed943ec4b8f0c66c66021ce802c4bf0d3fa74d074f83d2fb031bbcf109c46f4\"" Jul 7 00:06:18.219950 containerd[1456]: time="2025-07-07T00:06:18.219932280Z" level=info msg="StartContainer for \"aed943ec4b8f0c66c66021ce802c4bf0d3fa74d074f83d2fb031bbcf109c46f4\"" Jul 7 00:06:18.250520 systemd[1]: Started cri-containerd-aed943ec4b8f0c66c66021ce802c4bf0d3fa74d074f83d2fb031bbcf109c46f4.scope - libcontainer container aed943ec4b8f0c66c66021ce802c4bf0d3fa74d074f83d2fb031bbcf109c46f4. Jul 7 00:06:18.307393 containerd[1456]: time="2025-07-07T00:06:18.307038955Z" level=info msg="StartContainer for \"aed943ec4b8f0c66c66021ce802c4bf0d3fa74d074f83d2fb031bbcf109c46f4\" returns successfully" Jul 7 00:06:18.375742 kubelet[2515]: I0707 00:06:18.375684 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-94bd567f5-f7gkf" podStartSLOduration=1.123576833 podStartE2EDuration="3.375669227s" podCreationTimestamp="2025-07-07 00:06:15 +0000 UTC" firstStartedPulling="2025-07-07 00:06:15.952975054 +0000 UTC m=+27.920845367" lastFinishedPulling="2025-07-07 00:06:18.205067438 +0000 UTC m=+30.172937761" observedRunningTime="2025-07-07 00:06:18.375248691 +0000 UTC m=+30.343119014" watchObservedRunningTime="2025-07-07 00:06:18.375669227 +0000 UTC m=+30.343539540" Jul 7 00:06:18.752094 kubelet[2515]: I0707 00:06:18.751208 2515 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 7 00:06:18.752094 kubelet[2515]: E0707 00:06:18.751528 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:19.367292 kubelet[2515]: E0707 00:06:19.367244 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:19.867403 kernel: bpftool[4103]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Jul 7 00:06:20.098341 systemd-networkd[1382]: vxlan.calico: Link UP Jul 7 00:06:20.098352 systemd-networkd[1382]: vxlan.calico: Gained carrier Jul 7 00:06:22.115089 systemd-networkd[1382]: vxlan.calico: Gained IPv6LL Jul 7 00:06:24.145837 containerd[1456]: time="2025-07-07T00:06:24.145750763Z" level=info msg="StopPodSandbox for \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\"" Jul 7 00:06:24.240859 containerd[1456]: 2025-07-07 00:06:24.201 [INFO][4187] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Jul 7 00:06:24.240859 containerd[1456]: 2025-07-07 00:06:24.202 [INFO][4187] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" iface="eth0" netns="/var/run/netns/cni-99aff9d8-3445-b78a-329b-1ddc5db7f7f1" Jul 7 00:06:24.240859 containerd[1456]: 2025-07-07 00:06:24.203 [INFO][4187] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" iface="eth0" netns="/var/run/netns/cni-99aff9d8-3445-b78a-329b-1ddc5db7f7f1" Jul 7 00:06:24.240859 containerd[1456]: 2025-07-07 00:06:24.203 [INFO][4187] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" iface="eth0" netns="/var/run/netns/cni-99aff9d8-3445-b78a-329b-1ddc5db7f7f1" Jul 7 00:06:24.240859 containerd[1456]: 2025-07-07 00:06:24.203 [INFO][4187] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Jul 7 00:06:24.240859 containerd[1456]: 2025-07-07 00:06:24.203 [INFO][4187] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Jul 7 00:06:24.240859 containerd[1456]: 2025-07-07 00:06:24.227 [INFO][4195] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" HandleID="k8s-pod-network.b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:24.240859 containerd[1456]: 2025-07-07 00:06:24.227 [INFO][4195] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:24.240859 containerd[1456]: 2025-07-07 00:06:24.227 [INFO][4195] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:24.240859 containerd[1456]: 2025-07-07 00:06:24.233 [WARNING][4195] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" HandleID="k8s-pod-network.b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:24.240859 containerd[1456]: 2025-07-07 00:06:24.233 [INFO][4195] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" HandleID="k8s-pod-network.b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:24.240859 containerd[1456]: 2025-07-07 00:06:24.235 [INFO][4195] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:24.240859 containerd[1456]: 2025-07-07 00:06:24.237 [INFO][4187] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Jul 7 00:06:24.243317 containerd[1456]: time="2025-07-07T00:06:24.241122584Z" level=info msg="TearDown network for sandbox \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\" successfully" Jul 7 00:06:24.243317 containerd[1456]: time="2025-07-07T00:06:24.241181930Z" level=info msg="StopPodSandbox for \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\" returns successfully" Jul 7 00:06:24.244447 kubelet[2515]: E0707 00:06:24.243579 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:24.247117 containerd[1456]: time="2025-07-07T00:06:24.245457784Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-pdhcs,Uid:6b2ab80d-76dc-4520-bb95-f08a370b3d13,Namespace:kube-system,Attempt:1,}" Jul 7 00:06:24.246221 systemd[1]: run-netns-cni\x2d99aff9d8\x2d3445\x2db78a\x2d329b\x2d1ddc5db7f7f1.mount: Deactivated successfully. Jul 7 00:06:24.273318 kubelet[2515]: I0707 00:06:24.272458 2515 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 7 00:06:24.386635 systemd-networkd[1382]: cali2b843ae349c: Link UP Jul 7 00:06:24.388572 systemd-networkd[1382]: cali2b843ae349c: Gained carrier Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.308 [INFO][4202] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0 coredns-7c65d6cfc9- kube-system 6b2ab80d-76dc-4520-bb95-f08a370b3d13 960 0 2025-07-07 00:05:53 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s 172-234-31-125 coredns-7c65d6cfc9-pdhcs eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali2b843ae349c [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" Namespace="kube-system" Pod="coredns-7c65d6cfc9-pdhcs" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-" Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.309 [INFO][4202] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" Namespace="kube-system" Pod="coredns-7c65d6cfc9-pdhcs" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.339 [INFO][4230] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" HandleID="k8s-pod-network.b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.339 [INFO][4230] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" HandleID="k8s-pod-network.b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d4fd0), Attrs:map[string]string{"namespace":"kube-system", "node":"172-234-31-125", "pod":"coredns-7c65d6cfc9-pdhcs", "timestamp":"2025-07-07 00:06:24.339820295 +0000 UTC"}, Hostname:"172-234-31-125", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.340 [INFO][4230] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.340 [INFO][4230] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.340 [INFO][4230] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-234-31-125' Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.347 [INFO][4230] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" host="172-234-31-125" Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.351 [INFO][4230] ipam/ipam.go 394: Looking up existing affinities for host host="172-234-31-125" Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.356 [INFO][4230] ipam/ipam.go 511: Trying affinity for 192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.358 [INFO][4230] ipam/ipam.go 158: Attempting to load block cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.360 [INFO][4230] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.360 [INFO][4230] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.38.192/26 handle="k8s-pod-network.b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" host="172-234-31-125" Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.362 [INFO][4230] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4 Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.365 [INFO][4230] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.38.192/26 handle="k8s-pod-network.b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" host="172-234-31-125" Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.370 [INFO][4230] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.38.194/26] block=192.168.38.192/26 handle="k8s-pod-network.b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" host="172-234-31-125" Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.371 [INFO][4230] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.38.194/26] handle="k8s-pod-network.b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" host="172-234-31-125" Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.371 [INFO][4230] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:24.411986 containerd[1456]: 2025-07-07 00:06:24.371 [INFO][4230] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.38.194/26] IPv6=[] ContainerID="b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" HandleID="k8s-pod-network.b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:24.413282 containerd[1456]: 2025-07-07 00:06:24.375 [INFO][4202] cni-plugin/k8s.go 418: Populated endpoint ContainerID="b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" Namespace="kube-system" Pod="coredns-7c65d6cfc9-pdhcs" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"6b2ab80d-76dc-4520-bb95-f08a370b3d13", ResourceVersion:"960", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 5, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"", Pod:"coredns-7c65d6cfc9-pdhcs", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.38.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2b843ae349c", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:24.413282 containerd[1456]: 2025-07-07 00:06:24.375 [INFO][4202] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.38.194/32] ContainerID="b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" Namespace="kube-system" Pod="coredns-7c65d6cfc9-pdhcs" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:24.413282 containerd[1456]: 2025-07-07 00:06:24.375 [INFO][4202] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2b843ae349c ContainerID="b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" Namespace="kube-system" Pod="coredns-7c65d6cfc9-pdhcs" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:24.413282 containerd[1456]: 2025-07-07 00:06:24.389 [INFO][4202] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" Namespace="kube-system" Pod="coredns-7c65d6cfc9-pdhcs" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:24.413282 containerd[1456]: 2025-07-07 00:06:24.389 [INFO][4202] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" Namespace="kube-system" Pod="coredns-7c65d6cfc9-pdhcs" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"6b2ab80d-76dc-4520-bb95-f08a370b3d13", ResourceVersion:"960", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 5, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4", Pod:"coredns-7c65d6cfc9-pdhcs", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.38.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2b843ae349c", MAC:"de:5d:04:76:ac:fe", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:24.413282 containerd[1456]: 2025-07-07 00:06:24.404 [INFO][4202] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4" Namespace="kube-system" Pod="coredns-7c65d6cfc9-pdhcs" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:24.452553 containerd[1456]: time="2025-07-07T00:06:24.451534817Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:06:24.452869 containerd[1456]: time="2025-07-07T00:06:24.452723714Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:06:24.452869 containerd[1456]: time="2025-07-07T00:06:24.452814512Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:24.453095 containerd[1456]: time="2025-07-07T00:06:24.453012760Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:24.491769 systemd[1]: Started cri-containerd-b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4.scope - libcontainer container b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4. Jul 7 00:06:24.548206 containerd[1456]: time="2025-07-07T00:06:24.548128618Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-pdhcs,Uid:6b2ab80d-76dc-4520-bb95-f08a370b3d13,Namespace:kube-system,Attempt:1,} returns sandbox id \"b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4\"" Jul 7 00:06:24.549541 kubelet[2515]: E0707 00:06:24.549511 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:24.552672 containerd[1456]: time="2025-07-07T00:06:24.552625653Z" level=info msg="CreateContainer within sandbox \"b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jul 7 00:06:24.564683 containerd[1456]: time="2025-07-07T00:06:24.564638714Z" level=info msg="CreateContainer within sandbox \"b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"59919e9dfd537b4711b0a6e0b8a1eeef79bd5ad1eb34409b96f6504575a97f54\"" Jul 7 00:06:24.566207 containerd[1456]: time="2025-07-07T00:06:24.565304263Z" level=info msg="StartContainer for \"59919e9dfd537b4711b0a6e0b8a1eeef79bd5ad1eb34409b96f6504575a97f54\"" Jul 7 00:06:24.603542 systemd[1]: Started cri-containerd-59919e9dfd537b4711b0a6e0b8a1eeef79bd5ad1eb34409b96f6504575a97f54.scope - libcontainer container 59919e9dfd537b4711b0a6e0b8a1eeef79bd5ad1eb34409b96f6504575a97f54. Jul 7 00:06:24.634855 containerd[1456]: time="2025-07-07T00:06:24.634815068Z" level=info msg="StartContainer for \"59919e9dfd537b4711b0a6e0b8a1eeef79bd5ad1eb34409b96f6504575a97f54\" returns successfully" Jul 7 00:06:25.384706 kubelet[2515]: E0707 00:06:25.384657 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:25.413733 kubelet[2515]: I0707 00:06:25.412508 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-pdhcs" podStartSLOduration=32.412488396 podStartE2EDuration="32.412488396s" podCreationTimestamp="2025-07-07 00:05:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-07 00:06:25.401549933 +0000 UTC m=+37.369420246" watchObservedRunningTime="2025-07-07 00:06:25.412488396 +0000 UTC m=+37.380358709" Jul 7 00:06:26.149153 containerd[1456]: time="2025-07-07T00:06:26.147883559Z" level=info msg="StopPodSandbox for \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\"" Jul 7 00:06:26.151081 containerd[1456]: time="2025-07-07T00:06:26.150630627Z" level=info msg="StopPodSandbox for \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\"" Jul 7 00:06:26.157943 containerd[1456]: time="2025-07-07T00:06:26.157676473Z" level=info msg="StopPodSandbox for \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\"" Jul 7 00:06:26.283092 systemd-networkd[1382]: cali2b843ae349c: Gained IPv6LL Jul 7 00:06:26.354508 containerd[1456]: 2025-07-07 00:06:26.266 [INFO][4392] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Jul 7 00:06:26.354508 containerd[1456]: 2025-07-07 00:06:26.266 [INFO][4392] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" iface="eth0" netns="/var/run/netns/cni-4ff48958-45dd-f8d2-b328-72c4d8cb5d6c" Jul 7 00:06:26.354508 containerd[1456]: 2025-07-07 00:06:26.267 [INFO][4392] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" iface="eth0" netns="/var/run/netns/cni-4ff48958-45dd-f8d2-b328-72c4d8cb5d6c" Jul 7 00:06:26.354508 containerd[1456]: 2025-07-07 00:06:26.275 [INFO][4392] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" iface="eth0" netns="/var/run/netns/cni-4ff48958-45dd-f8d2-b328-72c4d8cb5d6c" Jul 7 00:06:26.354508 containerd[1456]: 2025-07-07 00:06:26.276 [INFO][4392] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Jul 7 00:06:26.354508 containerd[1456]: 2025-07-07 00:06:26.276 [INFO][4392] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Jul 7 00:06:26.354508 containerd[1456]: 2025-07-07 00:06:26.336 [INFO][4425] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" HandleID="k8s-pod-network.4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Workload="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:26.354508 containerd[1456]: 2025-07-07 00:06:26.337 [INFO][4425] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:26.354508 containerd[1456]: 2025-07-07 00:06:26.337 [INFO][4425] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:26.354508 containerd[1456]: 2025-07-07 00:06:26.346 [WARNING][4425] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" HandleID="k8s-pod-network.4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Workload="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:26.354508 containerd[1456]: 2025-07-07 00:06:26.346 [INFO][4425] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" HandleID="k8s-pod-network.4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Workload="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:26.354508 containerd[1456]: 2025-07-07 00:06:26.347 [INFO][4425] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:26.354508 containerd[1456]: 2025-07-07 00:06:26.350 [INFO][4392] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Jul 7 00:06:26.354508 containerd[1456]: time="2025-07-07T00:06:26.354503707Z" level=info msg="TearDown network for sandbox \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\" successfully" Jul 7 00:06:26.354992 containerd[1456]: time="2025-07-07T00:06:26.354528119Z" level=info msg="StopPodSandbox for \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\" returns successfully" Jul 7 00:06:26.359257 systemd[1]: run-netns-cni\x2d4ff48958\x2d45dd\x2df8d2\x2db328\x2d72c4d8cb5d6c.mount: Deactivated successfully. Jul 7 00:06:26.362825 containerd[1456]: time="2025-07-07T00:06:26.362140651Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-94599b4db-wbn2z,Uid:7f03664a-b5c1-41d9-b19a-954bfd2a5ea7,Namespace:calico-system,Attempt:1,}" Jul 7 00:06:26.383896 containerd[1456]: 2025-07-07 00:06:26.238 [INFO][4397] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Jul 7 00:06:26.383896 containerd[1456]: 2025-07-07 00:06:26.239 [INFO][4397] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" iface="eth0" netns="/var/run/netns/cni-14a5a712-e3f6-0dce-ba61-6f044b74c34b" Jul 7 00:06:26.383896 containerd[1456]: 2025-07-07 00:06:26.239 [INFO][4397] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" iface="eth0" netns="/var/run/netns/cni-14a5a712-e3f6-0dce-ba61-6f044b74c34b" Jul 7 00:06:26.383896 containerd[1456]: 2025-07-07 00:06:26.239 [INFO][4397] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" iface="eth0" netns="/var/run/netns/cni-14a5a712-e3f6-0dce-ba61-6f044b74c34b" Jul 7 00:06:26.383896 containerd[1456]: 2025-07-07 00:06:26.239 [INFO][4397] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Jul 7 00:06:26.383896 containerd[1456]: 2025-07-07 00:06:26.239 [INFO][4397] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Jul 7 00:06:26.383896 containerd[1456]: 2025-07-07 00:06:26.335 [INFO][4414] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" HandleID="k8s-pod-network.3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Workload="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:26.383896 containerd[1456]: 2025-07-07 00:06:26.341 [INFO][4414] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:26.383896 containerd[1456]: 2025-07-07 00:06:26.348 [INFO][4414] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:26.383896 containerd[1456]: 2025-07-07 00:06:26.359 [WARNING][4414] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" HandleID="k8s-pod-network.3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Workload="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:26.383896 containerd[1456]: 2025-07-07 00:06:26.359 [INFO][4414] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" HandleID="k8s-pod-network.3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Workload="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:26.383896 containerd[1456]: 2025-07-07 00:06:26.362 [INFO][4414] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:26.383896 containerd[1456]: 2025-07-07 00:06:26.372 [INFO][4397] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Jul 7 00:06:26.384423 containerd[1456]: time="2025-07-07T00:06:26.384132238Z" level=info msg="TearDown network for sandbox \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\" successfully" Jul 7 00:06:26.384423 containerd[1456]: time="2025-07-07T00:06:26.384168871Z" level=info msg="StopPodSandbox for \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\" returns successfully" Jul 7 00:06:26.389392 containerd[1456]: time="2025-07-07T00:06:26.388263765Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-d799z,Uid:06260b4f-49a9-4902-b2df-d9734428c436,Namespace:calico-system,Attempt:1,}" Jul 7 00:06:26.390270 kubelet[2515]: E0707 00:06:26.389144 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:26.393199 systemd[1]: run-netns-cni\x2d14a5a712\x2de3f6\x2d0dce\x2dba61\x2d6f044b74c34b.mount: Deactivated successfully. Jul 7 00:06:26.400710 containerd[1456]: 2025-07-07 00:06:26.250 [INFO][4396] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Jul 7 00:06:26.400710 containerd[1456]: 2025-07-07 00:06:26.251 [INFO][4396] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" iface="eth0" netns="/var/run/netns/cni-09428728-17fe-35d7-5f07-1705d2a7c1e1" Jul 7 00:06:26.400710 containerd[1456]: 2025-07-07 00:06:26.252 [INFO][4396] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" iface="eth0" netns="/var/run/netns/cni-09428728-17fe-35d7-5f07-1705d2a7c1e1" Jul 7 00:06:26.400710 containerd[1456]: 2025-07-07 00:06:26.252 [INFO][4396] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" iface="eth0" netns="/var/run/netns/cni-09428728-17fe-35d7-5f07-1705d2a7c1e1" Jul 7 00:06:26.400710 containerd[1456]: 2025-07-07 00:06:26.252 [INFO][4396] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Jul 7 00:06:26.400710 containerd[1456]: 2025-07-07 00:06:26.253 [INFO][4396] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Jul 7 00:06:26.400710 containerd[1456]: 2025-07-07 00:06:26.340 [INFO][4419] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" HandleID="k8s-pod-network.455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Workload="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:26.400710 containerd[1456]: 2025-07-07 00:06:26.341 [INFO][4419] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:26.400710 containerd[1456]: 2025-07-07 00:06:26.362 [INFO][4419] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:26.400710 containerd[1456]: 2025-07-07 00:06:26.370 [WARNING][4419] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" HandleID="k8s-pod-network.455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Workload="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:26.400710 containerd[1456]: 2025-07-07 00:06:26.370 [INFO][4419] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" HandleID="k8s-pod-network.455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Workload="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:26.400710 containerd[1456]: 2025-07-07 00:06:26.372 [INFO][4419] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:26.400710 containerd[1456]: 2025-07-07 00:06:26.387 [INFO][4396] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Jul 7 00:06:26.400710 containerd[1456]: time="2025-07-07T00:06:26.397600752Z" level=info msg="TearDown network for sandbox \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\" successfully" Jul 7 00:06:26.400710 containerd[1456]: time="2025-07-07T00:06:26.397616884Z" level=info msg="StopPodSandbox for \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\" returns successfully" Jul 7 00:06:26.401708 containerd[1456]: time="2025-07-07T00:06:26.401641162Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-9knjc,Uid:2001a2e7-8165-484d-b2ef-2ba8a539af9a,Namespace:calico-system,Attempt:1,}" Jul 7 00:06:26.405227 systemd[1]: run-netns-cni\x2d09428728\x2d17fe\x2d35d7\x2d5f07\x2d1705d2a7c1e1.mount: Deactivated successfully. Jul 7 00:06:26.627059 systemd-networkd[1382]: cali4a975bd8445: Link UP Jul 7 00:06:26.628533 systemd-networkd[1382]: cali4a975bd8445: Gained carrier Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.482 [INFO][4445] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0 goldmane-58fd7646b9- calico-system 06260b4f-49a9-4902-b2df-d9734428c436 987 0 2025-07-07 00:06:04 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:58fd7646b9 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s 172-234-31-125 goldmane-58fd7646b9-d799z eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali4a975bd8445 [] [] }} ContainerID="5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" Namespace="calico-system" Pod="goldmane-58fd7646b9-d799z" WorkloadEndpoint="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-" Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.483 [INFO][4445] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" Namespace="calico-system" Pod="goldmane-58fd7646b9-d799z" WorkloadEndpoint="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.552 [INFO][4470] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" HandleID="k8s-pod-network.5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" Workload="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.554 [INFO][4470] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" HandleID="k8s-pod-network.5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" Workload="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000100280), Attrs:map[string]string{"namespace":"calico-system", "node":"172-234-31-125", "pod":"goldmane-58fd7646b9-d799z", "timestamp":"2025-07-07 00:06:26.551498344 +0000 UTC"}, Hostname:"172-234-31-125", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.554 [INFO][4470] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.554 [INFO][4470] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.554 [INFO][4470] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-234-31-125' Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.566 [INFO][4470] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" host="172-234-31-125" Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.572 [INFO][4470] ipam/ipam.go 394: Looking up existing affinities for host host="172-234-31-125" Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.576 [INFO][4470] ipam/ipam.go 511: Trying affinity for 192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.584 [INFO][4470] ipam/ipam.go 158: Attempting to load block cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.588 [INFO][4470] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.588 [INFO][4470] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.38.192/26 handle="k8s-pod-network.5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" host="172-234-31-125" Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.592 [INFO][4470] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.600 [INFO][4470] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.38.192/26 handle="k8s-pod-network.5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" host="172-234-31-125" Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.615 [INFO][4470] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.38.195/26] block=192.168.38.192/26 handle="k8s-pod-network.5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" host="172-234-31-125" Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.615 [INFO][4470] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.38.195/26] handle="k8s-pod-network.5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" host="172-234-31-125" Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.615 [INFO][4470] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:26.657564 containerd[1456]: 2025-07-07 00:06:26.615 [INFO][4470] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.38.195/26] IPv6=[] ContainerID="5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" HandleID="k8s-pod-network.5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" Workload="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:26.658202 containerd[1456]: 2025-07-07 00:06:26.621 [INFO][4445] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" Namespace="calico-system" Pod="goldmane-58fd7646b9-d799z" WorkloadEndpoint="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"06260b4f-49a9-4902-b2df-d9734428c436", ResourceVersion:"987", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"", Pod:"goldmane-58fd7646b9-d799z", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.38.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali4a975bd8445", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:26.658202 containerd[1456]: 2025-07-07 00:06:26.621 [INFO][4445] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.38.195/32] ContainerID="5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" Namespace="calico-system" Pod="goldmane-58fd7646b9-d799z" WorkloadEndpoint="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:26.658202 containerd[1456]: 2025-07-07 00:06:26.621 [INFO][4445] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4a975bd8445 ContainerID="5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" Namespace="calico-system" Pod="goldmane-58fd7646b9-d799z" WorkloadEndpoint="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:26.658202 containerd[1456]: 2025-07-07 00:06:26.627 [INFO][4445] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" Namespace="calico-system" Pod="goldmane-58fd7646b9-d799z" WorkloadEndpoint="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:26.658202 containerd[1456]: 2025-07-07 00:06:26.630 [INFO][4445] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" Namespace="calico-system" Pod="goldmane-58fd7646b9-d799z" WorkloadEndpoint="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"06260b4f-49a9-4902-b2df-d9734428c436", ResourceVersion:"987", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece", Pod:"goldmane-58fd7646b9-d799z", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.38.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali4a975bd8445", MAC:"f6:2b:70:5e:85:df", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:26.658202 containerd[1456]: 2025-07-07 00:06:26.653 [INFO][4445] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece" Namespace="calico-system" Pod="goldmane-58fd7646b9-d799z" WorkloadEndpoint="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:26.687840 containerd[1456]: time="2025-07-07T00:06:26.687386852Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:06:26.687840 containerd[1456]: time="2025-07-07T00:06:26.687613370Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:06:26.687840 containerd[1456]: time="2025-07-07T00:06:26.687765662Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:26.688621 containerd[1456]: time="2025-07-07T00:06:26.688395072Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:26.715761 systemd[1]: Started cri-containerd-5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece.scope - libcontainer container 5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece. Jul 7 00:06:26.776064 systemd-networkd[1382]: cali94d73bf1cb2: Link UP Jul 7 00:06:26.784002 systemd-networkd[1382]: cali94d73bf1cb2: Gained carrier Jul 7 00:06:26.802671 containerd[1456]: time="2025-07-07T00:06:26.802159272Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-d799z,Uid:06260b4f-49a9-4902-b2df-d9734428c436,Namespace:calico-system,Attempt:1,} returns sandbox id \"5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece\"" Jul 7 00:06:26.818471 containerd[1456]: time="2025-07-07T00:06:26.816848862Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\"" Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.488 [INFO][4436] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0 calico-kube-controllers-94599b4db- calico-system 7f03664a-b5c1-41d9-b19a-954bfd2a5ea7 989 0 2025-07-07 00:06:05 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:94599b4db projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s 172-234-31-125 calico-kube-controllers-94599b4db-wbn2z eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali94d73bf1cb2 [] [] }} ContainerID="4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" Namespace="calico-system" Pod="calico-kube-controllers-94599b4db-wbn2z" WorkloadEndpoint="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-" Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.489 [INFO][4436] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" Namespace="calico-system" Pod="calico-kube-controllers-94599b4db-wbn2z" WorkloadEndpoint="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.561 [INFO][4475] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" HandleID="k8s-pod-network.4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" Workload="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.562 [INFO][4475] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" HandleID="k8s-pod-network.4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" Workload="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00031da40), Attrs:map[string]string{"namespace":"calico-system", "node":"172-234-31-125", "pod":"calico-kube-controllers-94599b4db-wbn2z", "timestamp":"2025-07-07 00:06:26.561884175 +0000 UTC"}, Hostname:"172-234-31-125", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.562 [INFO][4475] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.615 [INFO][4475] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.616 [INFO][4475] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-234-31-125' Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.667 [INFO][4475] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" host="172-234-31-125" Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.675 [INFO][4475] ipam/ipam.go 394: Looking up existing affinities for host host="172-234-31-125" Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.685 [INFO][4475] ipam/ipam.go 511: Trying affinity for 192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.691 [INFO][4475] ipam/ipam.go 158: Attempting to load block cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.694 [INFO][4475] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.695 [INFO][4475] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.38.192/26 handle="k8s-pod-network.4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" host="172-234-31-125" Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.702 [INFO][4475] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.720 [INFO][4475] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.38.192/26 handle="k8s-pod-network.4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" host="172-234-31-125" Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.746 [INFO][4475] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.38.196/26] block=192.168.38.192/26 handle="k8s-pod-network.4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" host="172-234-31-125" Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.746 [INFO][4475] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.38.196/26] handle="k8s-pod-network.4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" host="172-234-31-125" Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.746 [INFO][4475] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:26.823428 containerd[1456]: 2025-07-07 00:06:26.746 [INFO][4475] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.38.196/26] IPv6=[] ContainerID="4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" HandleID="k8s-pod-network.4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" Workload="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:26.824669 containerd[1456]: 2025-07-07 00:06:26.758 [INFO][4436] cni-plugin/k8s.go 418: Populated endpoint ContainerID="4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" Namespace="calico-system" Pod="calico-kube-controllers-94599b4db-wbn2z" WorkloadEndpoint="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0", GenerateName:"calico-kube-controllers-94599b4db-", Namespace:"calico-system", SelfLink:"", UID:"7f03664a-b5c1-41d9-b19a-954bfd2a5ea7", ResourceVersion:"989", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"94599b4db", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"", Pod:"calico-kube-controllers-94599b4db-wbn2z", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.38.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali94d73bf1cb2", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:26.824669 containerd[1456]: 2025-07-07 00:06:26.759 [INFO][4436] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.38.196/32] ContainerID="4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" Namespace="calico-system" Pod="calico-kube-controllers-94599b4db-wbn2z" WorkloadEndpoint="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:26.824669 containerd[1456]: 2025-07-07 00:06:26.759 [INFO][4436] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali94d73bf1cb2 ContainerID="4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" Namespace="calico-system" Pod="calico-kube-controllers-94599b4db-wbn2z" WorkloadEndpoint="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:26.824669 containerd[1456]: 2025-07-07 00:06:26.785 [INFO][4436] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" Namespace="calico-system" Pod="calico-kube-controllers-94599b4db-wbn2z" WorkloadEndpoint="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:26.824669 containerd[1456]: 2025-07-07 00:06:26.792 [INFO][4436] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" Namespace="calico-system" Pod="calico-kube-controllers-94599b4db-wbn2z" WorkloadEndpoint="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0", GenerateName:"calico-kube-controllers-94599b4db-", Namespace:"calico-system", SelfLink:"", UID:"7f03664a-b5c1-41d9-b19a-954bfd2a5ea7", ResourceVersion:"989", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"94599b4db", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b", Pod:"calico-kube-controllers-94599b4db-wbn2z", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.38.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali94d73bf1cb2", MAC:"82:93:87:c4:ec:bb", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:26.824669 containerd[1456]: 2025-07-07 00:06:26.819 [INFO][4436] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b" Namespace="calico-system" Pod="calico-kube-controllers-94599b4db-wbn2z" WorkloadEndpoint="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:26.856253 containerd[1456]: time="2025-07-07T00:06:26.856046390Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:06:26.856253 containerd[1456]: time="2025-07-07T00:06:26.856173920Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:06:26.856253 containerd[1456]: time="2025-07-07T00:06:26.856203262Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:26.857731 containerd[1456]: time="2025-07-07T00:06:26.857648076Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:26.901471 systemd-networkd[1382]: calif76aa88d3ce: Link UP Jul 7 00:06:26.902502 systemd[1]: Started cri-containerd-4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b.scope - libcontainer container 4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b. Jul 7 00:06:26.903774 systemd-networkd[1382]: calif76aa88d3ce: Gained carrier Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.488 [INFO][4447] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--234--31--125-k8s-csi--node--driver--9knjc-eth0 csi-node-driver- calico-system 2001a2e7-8165-484d-b2ef-2ba8a539af9a 988 0 2025-07-07 00:06:05 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:57bd658777 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s 172-234-31-125 csi-node-driver-9knjc eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calif76aa88d3ce [] [] }} ContainerID="9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" Namespace="calico-system" Pod="csi-node-driver-9knjc" WorkloadEndpoint="172--234--31--125-k8s-csi--node--driver--9knjc-" Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.489 [INFO][4447] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" Namespace="calico-system" Pod="csi-node-driver-9knjc" WorkloadEndpoint="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.569 [INFO][4477] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" HandleID="k8s-pod-network.9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" Workload="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.569 [INFO][4477] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" HandleID="k8s-pod-network.9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" Workload="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000156330), Attrs:map[string]string{"namespace":"calico-system", "node":"172-234-31-125", "pod":"csi-node-driver-9knjc", "timestamp":"2025-07-07 00:06:26.569475964 +0000 UTC"}, Hostname:"172-234-31-125", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.569 [INFO][4477] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.748 [INFO][4477] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.748 [INFO][4477] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-234-31-125' Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.787 [INFO][4477] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" host="172-234-31-125" Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.814 [INFO][4477] ipam/ipam.go 394: Looking up existing affinities for host host="172-234-31-125" Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.829 [INFO][4477] ipam/ipam.go 511: Trying affinity for 192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.832 [INFO][4477] ipam/ipam.go 158: Attempting to load block cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.841 [INFO][4477] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.841 [INFO][4477] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.38.192/26 handle="k8s-pod-network.9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" host="172-234-31-125" Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.846 [INFO][4477] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402 Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.853 [INFO][4477] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.38.192/26 handle="k8s-pod-network.9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" host="172-234-31-125" Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.863 [INFO][4477] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.38.197/26] block=192.168.38.192/26 handle="k8s-pod-network.9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" host="172-234-31-125" Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.864 [INFO][4477] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.38.197/26] handle="k8s-pod-network.9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" host="172-234-31-125" Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.864 [INFO][4477] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:26.937955 containerd[1456]: 2025-07-07 00:06:26.864 [INFO][4477] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.38.197/26] IPv6=[] ContainerID="9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" HandleID="k8s-pod-network.9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" Workload="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:26.938564 containerd[1456]: 2025-07-07 00:06:26.871 [INFO][4447] cni-plugin/k8s.go 418: Populated endpoint ContainerID="9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" Namespace="calico-system" Pod="csi-node-driver-9knjc" WorkloadEndpoint="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-csi--node--driver--9knjc-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"2001a2e7-8165-484d-b2ef-2ba8a539af9a", ResourceVersion:"988", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"", Pod:"csi-node-driver-9knjc", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.38.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif76aa88d3ce", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:26.938564 containerd[1456]: 2025-07-07 00:06:26.871 [INFO][4447] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.38.197/32] ContainerID="9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" Namespace="calico-system" Pod="csi-node-driver-9knjc" WorkloadEndpoint="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:26.938564 containerd[1456]: 2025-07-07 00:06:26.871 [INFO][4447] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif76aa88d3ce ContainerID="9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" Namespace="calico-system" Pod="csi-node-driver-9knjc" WorkloadEndpoint="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:26.938564 containerd[1456]: 2025-07-07 00:06:26.909 [INFO][4447] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" Namespace="calico-system" Pod="csi-node-driver-9knjc" WorkloadEndpoint="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:26.938564 containerd[1456]: 2025-07-07 00:06:26.910 [INFO][4447] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" Namespace="calico-system" Pod="csi-node-driver-9knjc" WorkloadEndpoint="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-csi--node--driver--9knjc-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"2001a2e7-8165-484d-b2ef-2ba8a539af9a", ResourceVersion:"988", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402", Pod:"csi-node-driver-9knjc", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.38.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif76aa88d3ce", MAC:"76:83:19:89:fe:0e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:26.938564 containerd[1456]: 2025-07-07 00:06:26.932 [INFO][4447] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402" Namespace="calico-system" Pod="csi-node-driver-9knjc" WorkloadEndpoint="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:26.979769 containerd[1456]: time="2025-07-07T00:06:26.979550379Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:06:26.979905 containerd[1456]: time="2025-07-07T00:06:26.979781848Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:06:26.979905 containerd[1456]: time="2025-07-07T00:06:26.979854963Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:26.981181 containerd[1456]: time="2025-07-07T00:06:26.980947299Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:27.021588 systemd[1]: Started cri-containerd-9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402.scope - libcontainer container 9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402. Jul 7 00:06:27.106977 containerd[1456]: time="2025-07-07T00:06:27.106795830Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-94599b4db-wbn2z,Uid:7f03664a-b5c1-41d9-b19a-954bfd2a5ea7,Namespace:calico-system,Attempt:1,} returns sandbox id \"4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b\"" Jul 7 00:06:27.133993 containerd[1456]: time="2025-07-07T00:06:27.133553072Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-9knjc,Uid:2001a2e7-8165-484d-b2ef-2ba8a539af9a,Namespace:calico-system,Attempt:1,} returns sandbox id \"9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402\"" Jul 7 00:06:27.145057 containerd[1456]: time="2025-07-07T00:06:27.144992969Z" level=info msg="StopPodSandbox for \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\"" Jul 7 00:06:27.146541 containerd[1456]: time="2025-07-07T00:06:27.146508002Z" level=info msg="StopPodSandbox for \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\"" Jul 7 00:06:27.297773 containerd[1456]: 2025-07-07 00:06:27.231 [INFO][4664] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Jul 7 00:06:27.297773 containerd[1456]: 2025-07-07 00:06:27.232 [INFO][4664] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" iface="eth0" netns="/var/run/netns/cni-7e8bf2af-4960-eacb-b619-9b7c74b6491c" Jul 7 00:06:27.297773 containerd[1456]: 2025-07-07 00:06:27.232 [INFO][4664] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" iface="eth0" netns="/var/run/netns/cni-7e8bf2af-4960-eacb-b619-9b7c74b6491c" Jul 7 00:06:27.297773 containerd[1456]: 2025-07-07 00:06:27.232 [INFO][4664] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" iface="eth0" netns="/var/run/netns/cni-7e8bf2af-4960-eacb-b619-9b7c74b6491c" Jul 7 00:06:27.297773 containerd[1456]: 2025-07-07 00:06:27.232 [INFO][4664] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Jul 7 00:06:27.297773 containerd[1456]: 2025-07-07 00:06:27.233 [INFO][4664] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Jul 7 00:06:27.297773 containerd[1456]: 2025-07-07 00:06:27.263 [INFO][4677] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" HandleID="k8s-pod-network.071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:27.297773 containerd[1456]: 2025-07-07 00:06:27.263 [INFO][4677] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:27.297773 containerd[1456]: 2025-07-07 00:06:27.264 [INFO][4677] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:27.297773 containerd[1456]: 2025-07-07 00:06:27.281 [WARNING][4677] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" HandleID="k8s-pod-network.071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:27.297773 containerd[1456]: 2025-07-07 00:06:27.281 [INFO][4677] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" HandleID="k8s-pod-network.071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:27.297773 containerd[1456]: 2025-07-07 00:06:27.283 [INFO][4677] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:27.297773 containerd[1456]: 2025-07-07 00:06:27.289 [INFO][4664] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Jul 7 00:06:27.301839 containerd[1456]: time="2025-07-07T00:06:27.301533411Z" level=info msg="TearDown network for sandbox \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\" successfully" Jul 7 00:06:27.301839 containerd[1456]: time="2025-07-07T00:06:27.301647609Z" level=info msg="StopPodSandbox for \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\" returns successfully" Jul 7 00:06:27.306196 kubelet[2515]: E0707 00:06:27.306157 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:27.312061 containerd[1456]: time="2025-07-07T00:06:27.312018288Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-x9bb2,Uid:178fc278-1010-4c58-8d67-4a6ee0c1491b,Namespace:kube-system,Attempt:1,}" Jul 7 00:06:27.317947 containerd[1456]: 2025-07-07 00:06:27.234 [INFO][4663] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Jul 7 00:06:27.317947 containerd[1456]: 2025-07-07 00:06:27.234 [INFO][4663] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" iface="eth0" netns="/var/run/netns/cni-d01017db-2a7f-0968-f765-7ecb9c591594" Jul 7 00:06:27.317947 containerd[1456]: 2025-07-07 00:06:27.235 [INFO][4663] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" iface="eth0" netns="/var/run/netns/cni-d01017db-2a7f-0968-f765-7ecb9c591594" Jul 7 00:06:27.317947 containerd[1456]: 2025-07-07 00:06:27.235 [INFO][4663] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" iface="eth0" netns="/var/run/netns/cni-d01017db-2a7f-0968-f765-7ecb9c591594" Jul 7 00:06:27.317947 containerd[1456]: 2025-07-07 00:06:27.235 [INFO][4663] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Jul 7 00:06:27.317947 containerd[1456]: 2025-07-07 00:06:27.235 [INFO][4663] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Jul 7 00:06:27.317947 containerd[1456]: 2025-07-07 00:06:27.293 [INFO][4682] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" HandleID="k8s-pod-network.3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:27.317947 containerd[1456]: 2025-07-07 00:06:27.293 [INFO][4682] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:27.317947 containerd[1456]: 2025-07-07 00:06:27.293 [INFO][4682] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:27.317947 containerd[1456]: 2025-07-07 00:06:27.301 [WARNING][4682] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" HandleID="k8s-pod-network.3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:27.317947 containerd[1456]: 2025-07-07 00:06:27.301 [INFO][4682] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" HandleID="k8s-pod-network.3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:27.317947 containerd[1456]: 2025-07-07 00:06:27.303 [INFO][4682] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:27.317947 containerd[1456]: 2025-07-07 00:06:27.313 [INFO][4663] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Jul 7 00:06:27.318934 containerd[1456]: time="2025-07-07T00:06:27.318893276Z" level=info msg="TearDown network for sandbox \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\" successfully" Jul 7 00:06:27.318934 containerd[1456]: time="2025-07-07T00:06:27.318920438Z" level=info msg="StopPodSandbox for \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\" returns successfully" Jul 7 00:06:27.323045 containerd[1456]: time="2025-07-07T00:06:27.323001120Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85b7446c9f-8xc2r,Uid:4741c1ac-aedc-48ec-8809-16303ada3c56,Namespace:calico-apiserver,Attempt:1,}" Jul 7 00:06:27.376033 systemd[1]: run-netns-cni\x2d7e8bf2af\x2d4960\x2deacb\x2db619\x2d9b7c74b6491c.mount: Deactivated successfully. Jul 7 00:06:27.376139 systemd[1]: run-netns-cni\x2dd01017db\x2d2a7f\x2d0968\x2df765\x2d7ecb9c591594.mount: Deactivated successfully. Jul 7 00:06:27.396935 kubelet[2515]: E0707 00:06:27.396914 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:27.555488 systemd-networkd[1382]: cali602eb21fd0a: Link UP Jul 7 00:06:27.559996 systemd-networkd[1382]: cali602eb21fd0a: Gained carrier Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.426 [INFO][4690] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0 coredns-7c65d6cfc9- kube-system 178fc278-1010-4c58-8d67-4a6ee0c1491b 1008 0 2025-07-07 00:05:53 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s 172-234-31-125 coredns-7c65d6cfc9-x9bb2 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali602eb21fd0a [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" Namespace="kube-system" Pod="coredns-7c65d6cfc9-x9bb2" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-" Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.426 [INFO][4690] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" Namespace="kube-system" Pod="coredns-7c65d6cfc9-x9bb2" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.476 [INFO][4714] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" HandleID="k8s-pod-network.b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.477 [INFO][4714] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" HandleID="k8s-pod-network.b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5890), Attrs:map[string]string{"namespace":"kube-system", "node":"172-234-31-125", "pod":"coredns-7c65d6cfc9-x9bb2", "timestamp":"2025-07-07 00:06:27.476919708 +0000 UTC"}, Hostname:"172-234-31-125", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.478 [INFO][4714] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.478 [INFO][4714] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.478 [INFO][4714] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-234-31-125' Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.498 [INFO][4714] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" host="172-234-31-125" Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.505 [INFO][4714] ipam/ipam.go 394: Looking up existing affinities for host host="172-234-31-125" Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.522 [INFO][4714] ipam/ipam.go 511: Trying affinity for 192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.527 [INFO][4714] ipam/ipam.go 158: Attempting to load block cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.529 [INFO][4714] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.529 [INFO][4714] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.38.192/26 handle="k8s-pod-network.b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" host="172-234-31-125" Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.530 [INFO][4714] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.534 [INFO][4714] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.38.192/26 handle="k8s-pod-network.b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" host="172-234-31-125" Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.546 [INFO][4714] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.38.198/26] block=192.168.38.192/26 handle="k8s-pod-network.b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" host="172-234-31-125" Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.546 [INFO][4714] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.38.198/26] handle="k8s-pod-network.b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" host="172-234-31-125" Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.546 [INFO][4714] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:27.573260 containerd[1456]: 2025-07-07 00:06:27.546 [INFO][4714] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.38.198/26] IPv6=[] ContainerID="b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" HandleID="k8s-pod-network.b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:27.573734 containerd[1456]: 2025-07-07 00:06:27.550 [INFO][4690] cni-plugin/k8s.go 418: Populated endpoint ContainerID="b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" Namespace="kube-system" Pod="coredns-7c65d6cfc9-x9bb2" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"178fc278-1010-4c58-8d67-4a6ee0c1491b", ResourceVersion:"1008", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 5, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"", Pod:"coredns-7c65d6cfc9-x9bb2", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.38.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali602eb21fd0a", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:27.573734 containerd[1456]: 2025-07-07 00:06:27.551 [INFO][4690] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.38.198/32] ContainerID="b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" Namespace="kube-system" Pod="coredns-7c65d6cfc9-x9bb2" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:27.573734 containerd[1456]: 2025-07-07 00:06:27.551 [INFO][4690] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali602eb21fd0a ContainerID="b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" Namespace="kube-system" Pod="coredns-7c65d6cfc9-x9bb2" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:27.573734 containerd[1456]: 2025-07-07 00:06:27.560 [INFO][4690] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" Namespace="kube-system" Pod="coredns-7c65d6cfc9-x9bb2" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:27.573734 containerd[1456]: 2025-07-07 00:06:27.561 [INFO][4690] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" Namespace="kube-system" Pod="coredns-7c65d6cfc9-x9bb2" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"178fc278-1010-4c58-8d67-4a6ee0c1491b", ResourceVersion:"1008", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 5, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f", Pod:"coredns-7c65d6cfc9-x9bb2", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.38.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali602eb21fd0a", MAC:"a6:69:5c:6c:97:f3", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:27.573734 containerd[1456]: 2025-07-07 00:06:27.571 [INFO][4690] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f" Namespace="kube-system" Pod="coredns-7c65d6cfc9-x9bb2" WorkloadEndpoint="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:27.611473 containerd[1456]: time="2025-07-07T00:06:27.610319218Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:06:27.611473 containerd[1456]: time="2025-07-07T00:06:27.610385733Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:06:27.612072 containerd[1456]: time="2025-07-07T00:06:27.610408304Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:27.615828 containerd[1456]: time="2025-07-07T00:06:27.613125705Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:27.659516 systemd[1]: Started cri-containerd-b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f.scope - libcontainer container b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f. Jul 7 00:06:27.680029 systemd-networkd[1382]: cali2d7d678cb83: Link UP Jul 7 00:06:27.682318 systemd-networkd[1382]: cali2d7d678cb83: Gained carrier Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.425 [INFO][4695] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0 calico-apiserver-85b7446c9f- calico-apiserver 4741c1ac-aedc-48ec-8809-16303ada3c56 1009 0 2025-07-07 00:06:02 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:85b7446c9f projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s 172-234-31-125 calico-apiserver-85b7446c9f-8xc2r eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali2d7d678cb83 [] [] }} ContainerID="2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-8xc2r" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-" Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.425 [INFO][4695] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-8xc2r" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.493 [INFO][4716] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" HandleID="k8s-pod-network.2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.493 [INFO][4716] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" HandleID="k8s-pod-network.2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000365640), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"172-234-31-125", "pod":"calico-apiserver-85b7446c9f-8xc2r", "timestamp":"2025-07-07 00:06:27.493622356 +0000 UTC"}, Hostname:"172-234-31-125", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.493 [INFO][4716] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.546 [INFO][4716] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.547 [INFO][4716] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-234-31-125' Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.597 [INFO][4716] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" host="172-234-31-125" Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.606 [INFO][4716] ipam/ipam.go 394: Looking up existing affinities for host host="172-234-31-125" Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.632 [INFO][4716] ipam/ipam.go 511: Trying affinity for 192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.635 [INFO][4716] ipam/ipam.go 158: Attempting to load block cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.638 [INFO][4716] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.639 [INFO][4716] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.38.192/26 handle="k8s-pod-network.2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" host="172-234-31-125" Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.642 [INFO][4716] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39 Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.657 [INFO][4716] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.38.192/26 handle="k8s-pod-network.2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" host="172-234-31-125" Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.666 [INFO][4716] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.38.199/26] block=192.168.38.192/26 handle="k8s-pod-network.2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" host="172-234-31-125" Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.666 [INFO][4716] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.38.199/26] handle="k8s-pod-network.2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" host="172-234-31-125" Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.666 [INFO][4716] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:27.705458 containerd[1456]: 2025-07-07 00:06:27.666 [INFO][4716] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.38.199/26] IPv6=[] ContainerID="2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" HandleID="k8s-pod-network.2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:27.706094 containerd[1456]: 2025-07-07 00:06:27.673 [INFO][4695] cni-plugin/k8s.go 418: Populated endpoint ContainerID="2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-8xc2r" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0", GenerateName:"calico-apiserver-85b7446c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"4741c1ac-aedc-48ec-8809-16303ada3c56", ResourceVersion:"1009", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85b7446c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"", Pod:"calico-apiserver-85b7446c9f-8xc2r", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.38.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali2d7d678cb83", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:27.706094 containerd[1456]: 2025-07-07 00:06:27.674 [INFO][4695] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.38.199/32] ContainerID="2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-8xc2r" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:27.706094 containerd[1456]: 2025-07-07 00:06:27.674 [INFO][4695] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2d7d678cb83 ContainerID="2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-8xc2r" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:27.706094 containerd[1456]: 2025-07-07 00:06:27.679 [INFO][4695] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-8xc2r" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:27.706094 containerd[1456]: 2025-07-07 00:06:27.685 [INFO][4695] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-8xc2r" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0", GenerateName:"calico-apiserver-85b7446c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"4741c1ac-aedc-48ec-8809-16303ada3c56", ResourceVersion:"1009", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85b7446c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39", Pod:"calico-apiserver-85b7446c9f-8xc2r", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.38.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali2d7d678cb83", MAC:"b2:92:2a:78:d6:41", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:27.706094 containerd[1456]: 2025-07-07 00:06:27.695 [INFO][4695] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-8xc2r" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:27.794216 containerd[1456]: time="2025-07-07T00:06:27.794031212Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:06:27.795632 containerd[1456]: time="2025-07-07T00:06:27.795436765Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:06:27.795632 containerd[1456]: time="2025-07-07T00:06:27.795467048Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:27.795632 containerd[1456]: time="2025-07-07T00:06:27.795548484Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:27.866523 systemd[1]: Started cri-containerd-2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39.scope - libcontainer container 2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39. Jul 7 00:06:27.924773 containerd[1456]: time="2025-07-07T00:06:27.923478577Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-x9bb2,Uid:178fc278-1010-4c58-8d67-4a6ee0c1491b,Namespace:kube-system,Attempt:1,} returns sandbox id \"b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f\"" Jul 7 00:06:27.924963 kubelet[2515]: E0707 00:06:27.924144 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:27.940392 containerd[1456]: time="2025-07-07T00:06:27.939016408Z" level=info msg="CreateContainer within sandbox \"b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jul 7 00:06:27.952393 containerd[1456]: time="2025-07-07T00:06:27.951891812Z" level=info msg="CreateContainer within sandbox \"b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"6eb33380ea2888efab77e986c85787a5e84a65331195c9c39a34587809f7a7a2\"" Jul 7 00:06:27.956003 containerd[1456]: time="2025-07-07T00:06:27.955973474Z" level=info msg="StartContainer for \"6eb33380ea2888efab77e986c85787a5e84a65331195c9c39a34587809f7a7a2\"" Jul 7 00:06:28.010524 systemd[1]: Started cri-containerd-6eb33380ea2888efab77e986c85787a5e84a65331195c9c39a34587809f7a7a2.scope - libcontainer container 6eb33380ea2888efab77e986c85787a5e84a65331195c9c39a34587809f7a7a2. Jul 7 00:06:28.071733 containerd[1456]: time="2025-07-07T00:06:28.069247271Z" level=info msg="StartContainer for \"6eb33380ea2888efab77e986c85787a5e84a65331195c9c39a34587809f7a7a2\" returns successfully" Jul 7 00:06:28.130654 systemd-networkd[1382]: cali4a975bd8445: Gained IPv6LL Jul 7 00:06:28.149384 containerd[1456]: time="2025-07-07T00:06:28.149217670Z" level=info msg="StopPodSandbox for \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\"" Jul 7 00:06:28.260047 systemd-networkd[1382]: calif76aa88d3ce: Gained IPv6LL Jul 7 00:06:28.345194 containerd[1456]: time="2025-07-07T00:06:28.345156967Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85b7446c9f-8xc2r,Uid:4741c1ac-aedc-48ec-8809-16303ada3c56,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39\"" Jul 7 00:06:28.409706 kubelet[2515]: E0707 00:06:28.409676 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:28.428251 containerd[1456]: 2025-07-07 00:06:28.285 [INFO][4874] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Jul 7 00:06:28.428251 containerd[1456]: 2025-07-07 00:06:28.285 [INFO][4874] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" iface="eth0" netns="/var/run/netns/cni-78553802-1f7f-6e15-3725-bad7c15c61b1" Jul 7 00:06:28.428251 containerd[1456]: 2025-07-07 00:06:28.285 [INFO][4874] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" iface="eth0" netns="/var/run/netns/cni-78553802-1f7f-6e15-3725-bad7c15c61b1" Jul 7 00:06:28.428251 containerd[1456]: 2025-07-07 00:06:28.285 [INFO][4874] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" iface="eth0" netns="/var/run/netns/cni-78553802-1f7f-6e15-3725-bad7c15c61b1" Jul 7 00:06:28.428251 containerd[1456]: 2025-07-07 00:06:28.285 [INFO][4874] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Jul 7 00:06:28.428251 containerd[1456]: 2025-07-07 00:06:28.286 [INFO][4874] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Jul 7 00:06:28.428251 containerd[1456]: 2025-07-07 00:06:28.388 [INFO][4883] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" HandleID="k8s-pod-network.09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:28.428251 containerd[1456]: 2025-07-07 00:06:28.388 [INFO][4883] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:28.428251 containerd[1456]: 2025-07-07 00:06:28.388 [INFO][4883] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:28.428251 containerd[1456]: 2025-07-07 00:06:28.402 [WARNING][4883] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" HandleID="k8s-pod-network.09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:28.428251 containerd[1456]: 2025-07-07 00:06:28.402 [INFO][4883] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" HandleID="k8s-pod-network.09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:28.428251 containerd[1456]: 2025-07-07 00:06:28.404 [INFO][4883] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:28.428251 containerd[1456]: 2025-07-07 00:06:28.413 [INFO][4874] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Jul 7 00:06:28.430675 containerd[1456]: time="2025-07-07T00:06:28.430002565Z" level=info msg="TearDown network for sandbox \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\" successfully" Jul 7 00:06:28.430675 containerd[1456]: time="2025-07-07T00:06:28.430029227Z" level=info msg="StopPodSandbox for \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\" returns successfully" Jul 7 00:06:28.431013 containerd[1456]: time="2025-07-07T00:06:28.430656460Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85b7446c9f-78mbn,Uid:15e90dc0-720a-4ae9-aac2-01720a32d5a8,Namespace:calico-apiserver,Attempt:1,}" Jul 7 00:06:28.438595 systemd[1]: run-netns-cni\x2d78553802\x2d1f7f\x2d6e15\x2d3725\x2dbad7c15c61b1.mount: Deactivated successfully. Jul 7 00:06:28.451159 kubelet[2515]: I0707 00:06:28.450203 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-x9bb2" podStartSLOduration=35.450184436 podStartE2EDuration="35.450184436s" podCreationTimestamp="2025-07-07 00:05:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-07 00:06:28.426947052 +0000 UTC m=+40.394817365" watchObservedRunningTime="2025-07-07 00:06:28.450184436 +0000 UTC m=+40.418054749" Jul 7 00:06:28.689082 systemd-networkd[1382]: cali5cdff996823: Link UP Jul 7 00:06:28.692614 systemd-networkd[1382]: cali5cdff996823: Gained carrier Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.534 [INFO][4900] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0 calico-apiserver-85b7446c9f- calico-apiserver 15e90dc0-720a-4ae9-aac2-01720a32d5a8 1025 0 2025-07-07 00:06:02 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:85b7446c9f projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s 172-234-31-125 calico-apiserver-85b7446c9f-78mbn eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali5cdff996823 [] [] }} ContainerID="ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-78mbn" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-" Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.534 [INFO][4900] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-78mbn" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.615 [INFO][4912] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" HandleID="k8s-pod-network.ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.616 [INFO][4912] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" HandleID="k8s-pod-network.ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5df0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"172-234-31-125", "pod":"calico-apiserver-85b7446c9f-78mbn", "timestamp":"2025-07-07 00:06:28.61511383 +0000 UTC"}, Hostname:"172-234-31-125", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.616 [INFO][4912] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.616 [INFO][4912] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.616 [INFO][4912] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172-234-31-125' Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.630 [INFO][4912] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" host="172-234-31-125" Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.636 [INFO][4912] ipam/ipam.go 394: Looking up existing affinities for host host="172-234-31-125" Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.650 [INFO][4912] ipam/ipam.go 511: Trying affinity for 192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.653 [INFO][4912] ipam/ipam.go 158: Attempting to load block cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.656 [INFO][4912] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.38.192/26 host="172-234-31-125" Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.656 [INFO][4912] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.38.192/26 handle="k8s-pod-network.ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" host="172-234-31-125" Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.658 [INFO][4912] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502 Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.662 [INFO][4912] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.38.192/26 handle="k8s-pod-network.ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" host="172-234-31-125" Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.670 [INFO][4912] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.38.200/26] block=192.168.38.192/26 handle="k8s-pod-network.ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" host="172-234-31-125" Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.670 [INFO][4912] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.38.200/26] handle="k8s-pod-network.ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" host="172-234-31-125" Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.670 [INFO][4912] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:28.727804 containerd[1456]: 2025-07-07 00:06:28.670 [INFO][4912] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.38.200/26] IPv6=[] ContainerID="ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" HandleID="k8s-pod-network.ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:28.728429 containerd[1456]: 2025-07-07 00:06:28.678 [INFO][4900] cni-plugin/k8s.go 418: Populated endpoint ContainerID="ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-78mbn" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0", GenerateName:"calico-apiserver-85b7446c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"15e90dc0-720a-4ae9-aac2-01720a32d5a8", ResourceVersion:"1025", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85b7446c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"", Pod:"calico-apiserver-85b7446c9f-78mbn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.38.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali5cdff996823", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:28.728429 containerd[1456]: 2025-07-07 00:06:28.680 [INFO][4900] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.38.200/32] ContainerID="ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-78mbn" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:28.728429 containerd[1456]: 2025-07-07 00:06:28.680 [INFO][4900] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5cdff996823 ContainerID="ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-78mbn" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:28.728429 containerd[1456]: 2025-07-07 00:06:28.696 [INFO][4900] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-78mbn" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:28.728429 containerd[1456]: 2025-07-07 00:06:28.700 [INFO][4900] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-78mbn" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0", GenerateName:"calico-apiserver-85b7446c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"15e90dc0-720a-4ae9-aac2-01720a32d5a8", ResourceVersion:"1025", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85b7446c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502", Pod:"calico-apiserver-85b7446c9f-78mbn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.38.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali5cdff996823", MAC:"76:89:a3:90:b0:f8", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:28.728429 containerd[1456]: 2025-07-07 00:06:28.720 [INFO][4900] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502" Namespace="calico-apiserver" Pod="calico-apiserver-85b7446c9f-78mbn" WorkloadEndpoint="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:28.770533 systemd-networkd[1382]: cali94d73bf1cb2: Gained IPv6LL Jul 7 00:06:28.779968 containerd[1456]: time="2025-07-07T00:06:28.778753275Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 7 00:06:28.779968 containerd[1456]: time="2025-07-07T00:06:28.778802839Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 7 00:06:28.779968 containerd[1456]: time="2025-07-07T00:06:28.778817670Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:28.779968 containerd[1456]: time="2025-07-07T00:06:28.778902376Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 7 00:06:28.833602 systemd[1]: Started cri-containerd-ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502.scope - libcontainer container ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502. Jul 7 00:06:28.834850 systemd-networkd[1382]: cali2d7d678cb83: Gained IPv6LL Jul 7 00:06:28.952658 containerd[1456]: time="2025-07-07T00:06:28.951773222Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-85b7446c9f-78mbn,Uid:15e90dc0-720a-4ae9-aac2-01720a32d5a8,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502\"" Jul 7 00:06:29.219826 systemd-networkd[1382]: cali602eb21fd0a: Gained IPv6LL Jul 7 00:06:29.363335 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount923673203.mount: Deactivated successfully. Jul 7 00:06:29.430687 kubelet[2515]: E0707 00:06:29.429824 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:29.795598 systemd-networkd[1382]: cali5cdff996823: Gained IPv6LL Jul 7 00:06:29.889560 containerd[1456]: time="2025-07-07T00:06:29.888602631Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:29.890864 containerd[1456]: time="2025-07-07T00:06:29.890691777Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.2: active requests=0, bytes read=66352308" Jul 7 00:06:29.892060 containerd[1456]: time="2025-07-07T00:06:29.891720614Z" level=info msg="ImageCreate event name:\"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:29.894616 containerd[1456]: time="2025-07-07T00:06:29.894562498Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:29.896053 containerd[1456]: time="2025-07-07T00:06:29.896007103Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" with image id \"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305\", size \"66352154\" in 3.079128959s" Jul 7 00:06:29.896234 containerd[1456]: time="2025-07-07T00:06:29.896217856Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" returns image reference \"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\"" Jul 7 00:06:29.898664 containerd[1456]: time="2025-07-07T00:06:29.898619922Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\"" Jul 7 00:06:29.900133 containerd[1456]: time="2025-07-07T00:06:29.900096069Z" level=info msg="CreateContainer within sandbox \"5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Jul 7 00:06:29.919504 containerd[1456]: time="2025-07-07T00:06:29.919469928Z" level=info msg="CreateContainer within sandbox \"5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"d86f065255af24b7f9f17a1e93168ee1899ea13e2b54bfce6b64a44abb0a4870\"" Jul 7 00:06:29.920689 containerd[1456]: time="2025-07-07T00:06:29.920320433Z" level=info msg="StartContainer for \"d86f065255af24b7f9f17a1e93168ee1899ea13e2b54bfce6b64a44abb0a4870\"" Jul 7 00:06:30.004488 systemd[1]: Started cri-containerd-d86f065255af24b7f9f17a1e93168ee1899ea13e2b54bfce6b64a44abb0a4870.scope - libcontainer container d86f065255af24b7f9f17a1e93168ee1899ea13e2b54bfce6b64a44abb0a4870. Jul 7 00:06:30.070425 containerd[1456]: time="2025-07-07T00:06:30.069738416Z" level=info msg="StartContainer for \"d86f065255af24b7f9f17a1e93168ee1899ea13e2b54bfce6b64a44abb0a4870\" returns successfully" Jul 7 00:06:30.433786 kubelet[2515]: E0707 00:06:30.433747 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:06:30.452407 kubelet[2515]: I0707 00:06:30.451644 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-58fd7646b9-d799z" podStartSLOduration=23.364649369 podStartE2EDuration="26.451614223s" podCreationTimestamp="2025-07-07 00:06:04 +0000 UTC" firstStartedPulling="2025-07-07 00:06:26.810763751 +0000 UTC m=+38.778634064" lastFinishedPulling="2025-07-07 00:06:29.897728605 +0000 UTC m=+41.865598918" observedRunningTime="2025-07-07 00:06:30.450450822 +0000 UTC m=+42.418321135" watchObservedRunningTime="2025-07-07 00:06:30.451614223 +0000 UTC m=+42.419484536" Jul 7 00:06:30.474827 systemd[1]: run-containerd-runc-k8s.io-d86f065255af24b7f9f17a1e93168ee1899ea13e2b54bfce6b64a44abb0a4870-runc.qq619b.mount: Deactivated successfully. Jul 7 00:06:32.013228 containerd[1456]: time="2025-07-07T00:06:32.013184685Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:32.015174 containerd[1456]: time="2025-07-07T00:06:32.015141600Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.2: active requests=0, bytes read=51276688" Jul 7 00:06:32.017193 containerd[1456]: time="2025-07-07T00:06:32.016199556Z" level=info msg="ImageCreate event name:\"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:32.022133 containerd[1456]: time="2025-07-07T00:06:32.022103881Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:32.023120 containerd[1456]: time="2025-07-07T00:06:32.023098395Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" with image id \"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce\", size \"52769359\" in 2.124254859s" Jul 7 00:06:32.023408 containerd[1456]: time="2025-07-07T00:06:32.023392051Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" returns image reference \"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\"" Jul 7 00:06:32.027678 containerd[1456]: time="2025-07-07T00:06:32.027661630Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\"" Jul 7 00:06:32.043722 containerd[1456]: time="2025-07-07T00:06:32.043697287Z" level=info msg="CreateContainer within sandbox \"4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Jul 7 00:06:32.059141 containerd[1456]: time="2025-07-07T00:06:32.059113822Z" level=info msg="CreateContainer within sandbox \"4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"9508c81aa97aa2789f7a6f45f6bc7bc23e419b3162f2cdc33249cf514d6aba85\"" Jul 7 00:06:32.060094 containerd[1456]: time="2025-07-07T00:06:32.060067152Z" level=info msg="StartContainer for \"9508c81aa97aa2789f7a6f45f6bc7bc23e419b3162f2cdc33249cf514d6aba85\"" Jul 7 00:06:32.114519 systemd[1]: Started cri-containerd-9508c81aa97aa2789f7a6f45f6bc7bc23e419b3162f2cdc33249cf514d6aba85.scope - libcontainer container 9508c81aa97aa2789f7a6f45f6bc7bc23e419b3162f2cdc33249cf514d6aba85. Jul 7 00:06:32.214482 containerd[1456]: time="2025-07-07T00:06:32.214422819Z" level=info msg="StartContainer for \"9508c81aa97aa2789f7a6f45f6bc7bc23e419b3162f2cdc33249cf514d6aba85\" returns successfully" Jul 7 00:06:32.456808 kubelet[2515]: I0707 00:06:32.456058 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-94599b4db-wbn2z" podStartSLOduration=22.541748608 podStartE2EDuration="27.456039693s" podCreationTimestamp="2025-07-07 00:06:05 +0000 UTC" firstStartedPulling="2025-07-07 00:06:27.110274728 +0000 UTC m=+39.078145051" lastFinishedPulling="2025-07-07 00:06:32.024565823 +0000 UTC m=+43.992436136" observedRunningTime="2025-07-07 00:06:32.453650646 +0000 UTC m=+44.421520959" watchObservedRunningTime="2025-07-07 00:06:32.456039693 +0000 UTC m=+44.423910006" Jul 7 00:06:32.763033 containerd[1456]: time="2025-07-07T00:06:32.762893328Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:32.765153 containerd[1456]: time="2025-07-07T00:06:32.765101726Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.2: active requests=0, bytes read=8759190" Jul 7 00:06:32.767559 containerd[1456]: time="2025-07-07T00:06:32.766977476Z" level=info msg="ImageCreate event name:\"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:32.772291 containerd[1456]: time="2025-07-07T00:06:32.772068748Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:32.773013 containerd[1456]: time="2025-07-07T00:06:32.772977837Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.2\" with image id \"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0\", size \"10251893\" in 745.216613ms" Jul 7 00:06:32.773051 containerd[1456]: time="2025-07-07T00:06:32.773015189Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\" returns image reference \"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\"" Jul 7 00:06:32.776436 containerd[1456]: time="2025-07-07T00:06:32.776089303Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Jul 7 00:06:32.782595 containerd[1456]: time="2025-07-07T00:06:32.780141820Z" level=info msg="CreateContainer within sandbox \"9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Jul 7 00:06:32.798511 containerd[1456]: time="2025-07-07T00:06:32.798481031Z" level=info msg="CreateContainer within sandbox \"9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"b492c7d5e265c4c20d8d59fa9d456cf1dcf0e4f3cbb7484debb2cf3c4ae76f88\"" Jul 7 00:06:32.798983 containerd[1456]: time="2025-07-07T00:06:32.798964717Z" level=info msg="StartContainer for \"b492c7d5e265c4c20d8d59fa9d456cf1dcf0e4f3cbb7484debb2cf3c4ae76f88\"" Jul 7 00:06:32.840879 systemd[1]: Started cri-containerd-b492c7d5e265c4c20d8d59fa9d456cf1dcf0e4f3cbb7484debb2cf3c4ae76f88.scope - libcontainer container b492c7d5e265c4c20d8d59fa9d456cf1dcf0e4f3cbb7484debb2cf3c4ae76f88. Jul 7 00:06:32.890093 containerd[1456]: time="2025-07-07T00:06:32.890039879Z" level=info msg="StartContainer for \"b492c7d5e265c4c20d8d59fa9d456cf1dcf0e4f3cbb7484debb2cf3c4ae76f88\" returns successfully" Jul 7 00:06:33.447599 kubelet[2515]: I0707 00:06:33.447566 2515 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 7 00:06:34.810390 containerd[1456]: time="2025-07-07T00:06:34.808644351Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:34.811093 containerd[1456]: time="2025-07-07T00:06:34.811050864Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.2: active requests=0, bytes read=47317977" Jul 7 00:06:34.811641 containerd[1456]: time="2025-07-07T00:06:34.811612840Z" level=info msg="ImageCreate event name:\"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:34.814710 containerd[1456]: time="2025-07-07T00:06:34.814677924Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:34.815388 containerd[1456]: time="2025-07-07T00:06:34.815338365Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" with image id \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\", size \"48810696\" in 2.038820308s" Jul 7 00:06:34.815388 containerd[1456]: time="2025-07-07T00:06:34.815382457Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\"" Jul 7 00:06:34.818441 containerd[1456]: time="2025-07-07T00:06:34.818411880Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Jul 7 00:06:34.820096 containerd[1456]: time="2025-07-07T00:06:34.820065677Z" level=info msg="CreateContainer within sandbox \"2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Jul 7 00:06:34.844625 containerd[1456]: time="2025-07-07T00:06:34.844569268Z" level=info msg="CreateContainer within sandbox \"2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"f40da2aaa0f94ea6d2ba28cd651f169086a1adaa629559d3418a09b4da239cf3\"" Jul 7 00:06:34.847269 containerd[1456]: time="2025-07-07T00:06:34.847235073Z" level=info msg="StartContainer for \"f40da2aaa0f94ea6d2ba28cd651f169086a1adaa629559d3418a09b4da239cf3\"" Jul 7 00:06:34.893703 systemd[1]: run-containerd-runc-k8s.io-f40da2aaa0f94ea6d2ba28cd651f169086a1adaa629559d3418a09b4da239cf3-runc.L0OryU.mount: Deactivated successfully. Jul 7 00:06:34.909500 systemd[1]: Started cri-containerd-f40da2aaa0f94ea6d2ba28cd651f169086a1adaa629559d3418a09b4da239cf3.scope - libcontainer container f40da2aaa0f94ea6d2ba28cd651f169086a1adaa629559d3418a09b4da239cf3. Jul 7 00:06:34.989187 containerd[1456]: time="2025-07-07T00:06:34.989046132Z" level=info msg="StartContainer for \"f40da2aaa0f94ea6d2ba28cd651f169086a1adaa629559d3418a09b4da239cf3\" returns successfully" Jul 7 00:06:35.033732 containerd[1456]: time="2025-07-07T00:06:35.033671841Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:35.036129 containerd[1456]: time="2025-07-07T00:06:35.034747458Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.2: active requests=0, bytes read=77" Jul 7 00:06:35.039285 containerd[1456]: time="2025-07-07T00:06:35.039243185Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" with image id \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\", size \"48810696\" in 220.798964ms" Jul 7 00:06:35.039433 containerd[1456]: time="2025-07-07T00:06:35.039404453Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\"" Jul 7 00:06:35.041451 containerd[1456]: time="2025-07-07T00:06:35.041418391Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\"" Jul 7 00:06:35.041718 containerd[1456]: time="2025-07-07T00:06:35.041686713Z" level=info msg="CreateContainer within sandbox \"ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Jul 7 00:06:35.059734 containerd[1456]: time="2025-07-07T00:06:35.059687505Z" level=info msg="CreateContainer within sandbox \"ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"b4c8e1d438b3ec1ba3fc7e4a9f99ca2ec44623253ea3cf08e2d9ad5cc709d3f8\"" Jul 7 00:06:35.060293 containerd[1456]: time="2025-07-07T00:06:35.060266361Z" level=info msg="StartContainer for \"b4c8e1d438b3ec1ba3fc7e4a9f99ca2ec44623253ea3cf08e2d9ad5cc709d3f8\"" Jul 7 00:06:35.106554 systemd[1]: Started cri-containerd-b4c8e1d438b3ec1ba3fc7e4a9f99ca2ec44623253ea3cf08e2d9ad5cc709d3f8.scope - libcontainer container b4c8e1d438b3ec1ba3fc7e4a9f99ca2ec44623253ea3cf08e2d9ad5cc709d3f8. Jul 7 00:06:35.211679 containerd[1456]: time="2025-07-07T00:06:35.211633839Z" level=info msg="StartContainer for \"b4c8e1d438b3ec1ba3fc7e4a9f99ca2ec44623253ea3cf08e2d9ad5cc709d3f8\" returns successfully" Jul 7 00:06:35.498197 kubelet[2515]: I0707 00:06:35.498133 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-85b7446c9f-78mbn" podStartSLOduration=27.413456109 podStartE2EDuration="33.498119641s" podCreationTimestamp="2025-07-07 00:06:02 +0000 UTC" firstStartedPulling="2025-07-07 00:06:28.955667282 +0000 UTC m=+40.923537595" lastFinishedPulling="2025-07-07 00:06:35.040330814 +0000 UTC m=+47.008201127" observedRunningTime="2025-07-07 00:06:35.483584331 +0000 UTC m=+47.451454644" watchObservedRunningTime="2025-07-07 00:06:35.498119641 +0000 UTC m=+47.465989964" Jul 7 00:06:35.499640 kubelet[2515]: I0707 00:06:35.499413 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-85b7446c9f-8xc2r" podStartSLOduration=27.039087501 podStartE2EDuration="33.499406537s" podCreationTimestamp="2025-07-07 00:06:02 +0000 UTC" firstStartedPulling="2025-07-07 00:06:28.3567374 +0000 UTC m=+40.324607713" lastFinishedPulling="2025-07-07 00:06:34.817056436 +0000 UTC m=+46.784926749" observedRunningTime="2025-07-07 00:06:35.496974171 +0000 UTC m=+47.464844484" watchObservedRunningTime="2025-07-07 00:06:35.499406537 +0000 UTC m=+47.467276850" Jul 7 00:06:36.155473 containerd[1456]: time="2025-07-07T00:06:36.155419377Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:36.157402 containerd[1456]: time="2025-07-07T00:06:36.156551194Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2: active requests=0, bytes read=14703784" Jul 7 00:06:36.157402 containerd[1456]: time="2025-07-07T00:06:36.156982392Z" level=info msg="ImageCreate event name:\"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:36.159633 containerd[1456]: time="2025-07-07T00:06:36.159492185Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 7 00:06:36.160120 containerd[1456]: time="2025-07-07T00:06:36.160048938Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" with image id \"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f\", size \"16196439\" in 1.118552673s" Jul 7 00:06:36.160120 containerd[1456]: time="2025-07-07T00:06:36.160074939Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" returns image reference \"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\"" Jul 7 00:06:36.164507 containerd[1456]: time="2025-07-07T00:06:36.164452639Z" level=info msg="CreateContainer within sandbox \"9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Jul 7 00:06:36.181430 containerd[1456]: time="2025-07-07T00:06:36.181298123Z" level=info msg="CreateContainer within sandbox \"9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"5fc7bb924f61498bc7dbaca23a522ed3a3bdd3f5ff7459b88bcdc9fa689bbe5e\"" Jul 7 00:06:36.182211 containerd[1456]: time="2025-07-07T00:06:36.182091706Z" level=info msg="StartContainer for \"5fc7bb924f61498bc7dbaca23a522ed3a3bdd3f5ff7459b88bcdc9fa689bbe5e\"" Jul 7 00:06:36.233686 systemd[1]: Started cri-containerd-5fc7bb924f61498bc7dbaca23a522ed3a3bdd3f5ff7459b88bcdc9fa689bbe5e.scope - libcontainer container 5fc7bb924f61498bc7dbaca23a522ed3a3bdd3f5ff7459b88bcdc9fa689bbe5e. Jul 7 00:06:36.283392 containerd[1456]: time="2025-07-07T00:06:36.283271805Z" level=info msg="StartContainer for \"5fc7bb924f61498bc7dbaca23a522ed3a3bdd3f5ff7459b88bcdc9fa689bbe5e\" returns successfully" Jul 7 00:06:36.469403 kubelet[2515]: I0707 00:06:36.467911 2515 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 7 00:06:36.881478 kubelet[2515]: I0707 00:06:36.877258 2515 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-9knjc" podStartSLOduration=22.850737103 podStartE2EDuration="31.877241951s" podCreationTimestamp="2025-07-07 00:06:05 +0000 UTC" firstStartedPulling="2025-07-07 00:06:27.134710937 +0000 UTC m=+39.102581250" lastFinishedPulling="2025-07-07 00:06:36.161215785 +0000 UTC m=+48.129086098" observedRunningTime="2025-07-07 00:06:36.479533803 +0000 UTC m=+48.447404116" watchObservedRunningTime="2025-07-07 00:06:36.877241951 +0000 UTC m=+48.845112264" Jul 7 00:06:37.250669 kubelet[2515]: I0707 00:06:37.250591 2515 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Jul 7 00:06:37.250669 kubelet[2515]: I0707 00:06:37.250622 2515 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Jul 7 00:06:42.655840 kubelet[2515]: I0707 00:06:42.655642 2515 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 7 00:06:48.130388 containerd[1456]: time="2025-07-07T00:06:48.130031320Z" level=info msg="StopPodSandbox for \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\"" Jul 7 00:06:48.255486 containerd[1456]: 2025-07-07 00:06:48.187 [WARNING][5333] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"06260b4f-49a9-4902-b2df-d9734428c436", ResourceVersion:"1056", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece", Pod:"goldmane-58fd7646b9-d799z", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.38.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali4a975bd8445", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:48.255486 containerd[1456]: 2025-07-07 00:06:48.190 [INFO][5333] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Jul 7 00:06:48.255486 containerd[1456]: 2025-07-07 00:06:48.190 [INFO][5333] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" iface="eth0" netns="" Jul 7 00:06:48.255486 containerd[1456]: 2025-07-07 00:06:48.190 [INFO][5333] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Jul 7 00:06:48.255486 containerd[1456]: 2025-07-07 00:06:48.190 [INFO][5333] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Jul 7 00:06:48.255486 containerd[1456]: 2025-07-07 00:06:48.235 [INFO][5342] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" HandleID="k8s-pod-network.3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Workload="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:48.255486 containerd[1456]: 2025-07-07 00:06:48.237 [INFO][5342] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:48.255486 containerd[1456]: 2025-07-07 00:06:48.237 [INFO][5342] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:48.255486 containerd[1456]: 2025-07-07 00:06:48.246 [WARNING][5342] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" HandleID="k8s-pod-network.3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Workload="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:48.255486 containerd[1456]: 2025-07-07 00:06:48.246 [INFO][5342] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" HandleID="k8s-pod-network.3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Workload="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:48.255486 containerd[1456]: 2025-07-07 00:06:48.248 [INFO][5342] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:48.255486 containerd[1456]: 2025-07-07 00:06:48.251 [INFO][5333] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Jul 7 00:06:48.255962 containerd[1456]: time="2025-07-07T00:06:48.255536480Z" level=info msg="TearDown network for sandbox \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\" successfully" Jul 7 00:06:48.255962 containerd[1456]: time="2025-07-07T00:06:48.255566021Z" level=info msg="StopPodSandbox for \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\" returns successfully" Jul 7 00:06:48.256343 containerd[1456]: time="2025-07-07T00:06:48.256310025Z" level=info msg="RemovePodSandbox for \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\"" Jul 7 00:06:48.256401 containerd[1456]: time="2025-07-07T00:06:48.256340675Z" level=info msg="Forcibly stopping sandbox \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\"" Jul 7 00:06:48.378210 containerd[1456]: 2025-07-07 00:06:48.329 [WARNING][5356] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"06260b4f-49a9-4902-b2df-d9734428c436", ResourceVersion:"1056", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"5c8e4570b66e869c4c3ee82e52e0de35a83a2a13a079f87e55f27e0f66687ece", Pod:"goldmane-58fd7646b9-d799z", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.38.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali4a975bd8445", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:48.378210 containerd[1456]: 2025-07-07 00:06:48.329 [INFO][5356] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Jul 7 00:06:48.378210 containerd[1456]: 2025-07-07 00:06:48.329 [INFO][5356] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" iface="eth0" netns="" Jul 7 00:06:48.378210 containerd[1456]: 2025-07-07 00:06:48.329 [INFO][5356] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Jul 7 00:06:48.378210 containerd[1456]: 2025-07-07 00:06:48.329 [INFO][5356] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Jul 7 00:06:48.378210 containerd[1456]: 2025-07-07 00:06:48.364 [INFO][5362] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" HandleID="k8s-pod-network.3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Workload="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:48.378210 containerd[1456]: 2025-07-07 00:06:48.364 [INFO][5362] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:48.378210 containerd[1456]: 2025-07-07 00:06:48.364 [INFO][5362] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:48.378210 containerd[1456]: 2025-07-07 00:06:48.369 [WARNING][5362] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" HandleID="k8s-pod-network.3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Workload="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:48.378210 containerd[1456]: 2025-07-07 00:06:48.369 [INFO][5362] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" HandleID="k8s-pod-network.3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Workload="172--234--31--125-k8s-goldmane--58fd7646b9--d799z-eth0" Jul 7 00:06:48.378210 containerd[1456]: 2025-07-07 00:06:48.371 [INFO][5362] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:48.378210 containerd[1456]: 2025-07-07 00:06:48.374 [INFO][5356] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f" Jul 7 00:06:48.378210 containerd[1456]: time="2025-07-07T00:06:48.376727899Z" level=info msg="TearDown network for sandbox \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\" successfully" Jul 7 00:06:48.389731 containerd[1456]: time="2025-07-07T00:06:48.389464188Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jul 7 00:06:48.389731 containerd[1456]: time="2025-07-07T00:06:48.389544439Z" level=info msg="RemovePodSandbox \"3a8d5db6af98a5fd280b2d5367f2aada12a3db18891c8f8950c1414cbd7bce1f\" returns successfully" Jul 7 00:06:48.390216 containerd[1456]: time="2025-07-07T00:06:48.390085170Z" level=info msg="StopPodSandbox for \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\"" Jul 7 00:06:48.470514 containerd[1456]: 2025-07-07 00:06:48.432 [WARNING][5376] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" WorkloadEndpoint="172--234--31--125-k8s-whisker--59c44fdcd5--rlcnr-eth0" Jul 7 00:06:48.470514 containerd[1456]: 2025-07-07 00:06:48.432 [INFO][5376] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Jul 7 00:06:48.470514 containerd[1456]: 2025-07-07 00:06:48.432 [INFO][5376] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" iface="eth0" netns="" Jul 7 00:06:48.470514 containerd[1456]: 2025-07-07 00:06:48.432 [INFO][5376] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Jul 7 00:06:48.470514 containerd[1456]: 2025-07-07 00:06:48.432 [INFO][5376] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Jul 7 00:06:48.470514 containerd[1456]: 2025-07-07 00:06:48.454 [INFO][5384] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" HandleID="k8s-pod-network.71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Workload="172--234--31--125-k8s-whisker--59c44fdcd5--rlcnr-eth0" Jul 7 00:06:48.470514 containerd[1456]: 2025-07-07 00:06:48.454 [INFO][5384] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:48.470514 containerd[1456]: 2025-07-07 00:06:48.454 [INFO][5384] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:48.470514 containerd[1456]: 2025-07-07 00:06:48.460 [WARNING][5384] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" HandleID="k8s-pod-network.71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Workload="172--234--31--125-k8s-whisker--59c44fdcd5--rlcnr-eth0" Jul 7 00:06:48.470514 containerd[1456]: 2025-07-07 00:06:48.460 [INFO][5384] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" HandleID="k8s-pod-network.71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Workload="172--234--31--125-k8s-whisker--59c44fdcd5--rlcnr-eth0" Jul 7 00:06:48.470514 containerd[1456]: 2025-07-07 00:06:48.462 [INFO][5384] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:48.470514 containerd[1456]: 2025-07-07 00:06:48.466 [INFO][5376] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Jul 7 00:06:48.471183 containerd[1456]: time="2025-07-07T00:06:48.471134447Z" level=info msg="TearDown network for sandbox \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\" successfully" Jul 7 00:06:48.471324 containerd[1456]: time="2025-07-07T00:06:48.471248899Z" level=info msg="StopPodSandbox for \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\" returns successfully" Jul 7 00:06:48.472628 containerd[1456]: time="2025-07-07T00:06:48.472605864Z" level=info msg="RemovePodSandbox for \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\"" Jul 7 00:06:48.472685 containerd[1456]: time="2025-07-07T00:06:48.472636875Z" level=info msg="Forcibly stopping sandbox \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\"" Jul 7 00:06:48.580534 containerd[1456]: 2025-07-07 00:06:48.529 [WARNING][5398] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" WorkloadEndpoint="172--234--31--125-k8s-whisker--59c44fdcd5--rlcnr-eth0" Jul 7 00:06:48.580534 containerd[1456]: 2025-07-07 00:06:48.530 [INFO][5398] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Jul 7 00:06:48.580534 containerd[1456]: 2025-07-07 00:06:48.530 [INFO][5398] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" iface="eth0" netns="" Jul 7 00:06:48.580534 containerd[1456]: 2025-07-07 00:06:48.530 [INFO][5398] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Jul 7 00:06:48.580534 containerd[1456]: 2025-07-07 00:06:48.530 [INFO][5398] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Jul 7 00:06:48.580534 containerd[1456]: 2025-07-07 00:06:48.562 [INFO][5406] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" HandleID="k8s-pod-network.71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Workload="172--234--31--125-k8s-whisker--59c44fdcd5--rlcnr-eth0" Jul 7 00:06:48.580534 containerd[1456]: 2025-07-07 00:06:48.563 [INFO][5406] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:48.580534 containerd[1456]: 2025-07-07 00:06:48.564 [INFO][5406] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:48.580534 containerd[1456]: 2025-07-07 00:06:48.572 [WARNING][5406] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" HandleID="k8s-pod-network.71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Workload="172--234--31--125-k8s-whisker--59c44fdcd5--rlcnr-eth0" Jul 7 00:06:48.580534 containerd[1456]: 2025-07-07 00:06:48.572 [INFO][5406] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" HandleID="k8s-pod-network.71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Workload="172--234--31--125-k8s-whisker--59c44fdcd5--rlcnr-eth0" Jul 7 00:06:48.580534 containerd[1456]: 2025-07-07 00:06:48.573 [INFO][5406] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:48.580534 containerd[1456]: 2025-07-07 00:06:48.576 [INFO][5398] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980" Jul 7 00:06:48.580534 containerd[1456]: time="2025-07-07T00:06:48.580426403Z" level=info msg="TearDown network for sandbox \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\" successfully" Jul 7 00:06:48.585858 containerd[1456]: time="2025-07-07T00:06:48.585625031Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jul 7 00:06:48.585858 containerd[1456]: time="2025-07-07T00:06:48.585744143Z" level=info msg="RemovePodSandbox \"71bcf8fa2c06f787e3a2e6c3a1ef4ac90f4a67c807641d9035f3af63734d5980\" returns successfully" Jul 7 00:06:48.586355 containerd[1456]: time="2025-07-07T00:06:48.586316064Z" level=info msg="StopPodSandbox for \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\"" Jul 7 00:06:48.661550 containerd[1456]: 2025-07-07 00:06:48.623 [WARNING][5421] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0", GenerateName:"calico-kube-controllers-94599b4db-", Namespace:"calico-system", SelfLink:"", UID:"7f03664a-b5c1-41d9-b19a-954bfd2a5ea7", ResourceVersion:"1077", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"94599b4db", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b", Pod:"calico-kube-controllers-94599b4db-wbn2z", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.38.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali94d73bf1cb2", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:48.661550 containerd[1456]: 2025-07-07 00:06:48.624 [INFO][5421] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Jul 7 00:06:48.661550 containerd[1456]: 2025-07-07 00:06:48.624 [INFO][5421] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" iface="eth0" netns="" Jul 7 00:06:48.661550 containerd[1456]: 2025-07-07 00:06:48.624 [INFO][5421] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Jul 7 00:06:48.661550 containerd[1456]: 2025-07-07 00:06:48.624 [INFO][5421] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Jul 7 00:06:48.661550 containerd[1456]: 2025-07-07 00:06:48.649 [INFO][5428] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" HandleID="k8s-pod-network.4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Workload="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:48.661550 containerd[1456]: 2025-07-07 00:06:48.649 [INFO][5428] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:48.661550 containerd[1456]: 2025-07-07 00:06:48.649 [INFO][5428] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:48.661550 containerd[1456]: 2025-07-07 00:06:48.655 [WARNING][5428] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" HandleID="k8s-pod-network.4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Workload="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:48.661550 containerd[1456]: 2025-07-07 00:06:48.655 [INFO][5428] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" HandleID="k8s-pod-network.4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Workload="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:48.661550 containerd[1456]: 2025-07-07 00:06:48.656 [INFO][5428] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:48.661550 containerd[1456]: 2025-07-07 00:06:48.658 [INFO][5421] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Jul 7 00:06:48.661550 containerd[1456]: time="2025-07-07T00:06:48.660522303Z" level=info msg="TearDown network for sandbox \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\" successfully" Jul 7 00:06:48.661550 containerd[1456]: time="2025-07-07T00:06:48.660553494Z" level=info msg="StopPodSandbox for \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\" returns successfully" Jul 7 00:06:48.661550 containerd[1456]: time="2025-07-07T00:06:48.661127435Z" level=info msg="RemovePodSandbox for \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\"" Jul 7 00:06:48.661550 containerd[1456]: time="2025-07-07T00:06:48.661154145Z" level=info msg="Forcibly stopping sandbox \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\"" Jul 7 00:06:48.746771 containerd[1456]: 2025-07-07 00:06:48.706 [WARNING][5442] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0", GenerateName:"calico-kube-controllers-94599b4db-", Namespace:"calico-system", SelfLink:"", UID:"7f03664a-b5c1-41d9-b19a-954bfd2a5ea7", ResourceVersion:"1077", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"94599b4db", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"4f8ce2d6916778df50db633919d9514f68c046b29bcab2dea6f31ccf7327237b", Pod:"calico-kube-controllers-94599b4db-wbn2z", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.38.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali94d73bf1cb2", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:48.746771 containerd[1456]: 2025-07-07 00:06:48.706 [INFO][5442] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Jul 7 00:06:48.746771 containerd[1456]: 2025-07-07 00:06:48.706 [INFO][5442] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" iface="eth0" netns="" Jul 7 00:06:48.746771 containerd[1456]: 2025-07-07 00:06:48.707 [INFO][5442] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Jul 7 00:06:48.746771 containerd[1456]: 2025-07-07 00:06:48.707 [INFO][5442] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Jul 7 00:06:48.746771 containerd[1456]: 2025-07-07 00:06:48.734 [INFO][5449] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" HandleID="k8s-pod-network.4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Workload="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:48.746771 containerd[1456]: 2025-07-07 00:06:48.734 [INFO][5449] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:48.746771 containerd[1456]: 2025-07-07 00:06:48.734 [INFO][5449] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:48.746771 containerd[1456]: 2025-07-07 00:06:48.740 [WARNING][5449] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" HandleID="k8s-pod-network.4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Workload="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:48.746771 containerd[1456]: 2025-07-07 00:06:48.740 [INFO][5449] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" HandleID="k8s-pod-network.4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Workload="172--234--31--125-k8s-calico--kube--controllers--94599b4db--wbn2z-eth0" Jul 7 00:06:48.746771 containerd[1456]: 2025-07-07 00:06:48.741 [INFO][5449] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:48.746771 containerd[1456]: 2025-07-07 00:06:48.743 [INFO][5442] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1" Jul 7 00:06:48.748642 containerd[1456]: time="2025-07-07T00:06:48.746861670Z" level=info msg="TearDown network for sandbox \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\" successfully" Jul 7 00:06:48.751231 containerd[1456]: time="2025-07-07T00:06:48.751180581Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jul 7 00:06:48.751307 containerd[1456]: time="2025-07-07T00:06:48.751255842Z" level=info msg="RemovePodSandbox \"4af5986a292667cfa323767b374388d12d18df55ed67bf02b2f4e57ff60dc3d1\" returns successfully" Jul 7 00:06:48.751985 containerd[1456]: time="2025-07-07T00:06:48.751704741Z" level=info msg="StopPodSandbox for \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\"" Jul 7 00:06:48.832329 containerd[1456]: 2025-07-07 00:06:48.794 [WARNING][5463] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0", GenerateName:"calico-apiserver-85b7446c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"4741c1ac-aedc-48ec-8809-16303ada3c56", ResourceVersion:"1138", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85b7446c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39", Pod:"calico-apiserver-85b7446c9f-8xc2r", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.38.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali2d7d678cb83", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:48.832329 containerd[1456]: 2025-07-07 00:06:48.794 [INFO][5463] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Jul 7 00:06:48.832329 containerd[1456]: 2025-07-07 00:06:48.794 [INFO][5463] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" iface="eth0" netns="" Jul 7 00:06:48.832329 containerd[1456]: 2025-07-07 00:06:48.795 [INFO][5463] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Jul 7 00:06:48.832329 containerd[1456]: 2025-07-07 00:06:48.795 [INFO][5463] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Jul 7 00:06:48.832329 containerd[1456]: 2025-07-07 00:06:48.818 [INFO][5470] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" HandleID="k8s-pod-network.3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:48.832329 containerd[1456]: 2025-07-07 00:06:48.818 [INFO][5470] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:48.832329 containerd[1456]: 2025-07-07 00:06:48.818 [INFO][5470] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:48.832329 containerd[1456]: 2025-07-07 00:06:48.823 [WARNING][5470] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" HandleID="k8s-pod-network.3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:48.832329 containerd[1456]: 2025-07-07 00:06:48.823 [INFO][5470] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" HandleID="k8s-pod-network.3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:48.832329 containerd[1456]: 2025-07-07 00:06:48.824 [INFO][5470] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:48.832329 containerd[1456]: 2025-07-07 00:06:48.829 [INFO][5463] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Jul 7 00:06:48.834336 containerd[1456]: time="2025-07-07T00:06:48.832469213Z" level=info msg="TearDown network for sandbox \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\" successfully" Jul 7 00:06:48.834336 containerd[1456]: time="2025-07-07T00:06:48.833057134Z" level=info msg="StopPodSandbox for \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\" returns successfully" Jul 7 00:06:48.834336 containerd[1456]: time="2025-07-07T00:06:48.833707577Z" level=info msg="RemovePodSandbox for \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\"" Jul 7 00:06:48.834336 containerd[1456]: time="2025-07-07T00:06:48.833733347Z" level=info msg="Forcibly stopping sandbox \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\"" Jul 7 00:06:48.917301 containerd[1456]: 2025-07-07 00:06:48.879 [WARNING][5484] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0", GenerateName:"calico-apiserver-85b7446c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"4741c1ac-aedc-48ec-8809-16303ada3c56", ResourceVersion:"1138", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85b7446c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"2642f47361a7e1dd2de5a9c3ff9545bfa276cde846ff3c795027975faca36b39", Pod:"calico-apiserver-85b7446c9f-8xc2r", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.38.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali2d7d678cb83", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:48.917301 containerd[1456]: 2025-07-07 00:06:48.879 [INFO][5484] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Jul 7 00:06:48.917301 containerd[1456]: 2025-07-07 00:06:48.879 [INFO][5484] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" iface="eth0" netns="" Jul 7 00:06:48.917301 containerd[1456]: 2025-07-07 00:06:48.879 [INFO][5484] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Jul 7 00:06:48.917301 containerd[1456]: 2025-07-07 00:06:48.879 [INFO][5484] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Jul 7 00:06:48.917301 containerd[1456]: 2025-07-07 00:06:48.905 [INFO][5491] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" HandleID="k8s-pod-network.3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:48.917301 containerd[1456]: 2025-07-07 00:06:48.906 [INFO][5491] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:48.917301 containerd[1456]: 2025-07-07 00:06:48.906 [INFO][5491] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:48.917301 containerd[1456]: 2025-07-07 00:06:48.910 [WARNING][5491] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" HandleID="k8s-pod-network.3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:48.917301 containerd[1456]: 2025-07-07 00:06:48.910 [INFO][5491] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" HandleID="k8s-pod-network.3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--8xc2r-eth0" Jul 7 00:06:48.917301 containerd[1456]: 2025-07-07 00:06:48.911 [INFO][5491] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:48.917301 containerd[1456]: 2025-07-07 00:06:48.913 [INFO][5484] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0" Jul 7 00:06:48.917301 containerd[1456]: time="2025-07-07T00:06:48.916546837Z" level=info msg="TearDown network for sandbox \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\" successfully" Jul 7 00:06:48.922377 containerd[1456]: time="2025-07-07T00:06:48.921437019Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jul 7 00:06:48.922377 containerd[1456]: time="2025-07-07T00:06:48.921508541Z" level=info msg="RemovePodSandbox \"3cd6c33502c6b28cdc6ca5bd7ace8541d64224045cd6bfcb9fca26343fca38c0\" returns successfully" Jul 7 00:06:48.922377 containerd[1456]: time="2025-07-07T00:06:48.921991850Z" level=info msg="StopPodSandbox for \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\"" Jul 7 00:06:49.024074 containerd[1456]: 2025-07-07 00:06:48.973 [WARNING][5507] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-csi--node--driver--9knjc-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"2001a2e7-8165-484d-b2ef-2ba8a539af9a", ResourceVersion:"1110", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402", Pod:"csi-node-driver-9knjc", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.38.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif76aa88d3ce", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:49.024074 containerd[1456]: 2025-07-07 00:06:48.974 [INFO][5507] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Jul 7 00:06:49.024074 containerd[1456]: 2025-07-07 00:06:48.974 [INFO][5507] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" iface="eth0" netns="" Jul 7 00:06:49.024074 containerd[1456]: 2025-07-07 00:06:48.974 [INFO][5507] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Jul 7 00:06:49.024074 containerd[1456]: 2025-07-07 00:06:48.974 [INFO][5507] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Jul 7 00:06:49.024074 containerd[1456]: 2025-07-07 00:06:49.005 [INFO][5514] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" HandleID="k8s-pod-network.455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Workload="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:49.024074 containerd[1456]: 2025-07-07 00:06:49.006 [INFO][5514] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:49.024074 containerd[1456]: 2025-07-07 00:06:49.006 [INFO][5514] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:49.024074 containerd[1456]: 2025-07-07 00:06:49.011 [WARNING][5514] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" HandleID="k8s-pod-network.455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Workload="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:49.024074 containerd[1456]: 2025-07-07 00:06:49.011 [INFO][5514] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" HandleID="k8s-pod-network.455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Workload="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:49.024074 containerd[1456]: 2025-07-07 00:06:49.013 [INFO][5514] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:49.024074 containerd[1456]: 2025-07-07 00:06:49.021 [INFO][5507] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Jul 7 00:06:49.026490 containerd[1456]: time="2025-07-07T00:06:49.024070814Z" level=info msg="TearDown network for sandbox \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\" successfully" Jul 7 00:06:49.026490 containerd[1456]: time="2025-07-07T00:06:49.024119425Z" level=info msg="StopPodSandbox for \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\" returns successfully" Jul 7 00:06:49.026490 containerd[1456]: time="2025-07-07T00:06:49.024877128Z" level=info msg="RemovePodSandbox for \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\"" Jul 7 00:06:49.026490 containerd[1456]: time="2025-07-07T00:06:49.024903348Z" level=info msg="Forcibly stopping sandbox \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\"" Jul 7 00:06:49.110088 containerd[1456]: 2025-07-07 00:06:49.072 [WARNING][5528] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-csi--node--driver--9knjc-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"2001a2e7-8165-484d-b2ef-2ba8a539af9a", ResourceVersion:"1110", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"9bbd452e965c3d567c327cf41baf8b87edc12582499a92efb920a85ef412d402", Pod:"csi-node-driver-9knjc", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.38.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif76aa88d3ce", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:49.110088 containerd[1456]: 2025-07-07 00:06:49.073 [INFO][5528] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Jul 7 00:06:49.110088 containerd[1456]: 2025-07-07 00:06:49.073 [INFO][5528] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" iface="eth0" netns="" Jul 7 00:06:49.110088 containerd[1456]: 2025-07-07 00:06:49.073 [INFO][5528] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Jul 7 00:06:49.110088 containerd[1456]: 2025-07-07 00:06:49.073 [INFO][5528] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Jul 7 00:06:49.110088 containerd[1456]: 2025-07-07 00:06:49.095 [INFO][5535] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" HandleID="k8s-pod-network.455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Workload="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:49.110088 containerd[1456]: 2025-07-07 00:06:49.095 [INFO][5535] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:49.110088 containerd[1456]: 2025-07-07 00:06:49.095 [INFO][5535] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:49.110088 containerd[1456]: 2025-07-07 00:06:49.102 [WARNING][5535] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" HandleID="k8s-pod-network.455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Workload="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:49.110088 containerd[1456]: 2025-07-07 00:06:49.102 [INFO][5535] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" HandleID="k8s-pod-network.455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Workload="172--234--31--125-k8s-csi--node--driver--9knjc-eth0" Jul 7 00:06:49.110088 containerd[1456]: 2025-07-07 00:06:49.104 [INFO][5535] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:49.110088 containerd[1456]: 2025-07-07 00:06:49.106 [INFO][5528] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7" Jul 7 00:06:49.110576 containerd[1456]: time="2025-07-07T00:06:49.110167103Z" level=info msg="TearDown network for sandbox \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\" successfully" Jul 7 00:06:49.129665 containerd[1456]: time="2025-07-07T00:06:49.129607053Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jul 7 00:06:49.129799 containerd[1456]: time="2025-07-07T00:06:49.129721246Z" level=info msg="RemovePodSandbox \"455730be63de502115c8e389c1d683f402f2b8a78a0ca61123cd03e4c4ed56e7\" returns successfully" Jul 7 00:06:49.131987 containerd[1456]: time="2025-07-07T00:06:49.131954675Z" level=info msg="StopPodSandbox for \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\"" Jul 7 00:06:49.267086 containerd[1456]: 2025-07-07 00:06:49.192 [WARNING][5549] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"6b2ab80d-76dc-4520-bb95-f08a370b3d13", ResourceVersion:"977", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 5, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4", Pod:"coredns-7c65d6cfc9-pdhcs", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.38.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2b843ae349c", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:49.267086 containerd[1456]: 2025-07-07 00:06:49.192 [INFO][5549] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Jul 7 00:06:49.267086 containerd[1456]: 2025-07-07 00:06:49.192 [INFO][5549] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" iface="eth0" netns="" Jul 7 00:06:49.267086 containerd[1456]: 2025-07-07 00:06:49.192 [INFO][5549] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Jul 7 00:06:49.267086 containerd[1456]: 2025-07-07 00:06:49.192 [INFO][5549] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Jul 7 00:06:49.267086 containerd[1456]: 2025-07-07 00:06:49.233 [INFO][5556] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" HandleID="k8s-pod-network.b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:49.267086 containerd[1456]: 2025-07-07 00:06:49.234 [INFO][5556] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:49.267086 containerd[1456]: 2025-07-07 00:06:49.234 [INFO][5556] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:49.267086 containerd[1456]: 2025-07-07 00:06:49.246 [WARNING][5556] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" HandleID="k8s-pod-network.b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:49.267086 containerd[1456]: 2025-07-07 00:06:49.246 [INFO][5556] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" HandleID="k8s-pod-network.b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:49.267086 containerd[1456]: 2025-07-07 00:06:49.259 [INFO][5556] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:49.267086 containerd[1456]: 2025-07-07 00:06:49.262 [INFO][5549] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Jul 7 00:06:49.267599 containerd[1456]: time="2025-07-07T00:06:49.267081982Z" level=info msg="TearDown network for sandbox \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\" successfully" Jul 7 00:06:49.267599 containerd[1456]: time="2025-07-07T00:06:49.267112993Z" level=info msg="StopPodSandbox for \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\" returns successfully" Jul 7 00:06:49.267911 containerd[1456]: time="2025-07-07T00:06:49.267870857Z" level=info msg="RemovePodSandbox for \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\"" Jul 7 00:06:49.268262 containerd[1456]: time="2025-07-07T00:06:49.268092651Z" level=info msg="Forcibly stopping sandbox \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\"" Jul 7 00:06:49.352199 containerd[1456]: 2025-07-07 00:06:49.306 [WARNING][5570] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"6b2ab80d-76dc-4520-bb95-f08a370b3d13", ResourceVersion:"977", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 5, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"b4194d85d734ed4f08ee2738a443d3b17eed92e1efaf3cd0d2f93c95b9f7d0c4", Pod:"coredns-7c65d6cfc9-pdhcs", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.38.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2b843ae349c", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:49.352199 containerd[1456]: 2025-07-07 00:06:49.307 [INFO][5570] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Jul 7 00:06:49.352199 containerd[1456]: 2025-07-07 00:06:49.307 [INFO][5570] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" iface="eth0" netns="" Jul 7 00:06:49.352199 containerd[1456]: 2025-07-07 00:06:49.307 [INFO][5570] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Jul 7 00:06:49.352199 containerd[1456]: 2025-07-07 00:06:49.307 [INFO][5570] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Jul 7 00:06:49.352199 containerd[1456]: 2025-07-07 00:06:49.336 [INFO][5577] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" HandleID="k8s-pod-network.b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:49.352199 containerd[1456]: 2025-07-07 00:06:49.336 [INFO][5577] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:49.352199 containerd[1456]: 2025-07-07 00:06:49.336 [INFO][5577] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:49.352199 containerd[1456]: 2025-07-07 00:06:49.344 [WARNING][5577] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" HandleID="k8s-pod-network.b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:49.352199 containerd[1456]: 2025-07-07 00:06:49.344 [INFO][5577] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" HandleID="k8s-pod-network.b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--pdhcs-eth0" Jul 7 00:06:49.352199 containerd[1456]: 2025-07-07 00:06:49.347 [INFO][5577] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:49.352199 containerd[1456]: 2025-07-07 00:06:49.349 [INFO][5570] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd" Jul 7 00:06:49.352688 containerd[1456]: time="2025-07-07T00:06:49.352280125Z" level=info msg="TearDown network for sandbox \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\" successfully" Jul 7 00:06:49.359983 containerd[1456]: time="2025-07-07T00:06:49.359935390Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jul 7 00:06:49.360086 containerd[1456]: time="2025-07-07T00:06:49.360053522Z" level=info msg="RemovePodSandbox \"b0031baa2ad6d266fb413e2c27cdb784f8f7f2d7eb8aadcff722c825f758c9dd\" returns successfully" Jul 7 00:06:49.360768 containerd[1456]: time="2025-07-07T00:06:49.360739043Z" level=info msg="StopPodSandbox for \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\"" Jul 7 00:06:49.472299 containerd[1456]: 2025-07-07 00:06:49.417 [WARNING][5591] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0", GenerateName:"calico-apiserver-85b7446c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"15e90dc0-720a-4ae9-aac2-01720a32d5a8", ResourceVersion:"1113", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85b7446c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502", Pod:"calico-apiserver-85b7446c9f-78mbn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.38.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali5cdff996823", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:49.472299 containerd[1456]: 2025-07-07 00:06:49.418 [INFO][5591] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Jul 7 00:06:49.472299 containerd[1456]: 2025-07-07 00:06:49.418 [INFO][5591] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" iface="eth0" netns="" Jul 7 00:06:49.472299 containerd[1456]: 2025-07-07 00:06:49.418 [INFO][5591] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Jul 7 00:06:49.472299 containerd[1456]: 2025-07-07 00:06:49.418 [INFO][5591] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Jul 7 00:06:49.472299 containerd[1456]: 2025-07-07 00:06:49.453 [INFO][5599] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" HandleID="k8s-pod-network.09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:49.472299 containerd[1456]: 2025-07-07 00:06:49.453 [INFO][5599] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:49.472299 containerd[1456]: 2025-07-07 00:06:49.453 [INFO][5599] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:49.472299 containerd[1456]: 2025-07-07 00:06:49.463 [WARNING][5599] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" HandleID="k8s-pod-network.09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:49.472299 containerd[1456]: 2025-07-07 00:06:49.463 [INFO][5599] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" HandleID="k8s-pod-network.09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:49.472299 containerd[1456]: 2025-07-07 00:06:49.465 [INFO][5599] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:49.472299 containerd[1456]: 2025-07-07 00:06:49.469 [INFO][5591] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Jul 7 00:06:49.472299 containerd[1456]: time="2025-07-07T00:06:49.472107916Z" level=info msg="TearDown network for sandbox \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\" successfully" Jul 7 00:06:49.472299 containerd[1456]: time="2025-07-07T00:06:49.472148366Z" level=info msg="StopPodSandbox for \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\" returns successfully" Jul 7 00:06:49.473712 containerd[1456]: time="2025-07-07T00:06:49.472705935Z" level=info msg="RemovePodSandbox for \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\"" Jul 7 00:06:49.473712 containerd[1456]: time="2025-07-07T00:06:49.472887879Z" level=info msg="Forcibly stopping sandbox \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\"" Jul 7 00:06:49.571933 containerd[1456]: 2025-07-07 00:06:49.508 [WARNING][5614] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0", GenerateName:"calico-apiserver-85b7446c9f-", Namespace:"calico-apiserver", SelfLink:"", UID:"15e90dc0-720a-4ae9-aac2-01720a32d5a8", ResourceVersion:"1113", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 6, 2, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"85b7446c9f", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"ca5fd2d981e84230e814344569bee81b2c59fc0554b99cab6c3f572925a18502", Pod:"calico-apiserver-85b7446c9f-78mbn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.38.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali5cdff996823", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:49.571933 containerd[1456]: 2025-07-07 00:06:49.509 [INFO][5614] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Jul 7 00:06:49.571933 containerd[1456]: 2025-07-07 00:06:49.509 [INFO][5614] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" iface="eth0" netns="" Jul 7 00:06:49.571933 containerd[1456]: 2025-07-07 00:06:49.509 [INFO][5614] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Jul 7 00:06:49.571933 containerd[1456]: 2025-07-07 00:06:49.509 [INFO][5614] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Jul 7 00:06:49.571933 containerd[1456]: 2025-07-07 00:06:49.555 [INFO][5621] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" HandleID="k8s-pod-network.09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:49.571933 containerd[1456]: 2025-07-07 00:06:49.555 [INFO][5621] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:49.571933 containerd[1456]: 2025-07-07 00:06:49.555 [INFO][5621] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:49.571933 containerd[1456]: 2025-07-07 00:06:49.562 [WARNING][5621] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" HandleID="k8s-pod-network.09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:49.571933 containerd[1456]: 2025-07-07 00:06:49.562 [INFO][5621] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" HandleID="k8s-pod-network.09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Workload="172--234--31--125-k8s-calico--apiserver--85b7446c9f--78mbn-eth0" Jul 7 00:06:49.571933 containerd[1456]: 2025-07-07 00:06:49.566 [INFO][5621] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:49.571933 containerd[1456]: 2025-07-07 00:06:49.569 [INFO][5614] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8" Jul 7 00:06:49.571933 containerd[1456]: time="2025-07-07T00:06:49.571869624Z" level=info msg="TearDown network for sandbox \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\" successfully" Jul 7 00:06:49.575149 containerd[1456]: time="2025-07-07T00:06:49.575109420Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jul 7 00:06:49.575272 containerd[1456]: time="2025-07-07T00:06:49.575242723Z" level=info msg="RemovePodSandbox \"09420bed5484828e3f45377245164bad7fb9fd375b0f7ebb2f450a4c34371eb8\" returns successfully" Jul 7 00:06:49.575800 containerd[1456]: time="2025-07-07T00:06:49.575768522Z" level=info msg="StopPodSandbox for \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\"" Jul 7 00:06:49.676003 containerd[1456]: 2025-07-07 00:06:49.615 [WARNING][5635] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"178fc278-1010-4c58-8d67-4a6ee0c1491b", ResourceVersion:"1032", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 5, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f", Pod:"coredns-7c65d6cfc9-x9bb2", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.38.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali602eb21fd0a", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:49.676003 containerd[1456]: 2025-07-07 00:06:49.616 [INFO][5635] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Jul 7 00:06:49.676003 containerd[1456]: 2025-07-07 00:06:49.616 [INFO][5635] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" iface="eth0" netns="" Jul 7 00:06:49.676003 containerd[1456]: 2025-07-07 00:06:49.616 [INFO][5635] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Jul 7 00:06:49.676003 containerd[1456]: 2025-07-07 00:06:49.616 [INFO][5635] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Jul 7 00:06:49.676003 containerd[1456]: 2025-07-07 00:06:49.650 [INFO][5643] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" HandleID="k8s-pod-network.071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:49.676003 containerd[1456]: 2025-07-07 00:06:49.650 [INFO][5643] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:49.676003 containerd[1456]: 2025-07-07 00:06:49.650 [INFO][5643] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:49.676003 containerd[1456]: 2025-07-07 00:06:49.664 [WARNING][5643] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" HandleID="k8s-pod-network.071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:49.676003 containerd[1456]: 2025-07-07 00:06:49.664 [INFO][5643] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" HandleID="k8s-pod-network.071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:49.676003 containerd[1456]: 2025-07-07 00:06:49.666 [INFO][5643] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:49.676003 containerd[1456]: 2025-07-07 00:06:49.672 [INFO][5635] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Jul 7 00:06:49.676476 containerd[1456]: time="2025-07-07T00:06:49.676066540Z" level=info msg="TearDown network for sandbox \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\" successfully" Jul 7 00:06:49.676476 containerd[1456]: time="2025-07-07T00:06:49.676108741Z" level=info msg="StopPodSandbox for \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\" returns successfully" Jul 7 00:06:49.677656 containerd[1456]: time="2025-07-07T00:06:49.677266881Z" level=info msg="RemovePodSandbox for \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\"" Jul 7 00:06:49.677656 containerd[1456]: time="2025-07-07T00:06:49.677313541Z" level=info msg="Forcibly stopping sandbox \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\"" Jul 7 00:06:49.788831 containerd[1456]: 2025-07-07 00:06:49.734 [WARNING][5657] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"178fc278-1010-4c58-8d67-4a6ee0c1491b", ResourceVersion:"1032", Generation:0, CreationTimestamp:time.Date(2025, time.July, 7, 0, 5, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172-234-31-125", ContainerID:"b822bdbef283700f0f16c2d83d0dd25bba930add95265d38eb774cb7f77d457f", Pod:"coredns-7c65d6cfc9-x9bb2", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.38.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali602eb21fd0a", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 7 00:06:49.788831 containerd[1456]: 2025-07-07 00:06:49.734 [INFO][5657] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Jul 7 00:06:49.788831 containerd[1456]: 2025-07-07 00:06:49.734 [INFO][5657] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" iface="eth0" netns="" Jul 7 00:06:49.788831 containerd[1456]: 2025-07-07 00:06:49.734 [INFO][5657] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Jul 7 00:06:49.788831 containerd[1456]: 2025-07-07 00:06:49.734 [INFO][5657] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Jul 7 00:06:49.788831 containerd[1456]: 2025-07-07 00:06:49.773 [INFO][5664] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" HandleID="k8s-pod-network.071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:49.788831 containerd[1456]: 2025-07-07 00:06:49.773 [INFO][5664] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 7 00:06:49.788831 containerd[1456]: 2025-07-07 00:06:49.773 [INFO][5664] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 7 00:06:49.788831 containerd[1456]: 2025-07-07 00:06:49.780 [WARNING][5664] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" HandleID="k8s-pod-network.071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:49.788831 containerd[1456]: 2025-07-07 00:06:49.780 [INFO][5664] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" HandleID="k8s-pod-network.071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Workload="172--234--31--125-k8s-coredns--7c65d6cfc9--x9bb2-eth0" Jul 7 00:06:49.788831 containerd[1456]: 2025-07-07 00:06:49.782 [INFO][5664] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 7 00:06:49.788831 containerd[1456]: 2025-07-07 00:06:49.786 [INFO][5657] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93" Jul 7 00:06:49.789272 containerd[1456]: time="2025-07-07T00:06:49.788872066Z" level=info msg="TearDown network for sandbox \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\" successfully" Jul 7 00:06:49.793192 containerd[1456]: time="2025-07-07T00:06:49.793140791Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jul 7 00:06:49.793304 containerd[1456]: time="2025-07-07T00:06:49.793226593Z" level=info msg="RemovePodSandbox \"071645c102ea661577224515efe308dbd54af837060b4c246d625dba0337ee93\" returns successfully" Jul 7 00:06:54.306763 systemd[1]: run-containerd-runc-k8s.io-4cf0bc139daeb67691822e7adc2074362d46fec62277a2028a3b713ddffc59b6-runc.hHR7QZ.mount: Deactivated successfully. Jul 7 00:07:01.144965 kubelet[2515]: E0707 00:07:01.144913 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:07:03.524737 systemd[1]: run-containerd-runc-k8s.io-9508c81aa97aa2789f7a6f45f6bc7bc23e419b3162f2cdc33249cf514d6aba85-runc.Iq2wBU.mount: Deactivated successfully. Jul 7 00:07:16.145095 kubelet[2515]: E0707 00:07:16.144119 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:07:19.145211 kubelet[2515]: E0707 00:07:19.144866 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:07:26.144417 kubelet[2515]: E0707 00:07:26.144137 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:07:27.503167 systemd[1]: Started sshd@7-172.234.31.125:22-172.236.228.224:32262.service - OpenSSH per-connection server daemon (172.236.228.224:32262). Jul 7 00:07:28.555235 sshd[5830]: Connection closed by 172.236.228.224 port 32262 [preauth] Jul 7 00:07:28.557073 systemd[1]: sshd@7-172.234.31.125:22-172.236.228.224:32262.service: Deactivated successfully. Jul 7 00:07:28.632586 systemd[1]: Started sshd@8-172.234.31.125:22-172.236.228.224:32278.service - OpenSSH per-connection server daemon (172.236.228.224:32278). Jul 7 00:07:29.601005 sshd[5835]: Connection closed by 172.236.228.224 port 32278 [preauth] Jul 7 00:07:29.603595 systemd[1]: sshd@8-172.234.31.125:22-172.236.228.224:32278.service: Deactivated successfully. Jul 7 00:07:29.681766 systemd[1]: Started sshd@9-172.234.31.125:22-172.236.228.224:32292.service - OpenSSH per-connection server daemon (172.236.228.224:32292). Jul 7 00:07:30.810910 sshd[5840]: Connection closed by 172.236.228.224 port 32292 [preauth] Jul 7 00:07:30.813692 systemd[1]: sshd@9-172.234.31.125:22-172.236.228.224:32292.service: Deactivated successfully. Jul 7 00:07:33.520879 systemd[1]: run-containerd-runc-k8s.io-9508c81aa97aa2789f7a6f45f6bc7bc23e419b3162f2cdc33249cf514d6aba85-runc.VJYL7B.mount: Deactivated successfully. Jul 7 00:07:35.144893 kubelet[2515]: E0707 00:07:35.144498 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:07:45.143986 kubelet[2515]: E0707 00:07:45.143934 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:07:56.026681 systemd[1]: Started sshd@10-172.234.31.125:22-147.75.109.163:52588.service - OpenSSH per-connection server daemon (147.75.109.163:52588). Jul 7 00:07:56.392664 sshd[5939]: Accepted publickey for core from 147.75.109.163 port 52588 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:07:56.394338 sshd[5939]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:07:56.400172 systemd-logind[1440]: New session 8 of user core. Jul 7 00:07:56.407677 systemd[1]: Started session-8.scope - Session 8 of User core. Jul 7 00:07:56.741405 sshd[5939]: pam_unix(sshd:session): session closed for user core Jul 7 00:07:56.747073 systemd[1]: sshd@10-172.234.31.125:22-147.75.109.163:52588.service: Deactivated successfully. Jul 7 00:07:56.749583 systemd[1]: session-8.scope: Deactivated successfully. Jul 7 00:07:56.752978 systemd-logind[1440]: Session 8 logged out. Waiting for processes to exit. Jul 7 00:07:56.754325 systemd-logind[1440]: Removed session 8. Jul 7 00:08:01.148059 kubelet[2515]: E0707 00:08:01.148020 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:08:01.809764 systemd[1]: Started sshd@11-172.234.31.125:22-147.75.109.163:52596.service - OpenSSH per-connection server daemon (147.75.109.163:52596). Jul 7 00:08:02.162485 sshd[6000]: Accepted publickey for core from 147.75.109.163 port 52596 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:08:02.164247 sshd[6000]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:08:02.170043 systemd-logind[1440]: New session 9 of user core. Jul 7 00:08:02.176526 systemd[1]: Started session-9.scope - Session 9 of User core. Jul 7 00:08:02.511404 sshd[6000]: pam_unix(sshd:session): session closed for user core Jul 7 00:08:02.514971 systemd-logind[1440]: Session 9 logged out. Waiting for processes to exit. Jul 7 00:08:02.517277 systemd[1]: sshd@11-172.234.31.125:22-147.75.109.163:52596.service: Deactivated successfully. Jul 7 00:08:02.520050 systemd[1]: session-9.scope: Deactivated successfully. Jul 7 00:08:02.521915 systemd-logind[1440]: Removed session 9. Jul 7 00:08:03.524721 systemd[1]: run-containerd-runc-k8s.io-9508c81aa97aa2789f7a6f45f6bc7bc23e419b3162f2cdc33249cf514d6aba85-runc.p8lyx3.mount: Deactivated successfully. Jul 7 00:08:07.583672 systemd[1]: Started sshd@12-172.234.31.125:22-147.75.109.163:34338.service - OpenSSH per-connection server daemon (147.75.109.163:34338). Jul 7 00:08:07.957421 sshd[6037]: Accepted publickey for core from 147.75.109.163 port 34338 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:08:07.959869 sshd[6037]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:08:07.965432 systemd-logind[1440]: New session 10 of user core. Jul 7 00:08:07.972519 systemd[1]: Started session-10.scope - Session 10 of User core. Jul 7 00:08:08.304609 sshd[6037]: pam_unix(sshd:session): session closed for user core Jul 7 00:08:08.310919 systemd-logind[1440]: Session 10 logged out. Waiting for processes to exit. Jul 7 00:08:08.311200 systemd[1]: sshd@12-172.234.31.125:22-147.75.109.163:34338.service: Deactivated successfully. Jul 7 00:08:08.315429 systemd[1]: session-10.scope: Deactivated successfully. Jul 7 00:08:08.316806 systemd-logind[1440]: Removed session 10. Jul 7 00:08:08.367756 systemd[1]: Started sshd@13-172.234.31.125:22-147.75.109.163:34354.service - OpenSSH per-connection server daemon (147.75.109.163:34354). Jul 7 00:08:08.733778 sshd[6051]: Accepted publickey for core from 147.75.109.163 port 34354 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:08:08.736187 sshd[6051]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:08:08.743552 systemd-logind[1440]: New session 11 of user core. Jul 7 00:08:08.749552 systemd[1]: Started session-11.scope - Session 11 of User core. Jul 7 00:08:09.096557 sshd[6051]: pam_unix(sshd:session): session closed for user core Jul 7 00:08:09.102156 systemd-logind[1440]: Session 11 logged out. Waiting for processes to exit. Jul 7 00:08:09.102742 systemd[1]: sshd@13-172.234.31.125:22-147.75.109.163:34354.service: Deactivated successfully. Jul 7 00:08:09.105623 systemd[1]: session-11.scope: Deactivated successfully. Jul 7 00:08:09.106870 systemd-logind[1440]: Removed session 11. Jul 7 00:08:09.170870 systemd[1]: Started sshd@14-172.234.31.125:22-147.75.109.163:34366.service - OpenSSH per-connection server daemon (147.75.109.163:34366). Jul 7 00:08:09.525983 sshd[6062]: Accepted publickey for core from 147.75.109.163 port 34366 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:08:09.528100 sshd[6062]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:08:09.533234 systemd-logind[1440]: New session 12 of user core. Jul 7 00:08:09.535521 systemd[1]: Started session-12.scope - Session 12 of User core. Jul 7 00:08:09.861838 sshd[6062]: pam_unix(sshd:session): session closed for user core Jul 7 00:08:09.866892 systemd[1]: sshd@14-172.234.31.125:22-147.75.109.163:34366.service: Deactivated successfully. Jul 7 00:08:09.871258 systemd[1]: session-12.scope: Deactivated successfully. Jul 7 00:08:09.873798 systemd-logind[1440]: Session 12 logged out. Waiting for processes to exit. Jul 7 00:08:09.876048 systemd-logind[1440]: Removed session 12. Jul 7 00:08:14.937661 systemd[1]: Started sshd@15-172.234.31.125:22-147.75.109.163:34368.service - OpenSSH per-connection server daemon (147.75.109.163:34368). Jul 7 00:08:15.318510 sshd[6075]: Accepted publickey for core from 147.75.109.163 port 34368 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:08:15.319992 sshd[6075]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:08:15.326638 systemd-logind[1440]: New session 13 of user core. Jul 7 00:08:15.332668 systemd[1]: Started session-13.scope - Session 13 of User core. Jul 7 00:08:15.658796 sshd[6075]: pam_unix(sshd:session): session closed for user core Jul 7 00:08:15.663793 systemd[1]: sshd@15-172.234.31.125:22-147.75.109.163:34368.service: Deactivated successfully. Jul 7 00:08:15.667453 systemd[1]: session-13.scope: Deactivated successfully. Jul 7 00:08:15.668249 systemd-logind[1440]: Session 13 logged out. Waiting for processes to exit. Jul 7 00:08:15.669415 systemd-logind[1440]: Removed session 13. Jul 7 00:08:15.724604 systemd[1]: Started sshd@16-172.234.31.125:22-147.75.109.163:34370.service - OpenSSH per-connection server daemon (147.75.109.163:34370). Jul 7 00:08:16.077555 sshd[6088]: Accepted publickey for core from 147.75.109.163 port 34370 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:08:16.079968 sshd[6088]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:08:16.085411 systemd-logind[1440]: New session 14 of user core. Jul 7 00:08:16.088478 systemd[1]: Started session-14.scope - Session 14 of User core. Jul 7 00:08:16.704732 sshd[6088]: pam_unix(sshd:session): session closed for user core Jul 7 00:08:16.708623 systemd[1]: sshd@16-172.234.31.125:22-147.75.109.163:34370.service: Deactivated successfully. Jul 7 00:08:16.710812 systemd[1]: session-14.scope: Deactivated successfully. Jul 7 00:08:16.712144 systemd-logind[1440]: Session 14 logged out. Waiting for processes to exit. Jul 7 00:08:16.714283 systemd-logind[1440]: Removed session 14. Jul 7 00:08:16.781593 systemd[1]: Started sshd@17-172.234.31.125:22-147.75.109.163:43704.service - OpenSSH per-connection server daemon (147.75.109.163:43704). Jul 7 00:08:17.157685 sshd[6100]: Accepted publickey for core from 147.75.109.163 port 43704 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:08:17.158534 sshd[6100]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:08:17.163031 systemd-logind[1440]: New session 15 of user core. Jul 7 00:08:17.169806 systemd[1]: Started session-15.scope - Session 15 of User core. Jul 7 00:08:19.144325 kubelet[2515]: E0707 00:08:19.144231 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:08:19.300801 sshd[6100]: pam_unix(sshd:session): session closed for user core Jul 7 00:08:19.305433 systemd-logind[1440]: Session 15 logged out. Waiting for processes to exit. Jul 7 00:08:19.305481 systemd[1]: sshd@17-172.234.31.125:22-147.75.109.163:43704.service: Deactivated successfully. Jul 7 00:08:19.307979 systemd[1]: session-15.scope: Deactivated successfully. Jul 7 00:08:19.309818 systemd-logind[1440]: Removed session 15. Jul 7 00:08:19.371590 systemd[1]: Started sshd@18-172.234.31.125:22-147.75.109.163:43718.service - OpenSSH per-connection server daemon (147.75.109.163:43718). Jul 7 00:08:19.736508 sshd[6119]: Accepted publickey for core from 147.75.109.163 port 43718 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:08:19.738172 sshd[6119]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:08:19.742843 systemd-logind[1440]: New session 16 of user core. Jul 7 00:08:19.745904 systemd[1]: Started session-16.scope - Session 16 of User core. Jul 7 00:08:20.201075 sshd[6119]: pam_unix(sshd:session): session closed for user core Jul 7 00:08:20.205441 systemd-logind[1440]: Session 16 logged out. Waiting for processes to exit. Jul 7 00:08:20.206744 systemd[1]: sshd@18-172.234.31.125:22-147.75.109.163:43718.service: Deactivated successfully. Jul 7 00:08:20.209264 systemd[1]: session-16.scope: Deactivated successfully. Jul 7 00:08:20.211034 systemd-logind[1440]: Removed session 16. Jul 7 00:08:20.272614 systemd[1]: Started sshd@19-172.234.31.125:22-147.75.109.163:43732.service - OpenSSH per-connection server daemon (147.75.109.163:43732). Jul 7 00:08:20.640736 sshd[6130]: Accepted publickey for core from 147.75.109.163 port 43732 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:08:20.642725 sshd[6130]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:08:20.648826 systemd-logind[1440]: New session 17 of user core. Jul 7 00:08:20.653492 systemd[1]: Started session-17.scope - Session 17 of User core. Jul 7 00:08:20.987688 sshd[6130]: pam_unix(sshd:session): session closed for user core Jul 7 00:08:20.992831 systemd[1]: sshd@19-172.234.31.125:22-147.75.109.163:43732.service: Deactivated successfully. Jul 7 00:08:20.995936 systemd[1]: session-17.scope: Deactivated successfully. Jul 7 00:08:20.996897 systemd-logind[1440]: Session 17 logged out. Waiting for processes to exit. Jul 7 00:08:20.998158 systemd-logind[1440]: Removed session 17. Jul 7 00:08:26.057686 systemd[1]: Started sshd@20-172.234.31.125:22-147.75.109.163:44978.service - OpenSSH per-connection server daemon (147.75.109.163:44978). Jul 7 00:08:26.425954 sshd[6190]: Accepted publickey for core from 147.75.109.163 port 44978 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:08:26.428148 sshd[6190]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:08:26.433486 systemd-logind[1440]: New session 18 of user core. Jul 7 00:08:26.437479 systemd[1]: Started session-18.scope - Session 18 of User core. Jul 7 00:08:26.763593 sshd[6190]: pam_unix(sshd:session): session closed for user core Jul 7 00:08:26.768568 systemd[1]: sshd@20-172.234.31.125:22-147.75.109.163:44978.service: Deactivated successfully. Jul 7 00:08:26.773676 systemd[1]: session-18.scope: Deactivated successfully. Jul 7 00:08:26.776002 systemd-logind[1440]: Session 18 logged out. Waiting for processes to exit. Jul 7 00:08:26.778160 systemd-logind[1440]: Removed session 18. Jul 7 00:08:29.144353 kubelet[2515]: E0707 00:08:29.144294 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:08:31.144405 kubelet[2515]: E0707 00:08:31.144283 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:08:31.835753 systemd[1]: Started sshd@21-172.234.31.125:22-147.75.109.163:44980.service - OpenSSH per-connection server daemon (147.75.109.163:44980). Jul 7 00:08:32.196315 sshd[6203]: Accepted publickey for core from 147.75.109.163 port 44980 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:08:32.197950 sshd[6203]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:08:32.203160 systemd-logind[1440]: New session 19 of user core. Jul 7 00:08:32.208485 systemd[1]: Started session-19.scope - Session 19 of User core. Jul 7 00:08:32.543999 sshd[6203]: pam_unix(sshd:session): session closed for user core Jul 7 00:08:32.557236 systemd[1]: sshd@21-172.234.31.125:22-147.75.109.163:44980.service: Deactivated successfully. Jul 7 00:08:32.565255 systemd[1]: session-19.scope: Deactivated successfully. Jul 7 00:08:32.566343 systemd-logind[1440]: Session 19 logged out. Waiting for processes to exit. Jul 7 00:08:32.568326 systemd-logind[1440]: Removed session 19. Jul 7 00:08:33.529934 systemd[1]: run-containerd-runc-k8s.io-9508c81aa97aa2789f7a6f45f6bc7bc23e419b3162f2cdc33249cf514d6aba85-runc.ZpQHM8.mount: Deactivated successfully. Jul 7 00:08:36.145157 kubelet[2515]: E0707 00:08:36.144777 2515 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 172.232.0.9 172.232.0.19 172.232.0.20" Jul 7 00:08:37.607416 systemd[1]: Started sshd@22-172.234.31.125:22-147.75.109.163:59588.service - OpenSSH per-connection server daemon (147.75.109.163:59588). Jul 7 00:08:37.965530 sshd[6235]: Accepted publickey for core from 147.75.109.163 port 59588 ssh2: RSA SHA256:E8VvoKFBGB97naanZasa7CRd5+36IUUTzq11yQ2Rock Jul 7 00:08:37.967187 sshd[6235]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 7 00:08:37.971658 systemd-logind[1440]: New session 20 of user core. Jul 7 00:08:37.974571 systemd[1]: Started session-20.scope - Session 20 of User core. Jul 7 00:08:38.287233 sshd[6235]: pam_unix(sshd:session): session closed for user core Jul 7 00:08:38.291572 systemd[1]: sshd@22-172.234.31.125:22-147.75.109.163:59588.service: Deactivated successfully. Jul 7 00:08:38.293998 systemd[1]: session-20.scope: Deactivated successfully. Jul 7 00:08:38.295640 systemd-logind[1440]: Session 20 logged out. Waiting for processes to exit. Jul 7 00:08:38.296663 systemd-logind[1440]: Removed session 20.