May 27 03:14:45.918753 kernel: Linux version 6.12.30-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT_DYNAMIC Tue May 27 01:09:43 -00 2025 May 27 03:14:45.918838 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=f6c186658a19d5a08471ef76df75f82494b37b46908f9237b2c3cf497da860c6 May 27 03:14:45.918855 kernel: BIOS-provided physical RAM map: May 27 03:14:45.918867 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable May 27 03:14:45.918877 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved May 27 03:14:45.918887 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved May 27 03:14:45.918900 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdafff] usable May 27 03:14:45.918917 kernel: BIOS-e820: [mem 0x000000007ffdb000-0x000000007fffffff] reserved May 27 03:14:45.918937 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved May 27 03:14:45.918949 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved May 27 03:14:45.918960 kernel: NX (Execute Disable) protection: active May 27 03:14:45.918972 kernel: APIC: Static calls initialized May 27 03:14:45.918983 kernel: SMBIOS 2.8 present. May 27 03:14:45.918995 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 May 27 03:14:45.919018 kernel: DMI: Memory slots populated: 1/1 May 27 03:14:45.919040 kernel: Hypervisor detected: KVM May 27 03:14:45.920961 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 May 27 03:14:45.920990 kernel: kvm-clock: using sched offset of 4543416430 cycles May 27 03:14:45.921005 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns May 27 03:14:45.921018 kernel: tsc: Detected 2494.136 MHz processor May 27 03:14:45.921032 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved May 27 03:14:45.921046 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable May 27 03:14:45.921070 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 May 27 03:14:45.921101 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs May 27 03:14:45.921114 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT May 27 03:14:45.921127 kernel: ACPI: Early table checksum verification disabled May 27 03:14:45.921139 kernel: ACPI: RSDP 0x00000000000F5950 000014 (v00 BOCHS ) May 27 03:14:45.921153 kernel: ACPI: RSDT 0x000000007FFE1986 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 27 03:14:45.921167 kernel: ACPI: FACP 0x000000007FFE176A 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 27 03:14:45.921179 kernel: ACPI: DSDT 0x000000007FFE0040 00172A (v01 BOCHS BXPC 00000001 BXPC 00000001) May 27 03:14:45.921192 kernel: ACPI: FACS 0x000000007FFE0000 000040 May 27 03:14:45.921205 kernel: ACPI: APIC 0x000000007FFE17DE 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 27 03:14:45.921227 kernel: ACPI: HPET 0x000000007FFE185E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 27 03:14:45.921240 kernel: ACPI: SRAT 0x000000007FFE1896 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 27 03:14:45.921253 kernel: ACPI: WAET 0x000000007FFE195E 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 27 03:14:45.921267 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe176a-0x7ffe17dd] May 27 03:14:45.921280 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe1769] May 27 03:14:45.921293 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] May 27 03:14:45.921306 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17de-0x7ffe185d] May 27 03:14:45.921320 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe185e-0x7ffe1895] May 27 03:14:45.921349 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe1896-0x7ffe195d] May 27 03:14:45.921363 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe195e-0x7ffe1985] May 27 03:14:45.921376 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] May 27 03:14:45.921391 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] May 27 03:14:45.921405 kernel: NUMA: Node 0 [mem 0x00001000-0x0009ffff] + [mem 0x00100000-0x7ffdafff] -> [mem 0x00001000-0x7ffdafff] May 27 03:14:45.921419 kernel: NODE_DATA(0) allocated [mem 0x7ffd3dc0-0x7ffdafff] May 27 03:14:45.921439 kernel: Zone ranges: May 27 03:14:45.921453 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] May 27 03:14:45.921467 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdafff] May 27 03:14:45.921481 kernel: Normal empty May 27 03:14:45.921494 kernel: Device empty May 27 03:14:45.921508 kernel: Movable zone start for each node May 27 03:14:45.921521 kernel: Early memory node ranges May 27 03:14:45.921535 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] May 27 03:14:45.921549 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdafff] May 27 03:14:45.921571 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdafff] May 27 03:14:45.921585 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 27 03:14:45.921599 kernel: On node 0, zone DMA: 97 pages in unavailable ranges May 27 03:14:45.921613 kernel: On node 0, zone DMA32: 37 pages in unavailable ranges May 27 03:14:45.921626 kernel: ACPI: PM-Timer IO Port: 0x608 May 27 03:14:45.921640 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) May 27 03:14:45.921656 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 May 27 03:14:45.921671 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) May 27 03:14:45.921687 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) May 27 03:14:45.921708 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) May 27 03:14:45.921726 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) May 27 03:14:45.921740 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) May 27 03:14:45.921754 kernel: ACPI: Using ACPI (MADT) for SMP configuration information May 27 03:14:45.921768 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 May 27 03:14:45.921782 kernel: TSC deadline timer available May 27 03:14:45.921795 kernel: CPU topo: Max. logical packages: 1 May 27 03:14:45.921809 kernel: CPU topo: Max. logical dies: 1 May 27 03:14:45.921823 kernel: CPU topo: Max. dies per package: 1 May 27 03:14:45.921836 kernel: CPU topo: Max. threads per core: 1 May 27 03:14:45.921858 kernel: CPU topo: Num. cores per package: 2 May 27 03:14:45.921871 kernel: CPU topo: Num. threads per package: 2 May 27 03:14:45.921884 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs May 27 03:14:45.921898 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() May 27 03:14:45.921911 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices May 27 03:14:45.921925 kernel: Booting paravirtualized kernel on KVM May 27 03:14:45.921939 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns May 27 03:14:45.921953 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 May 27 03:14:45.921967 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u1048576 May 27 03:14:45.921989 kernel: pcpu-alloc: s207832 r8192 d29736 u1048576 alloc=1*2097152 May 27 03:14:45.922002 kernel: pcpu-alloc: [0] 0 1 May 27 03:14:45.922015 kernel: kvm-guest: PV spinlocks disabled, no host support May 27 03:14:45.922031 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=f6c186658a19d5a08471ef76df75f82494b37b46908f9237b2c3cf497da860c6 May 27 03:14:45.922045 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 27 03:14:45.922068 kernel: random: crng init done May 27 03:14:45.922082 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 27 03:14:45.922096 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) May 27 03:14:45.922118 kernel: Fallback order for Node 0: 0 May 27 03:14:45.922131 kernel: Built 1 zonelists, mobility grouping on. Total pages: 524153 May 27 03:14:45.922144 kernel: Policy zone: DMA32 May 27 03:14:45.922158 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 27 03:14:45.922171 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 May 27 03:14:45.922185 kernel: Kernel/User page tables isolation: enabled May 27 03:14:45.922198 kernel: ftrace: allocating 40081 entries in 157 pages May 27 03:14:45.922211 kernel: ftrace: allocated 157 pages with 5 groups May 27 03:14:45.922225 kernel: Dynamic Preempt: voluntary May 27 03:14:45.922248 kernel: rcu: Preemptible hierarchical RCU implementation. May 27 03:14:45.922264 kernel: rcu: RCU event tracing is enabled. May 27 03:14:45.922278 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. May 27 03:14:45.922292 kernel: Trampoline variant of Tasks RCU enabled. May 27 03:14:45.922306 kernel: Rude variant of Tasks RCU enabled. May 27 03:14:45.922319 kernel: Tracing variant of Tasks RCU enabled. May 27 03:14:45.922331 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 27 03:14:45.922346 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 May 27 03:14:45.922361 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 27 03:14:45.922386 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 27 03:14:45.922400 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 27 03:14:45.922411 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 May 27 03:14:45.922422 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. May 27 03:14:45.922436 kernel: Console: colour VGA+ 80x25 May 27 03:14:45.922450 kernel: printk: legacy console [tty0] enabled May 27 03:14:45.922464 kernel: printk: legacy console [ttyS0] enabled May 27 03:14:45.922478 kernel: ACPI: Core revision 20240827 May 27 03:14:45.922490 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns May 27 03:14:45.922530 kernel: APIC: Switch to symmetric I/O mode setup May 27 03:14:45.922543 kernel: x2apic enabled May 27 03:14:45.922557 kernel: APIC: Switched APIC routing to: physical x2apic May 27 03:14:45.922577 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 May 27 03:14:45.922594 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x23f39654230, max_idle_ns: 440795207432 ns May 27 03:14:45.922607 kernel: Calibrating delay loop (skipped) preset value.. 4988.27 BogoMIPS (lpj=2494136) May 27 03:14:45.922621 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 May 27 03:14:45.922634 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 May 27 03:14:45.922648 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization May 27 03:14:45.922669 kernel: Spectre V2 : Mitigation: Retpolines May 27 03:14:45.922682 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT May 27 03:14:45.922696 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls May 27 03:14:45.922710 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier May 27 03:14:45.922724 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl May 27 03:14:45.922737 kernel: MDS: Mitigation: Clear CPU buffers May 27 03:14:45.922751 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode May 27 03:14:45.922774 kernel: ITS: Mitigation: Aligned branch/return thunks May 27 03:14:45.922787 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' May 27 03:14:45.922801 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' May 27 03:14:45.922816 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' May 27 03:14:45.922830 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 May 27 03:14:45.922844 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. May 27 03:14:45.922859 kernel: Freeing SMP alternatives memory: 32K May 27 03:14:45.922873 kernel: pid_max: default: 32768 minimum: 301 May 27 03:14:45.922887 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima May 27 03:14:45.922909 kernel: landlock: Up and running. May 27 03:14:45.922923 kernel: SELinux: Initializing. May 27 03:14:45.922937 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) May 27 03:14:45.922952 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) May 27 03:14:45.922966 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x4f, stepping: 0x1) May 27 03:14:45.922981 kernel: Performance Events: unsupported p6 CPU model 79 no PMU driver, software events only. May 27 03:14:45.922994 kernel: signal: max sigframe size: 1776 May 27 03:14:45.923008 kernel: rcu: Hierarchical SRCU implementation. May 27 03:14:45.923023 kernel: rcu: Max phase no-delay instances is 400. May 27 03:14:45.923045 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level May 27 03:14:45.925121 kernel: NMI watchdog: Perf NMI watchdog permanently disabled May 27 03:14:45.925141 kernel: smp: Bringing up secondary CPUs ... May 27 03:14:45.925157 kernel: smpboot: x86: Booting SMP configuration: May 27 03:14:45.925178 kernel: .... node #0, CPUs: #1 May 27 03:14:45.925193 kernel: smp: Brought up 1 node, 2 CPUs May 27 03:14:45.925210 kernel: smpboot: Total of 2 processors activated (9976.54 BogoMIPS) May 27 03:14:45.925236 kernel: Memory: 1966908K/2096612K available (14336K kernel code, 2430K rwdata, 9952K rodata, 54416K init, 2552K bss, 125140K reserved, 0K cma-reserved) May 27 03:14:45.925251 kernel: devtmpfs: initialized May 27 03:14:45.925280 kernel: x86/mm: Memory block size: 128MB May 27 03:14:45.925297 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 27 03:14:45.925313 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) May 27 03:14:45.925328 kernel: pinctrl core: initialized pinctrl subsystem May 27 03:14:45.925343 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 27 03:14:45.925357 kernel: audit: initializing netlink subsys (disabled) May 27 03:14:45.925371 kernel: audit: type=2000 audit(1748315682.204:1): state=initialized audit_enabled=0 res=1 May 27 03:14:45.925386 kernel: thermal_sys: Registered thermal governor 'step_wise' May 27 03:14:45.925401 kernel: thermal_sys: Registered thermal governor 'user_space' May 27 03:14:45.925424 kernel: cpuidle: using governor menu May 27 03:14:45.925439 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 27 03:14:45.925454 kernel: dca service started, version 1.12.1 May 27 03:14:45.925469 kernel: PCI: Using configuration type 1 for base access May 27 03:14:45.925485 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. May 27 03:14:45.925500 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages May 27 03:14:45.925516 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page May 27 03:14:45.925531 kernel: ACPI: Added _OSI(Module Device) May 27 03:14:45.925545 kernel: ACPI: Added _OSI(Processor Device) May 27 03:14:45.925568 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 27 03:14:45.925584 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 27 03:14:45.925598 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded May 27 03:14:45.925613 kernel: ACPI: Interpreter enabled May 27 03:14:45.925627 kernel: ACPI: PM: (supports S0 S5) May 27 03:14:45.925641 kernel: ACPI: Using IOAPIC for interrupt routing May 27 03:14:45.925657 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug May 27 03:14:45.925672 kernel: PCI: Using E820 reservations for host bridge windows May 27 03:14:45.925687 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F May 27 03:14:45.925702 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) May 27 03:14:45.926017 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] May 27 03:14:45.926221 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] May 27 03:14:45.926372 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge May 27 03:14:45.926392 kernel: acpiphp: Slot [3] registered May 27 03:14:45.926407 kernel: acpiphp: Slot [4] registered May 27 03:14:45.926422 kernel: acpiphp: Slot [5] registered May 27 03:14:45.926437 kernel: acpiphp: Slot [6] registered May 27 03:14:45.926469 kernel: acpiphp: Slot [7] registered May 27 03:14:45.926484 kernel: acpiphp: Slot [8] registered May 27 03:14:45.926499 kernel: acpiphp: Slot [9] registered May 27 03:14:45.926513 kernel: acpiphp: Slot [10] registered May 27 03:14:45.926528 kernel: acpiphp: Slot [11] registered May 27 03:14:45.926544 kernel: acpiphp: Slot [12] registered May 27 03:14:45.926559 kernel: acpiphp: Slot [13] registered May 27 03:14:45.926574 kernel: acpiphp: Slot [14] registered May 27 03:14:45.926589 kernel: acpiphp: Slot [15] registered May 27 03:14:45.926612 kernel: acpiphp: Slot [16] registered May 27 03:14:45.926627 kernel: acpiphp: Slot [17] registered May 27 03:14:45.926640 kernel: acpiphp: Slot [18] registered May 27 03:14:45.926655 kernel: acpiphp: Slot [19] registered May 27 03:14:45.926670 kernel: acpiphp: Slot [20] registered May 27 03:14:45.926684 kernel: acpiphp: Slot [21] registered May 27 03:14:45.926699 kernel: acpiphp: Slot [22] registered May 27 03:14:45.926714 kernel: acpiphp: Slot [23] registered May 27 03:14:45.926729 kernel: acpiphp: Slot [24] registered May 27 03:14:45.926744 kernel: acpiphp: Slot [25] registered May 27 03:14:45.926793 kernel: acpiphp: Slot [26] registered May 27 03:14:45.926816 kernel: acpiphp: Slot [27] registered May 27 03:14:45.926832 kernel: acpiphp: Slot [28] registered May 27 03:14:45.926847 kernel: acpiphp: Slot [29] registered May 27 03:14:45.926861 kernel: acpiphp: Slot [30] registered May 27 03:14:45.926875 kernel: acpiphp: Slot [31] registered May 27 03:14:45.926891 kernel: PCI host bridge to bus 0000:00 May 27 03:14:45.928102 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] May 27 03:14:45.928292 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] May 27 03:14:45.928424 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] May 27 03:14:45.928553 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] May 27 03:14:45.928742 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] May 27 03:14:45.928876 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] May 27 03:14:45.929109 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 conventional PCI endpoint May 27 03:14:45.929291 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 conventional PCI endpoint May 27 03:14:45.929495 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 conventional PCI endpoint May 27 03:14:45.929639 kernel: pci 0000:00:01.1: BAR 4 [io 0xc1e0-0xc1ef] May 27 03:14:45.929788 kernel: pci 0000:00:01.1: BAR 0 [io 0x01f0-0x01f7]: legacy IDE quirk May 27 03:14:45.929934 kernel: pci 0000:00:01.1: BAR 1 [io 0x03f6]: legacy IDE quirk May 27 03:14:45.931899 kernel: pci 0000:00:01.1: BAR 2 [io 0x0170-0x0177]: legacy IDE quirk May 27 03:14:45.932852 kernel: pci 0000:00:01.1: BAR 3 [io 0x0376]: legacy IDE quirk May 27 03:14:45.933041 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 conventional PCI endpoint May 27 03:14:45.933244 kernel: pci 0000:00:01.2: BAR 4 [io 0xc180-0xc19f] May 27 03:14:45.933360 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 conventional PCI endpoint May 27 03:14:45.933459 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI May 27 03:14:45.933582 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB May 27 03:14:45.933751 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 conventional PCI endpoint May 27 03:14:45.933854 kernel: pci 0000:00:02.0: BAR 0 [mem 0xfe000000-0xfe7fffff pref] May 27 03:14:45.933966 kernel: pci 0000:00:02.0: BAR 2 [mem 0xfe800000-0xfe803fff 64bit pref] May 27 03:14:45.934074 kernel: pci 0000:00:02.0: BAR 4 [mem 0xfebf0000-0xfebf0fff] May 27 03:14:45.934173 kernel: pci 0000:00:02.0: ROM [mem 0xfebe0000-0xfebeffff pref] May 27 03:14:45.934272 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] May 27 03:14:45.934384 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint May 27 03:14:45.934482 kernel: pci 0000:00:03.0: BAR 0 [io 0xc1a0-0xc1bf] May 27 03:14:45.934592 kernel: pci 0000:00:03.0: BAR 1 [mem 0xfebf1000-0xfebf1fff] May 27 03:14:45.934691 kernel: pci 0000:00:03.0: BAR 4 [mem 0xfe804000-0xfe807fff 64bit pref] May 27 03:14:45.934859 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint May 27 03:14:45.934968 kernel: pci 0000:00:04.0: BAR 0 [io 0xc1c0-0xc1df] May 27 03:14:45.935078 kernel: pci 0000:00:04.0: BAR 1 [mem 0xfebf2000-0xfebf2fff] May 27 03:14:45.935178 kernel: pci 0000:00:04.0: BAR 4 [mem 0xfe808000-0xfe80bfff 64bit pref] May 27 03:14:45.935302 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 conventional PCI endpoint May 27 03:14:45.935413 kernel: pci 0000:00:05.0: BAR 0 [io 0xc100-0xc13f] May 27 03:14:45.935511 kernel: pci 0000:00:05.0: BAR 1 [mem 0xfebf3000-0xfebf3fff] May 27 03:14:45.935608 kernel: pci 0000:00:05.0: BAR 4 [mem 0xfe80c000-0xfe80ffff 64bit pref] May 27 03:14:45.935727 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint May 27 03:14:45.935828 kernel: pci 0000:00:06.0: BAR 0 [io 0xc000-0xc07f] May 27 03:14:45.935924 kernel: pci 0000:00:06.0: BAR 1 [mem 0xfebf4000-0xfebf4fff] May 27 03:14:45.936020 kernel: pci 0000:00:06.0: BAR 4 [mem 0xfe810000-0xfe813fff 64bit pref] May 27 03:14:45.936154 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint May 27 03:14:45.936270 kernel: pci 0000:00:07.0: BAR 0 [io 0xc080-0xc0ff] May 27 03:14:45.936416 kernel: pci 0000:00:07.0: BAR 1 [mem 0xfebf5000-0xfebf5fff] May 27 03:14:45.936539 kernel: pci 0000:00:07.0: BAR 4 [mem 0xfe814000-0xfe817fff 64bit pref] May 27 03:14:45.936659 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 conventional PCI endpoint May 27 03:14:45.936808 kernel: pci 0000:00:08.0: BAR 0 [io 0xc140-0xc17f] May 27 03:14:45.936954 kernel: pci 0000:00:08.0: BAR 4 [mem 0xfe818000-0xfe81bfff 64bit pref] May 27 03:14:45.936969 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 May 27 03:14:45.936979 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 May 27 03:14:45.936988 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 May 27 03:14:45.936997 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 May 27 03:14:45.937006 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 May 27 03:14:45.937015 kernel: iommu: Default domain type: Translated May 27 03:14:45.937025 kernel: iommu: DMA domain TLB invalidation policy: lazy mode May 27 03:14:45.937034 kernel: PCI: Using ACPI for IRQ routing May 27 03:14:45.937084 kernel: PCI: pci_cache_line_size set to 64 bytes May 27 03:14:45.937093 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] May 27 03:14:45.937102 kernel: e820: reserve RAM buffer [mem 0x7ffdb000-0x7fffffff] May 27 03:14:45.937235 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device May 27 03:14:45.937333 kernel: pci 0000:00:02.0: vgaarb: bridge control possible May 27 03:14:45.937428 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none May 27 03:14:45.937441 kernel: vgaarb: loaded May 27 03:14:45.937450 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 May 27 03:14:45.937471 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter May 27 03:14:45.937479 kernel: clocksource: Switched to clocksource kvm-clock May 27 03:14:45.937489 kernel: VFS: Disk quotas dquot_6.6.0 May 27 03:14:45.937499 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 27 03:14:45.937513 kernel: pnp: PnP ACPI init May 27 03:14:45.937527 kernel: pnp: PnP ACPI: found 4 devices May 27 03:14:45.937540 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns May 27 03:14:45.937552 kernel: NET: Registered PF_INET protocol family May 27 03:14:45.937565 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) May 27 03:14:45.937589 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) May 27 03:14:45.937605 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 27 03:14:45.937620 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) May 27 03:14:45.937632 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) May 27 03:14:45.937641 kernel: TCP: Hash tables configured (established 16384 bind 16384) May 27 03:14:45.937650 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) May 27 03:14:45.937659 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) May 27 03:14:45.937668 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 27 03:14:45.937678 kernel: NET: Registered PF_XDP protocol family May 27 03:14:45.937805 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] May 27 03:14:45.937922 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] May 27 03:14:45.938226 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] May 27 03:14:45.938320 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] May 27 03:14:45.938406 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] May 27 03:14:45.938541 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release May 27 03:14:45.938646 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers May 27 03:14:45.938672 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 May 27 03:14:45.938817 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x720 took 28205 usecs May 27 03:14:45.938836 kernel: PCI: CLS 0 bytes, default 64 May 27 03:14:45.938851 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer May 27 03:14:45.938864 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x23f39654230, max_idle_ns: 440795207432 ns May 27 03:14:45.938873 kernel: Initialise system trusted keyrings May 27 03:14:45.938882 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 May 27 03:14:45.938892 kernel: Key type asymmetric registered May 27 03:14:45.938900 kernel: Asymmetric key parser 'x509' registered May 27 03:14:45.938909 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) May 27 03:14:45.938930 kernel: io scheduler mq-deadline registered May 27 03:14:45.938939 kernel: io scheduler kyber registered May 27 03:14:45.938949 kernel: io scheduler bfq registered May 27 03:14:45.938958 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 May 27 03:14:45.938967 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 May 27 03:14:45.938976 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 May 27 03:14:45.938985 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 May 27 03:14:45.938994 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 27 03:14:45.939003 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A May 27 03:14:45.939018 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 May 27 03:14:45.939027 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 May 27 03:14:45.939036 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 May 27 03:14:45.939221 kernel: rtc_cmos 00:03: RTC can wake from S4 May 27 03:14:45.939317 kernel: rtc_cmos 00:03: registered as rtc0 May 27 03:14:45.939414 kernel: rtc_cmos 00:03: setting system clock to 2025-05-27T03:14:45 UTC (1748315685) May 27 03:14:45.939426 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 May 27 03:14:45.939522 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram May 27 03:14:45.939533 kernel: intel_pstate: CPU model not supported May 27 03:14:45.939542 kernel: NET: Registered PF_INET6 protocol family May 27 03:14:45.939551 kernel: Segment Routing with IPv6 May 27 03:14:45.939561 kernel: In-situ OAM (IOAM) with IPv6 May 27 03:14:45.939573 kernel: NET: Registered PF_PACKET protocol family May 27 03:14:45.939586 kernel: Key type dns_resolver registered May 27 03:14:45.939597 kernel: IPI shorthand broadcast: enabled May 27 03:14:45.939610 kernel: sched_clock: Marking stable (3299003998, 87304075)->(3402972063, -16663990) May 27 03:14:45.939622 kernel: registered taskstats version 1 May 27 03:14:45.939647 kernel: Loading compiled-in X.509 certificates May 27 03:14:45.939656 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.30-flatcar: ba9eddccb334a70147f3ddfe4fbde029feaa991d' May 27 03:14:45.939665 kernel: Demotion targets for Node 0: null May 27 03:14:45.939674 kernel: Key type .fscrypt registered May 27 03:14:45.939683 kernel: Key type fscrypt-provisioning registered May 27 03:14:45.939733 kernel: ima: No TPM chip found, activating TPM-bypass! May 27 03:14:45.939748 kernel: ima: Allocated hash algorithm: sha1 May 27 03:14:45.939757 kernel: ima: No architecture policies found May 27 03:14:45.939778 kernel: clk: Disabling unused clocks May 27 03:14:45.939788 kernel: Warning: unable to open an initial console. May 27 03:14:45.939799 kernel: Freeing unused kernel image (initmem) memory: 54416K May 27 03:14:45.939808 kernel: Write protecting the kernel read-only data: 24576k May 27 03:14:45.939817 kernel: Freeing unused kernel image (rodata/data gap) memory: 288K May 27 03:14:45.939827 kernel: Run /init as init process May 27 03:14:45.939836 kernel: with arguments: May 27 03:14:45.939845 kernel: /init May 27 03:14:45.939861 kernel: with environment: May 27 03:14:45.939875 kernel: HOME=/ May 27 03:14:45.939884 kernel: TERM=linux May 27 03:14:45.939893 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 27 03:14:45.939905 systemd[1]: Successfully made /usr/ read-only. May 27 03:14:45.939918 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 27 03:14:45.939928 systemd[1]: Detected virtualization kvm. May 27 03:14:45.939938 systemd[1]: Detected architecture x86-64. May 27 03:14:45.939947 systemd[1]: Running in initrd. May 27 03:14:45.939963 systemd[1]: No hostname configured, using default hostname. May 27 03:14:45.939973 systemd[1]: Hostname set to . May 27 03:14:45.939982 systemd[1]: Initializing machine ID from VM UUID. May 27 03:14:45.939992 systemd[1]: Queued start job for default target initrd.target. May 27 03:14:45.940002 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 27 03:14:45.940012 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 27 03:14:45.940022 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... May 27 03:14:45.940032 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 27 03:14:45.940048 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... May 27 03:14:45.940111 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... May 27 03:14:45.940123 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... May 27 03:14:45.940138 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... May 27 03:14:45.940149 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 27 03:14:45.940160 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 27 03:14:45.940169 systemd[1]: Reached target paths.target - Path Units. May 27 03:14:45.940180 systemd[1]: Reached target slices.target - Slice Units. May 27 03:14:45.940189 systemd[1]: Reached target swap.target - Swaps. May 27 03:14:45.940199 systemd[1]: Reached target timers.target - Timer Units. May 27 03:14:45.940209 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. May 27 03:14:45.940219 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 27 03:14:45.940234 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). May 27 03:14:45.940244 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. May 27 03:14:45.940254 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 27 03:14:45.940264 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 27 03:14:45.940273 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 27 03:14:45.940283 systemd[1]: Reached target sockets.target - Socket Units. May 27 03:14:45.940293 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... May 27 03:14:45.940303 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 27 03:14:45.940323 systemd[1]: Finished network-cleanup.service - Network Cleanup. May 27 03:14:45.940337 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). May 27 03:14:45.940351 systemd[1]: Starting systemd-fsck-usr.service... May 27 03:14:45.940365 systemd[1]: Starting systemd-journald.service - Journal Service... May 27 03:14:45.940378 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 27 03:14:45.940388 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 27 03:14:45.940398 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. May 27 03:14:45.940415 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 27 03:14:45.940425 systemd[1]: Finished systemd-fsck-usr.service. May 27 03:14:45.940436 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 27 03:14:45.940494 systemd-journald[210]: Collecting audit messages is disabled. May 27 03:14:45.940538 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 27 03:14:45.940553 kernel: Bridge firewalling registered May 27 03:14:45.940568 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 27 03:14:45.940585 systemd-journald[210]: Journal started May 27 03:14:45.940625 systemd-journald[210]: Runtime Journal (/run/log/journal/7a8f55516abd40b095f542f0efd4ee1d) is 4.9M, max 39.5M, 34.6M free. May 27 03:14:45.900197 systemd-modules-load[211]: Inserted module 'overlay' May 27 03:14:45.931919 systemd-modules-load[211]: Inserted module 'br_netfilter' May 27 03:14:45.970077 systemd[1]: Started systemd-journald.service - Journal Service. May 27 03:14:45.970688 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 27 03:14:45.972098 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 27 03:14:45.975577 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 27 03:14:45.976751 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 27 03:14:45.980227 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 27 03:14:45.981547 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 27 03:14:46.006015 systemd-tmpfiles[233]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. May 27 03:14:46.009083 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 27 03:14:46.016129 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 27 03:14:46.020188 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 27 03:14:46.022409 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 27 03:14:46.025185 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... May 27 03:14:46.028576 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 27 03:14:46.056096 dracut-cmdline[250]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=f6c186658a19d5a08471ef76df75f82494b37b46908f9237b2c3cf497da860c6 May 27 03:14:46.065270 systemd-resolved[249]: Positive Trust Anchors: May 27 03:14:46.065285 systemd-resolved[249]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 27 03:14:46.065323 systemd-resolved[249]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 27 03:14:46.068716 systemd-resolved[249]: Defaulting to hostname 'linux'. May 27 03:14:46.070427 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 27 03:14:46.070935 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 27 03:14:46.176114 kernel: SCSI subsystem initialized May 27 03:14:46.187088 kernel: Loading iSCSI transport class v2.0-870. May 27 03:14:46.199099 kernel: iscsi: registered transport (tcp) May 27 03:14:46.222094 kernel: iscsi: registered transport (qla4xxx) May 27 03:14:46.223078 kernel: QLogic iSCSI HBA Driver May 27 03:14:46.255148 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 27 03:14:46.284129 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 27 03:14:46.286466 systemd[1]: Reached target network-pre.target - Preparation for Network. May 27 03:14:46.342727 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. May 27 03:14:46.344917 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... May 27 03:14:46.407112 kernel: raid6: avx2x4 gen() 17251 MB/s May 27 03:14:46.424120 kernel: raid6: avx2x2 gen() 17509 MB/s May 27 03:14:46.441350 kernel: raid6: avx2x1 gen() 13337 MB/s May 27 03:14:46.441456 kernel: raid6: using algorithm avx2x2 gen() 17509 MB/s May 27 03:14:46.459310 kernel: raid6: .... xor() 20671 MB/s, rmw enabled May 27 03:14:46.459424 kernel: raid6: using avx2x2 recovery algorithm May 27 03:14:46.490110 kernel: xor: automatically using best checksumming function avx May 27 03:14:46.699087 kernel: Btrfs loaded, zoned=no, fsverity=no May 27 03:14:46.707079 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. May 27 03:14:46.709731 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 27 03:14:46.739723 systemd-udevd[461]: Using default interface naming scheme 'v255'. May 27 03:14:46.746312 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 27 03:14:46.749361 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... May 27 03:14:46.777780 dracut-pre-trigger[467]: rd.md=0: removing MD RAID activation May 27 03:14:46.809382 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. May 27 03:14:46.811795 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 27 03:14:46.885405 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 27 03:14:46.888196 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... May 27 03:14:46.959080 kernel: virtio_blk virtio4: 1/0/0 default/read/poll queues May 27 03:14:46.963921 kernel: virtio_scsi virtio3: 2/0/0 default/read/poll queues May 27 03:14:46.966398 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) May 27 03:14:46.969083 kernel: scsi host0: Virtio SCSI HBA May 27 03:14:46.993119 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. May 27 03:14:46.993204 kernel: cryptd: max_cpu_qlen set to 1000 May 27 03:14:46.993228 kernel: GPT:9289727 != 125829119 May 27 03:14:46.995190 kernel: GPT:Alternate GPT header not at the end of the disk. May 27 03:14:46.995255 kernel: GPT:9289727 != 125829119 May 27 03:14:46.996161 kernel: GPT: Use GNU Parted to correct GPT errors. May 27 03:14:46.996239 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 27 03:14:47.034095 kernel: AES CTR mode by8 optimization enabled May 27 03:14:47.042751 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 27 03:14:47.042891 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 27 03:14:47.051814 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... May 27 03:14:47.059392 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 27 03:14:47.063653 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 27 03:14:47.072079 kernel: libata version 3.00 loaded. May 27 03:14:47.076089 kernel: virtio_blk virtio5: 1/0/0 default/read/poll queues May 27 03:14:47.080457 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 May 27 03:14:47.083484 kernel: ata_piix 0000:00:01.1: version 2.13 May 27 03:14:47.086519 kernel: ACPI: bus type USB registered May 27 03:14:47.086583 kernel: scsi host1: ata_piix May 27 03:14:47.086815 kernel: usbcore: registered new interface driver usbfs May 27 03:14:47.086829 kernel: virtio_blk virtio5: [vdb] 980 512-byte logical blocks (502 kB/490 KiB) May 27 03:14:47.089105 kernel: usbcore: registered new interface driver hub May 27 03:14:47.090682 kernel: scsi host2: ata_piix May 27 03:14:47.090782 kernel: usbcore: registered new device driver usb May 27 03:14:47.093016 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 lpm-pol 0 May 27 03:14:47.093114 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 lpm-pol 0 May 27 03:14:47.143151 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 27 03:14:47.315481 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller May 27 03:14:47.315870 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 May 27 03:14:47.324676 kernel: uhci_hcd 0000:00:01.2: detected 2 ports May 27 03:14:47.325094 kernel: uhci_hcd 0000:00:01.2: irq 11, io port 0x0000c180 May 27 03:14:47.327170 kernel: hub 1-0:1.0: USB hub found May 27 03:14:47.327487 kernel: hub 1-0:1.0: 2 ports detected May 27 03:14:47.332229 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. May 27 03:14:47.342472 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. May 27 03:14:47.352810 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 27 03:14:47.353707 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. May 27 03:14:47.361887 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. May 27 03:14:47.362397 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. May 27 03:14:47.363475 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. May 27 03:14:47.364211 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 27 03:14:47.365169 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 27 03:14:47.366791 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... May 27 03:14:47.368036 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... May 27 03:14:47.390225 disk-uuid[613]: Primary Header is updated. May 27 03:14:47.390225 disk-uuid[613]: Secondary Entries is updated. May 27 03:14:47.390225 disk-uuid[613]: Secondary Header is updated. May 27 03:14:47.397833 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. May 27 03:14:47.409105 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 27 03:14:48.421140 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 27 03:14:48.421961 disk-uuid[616]: The operation has completed successfully. May 27 03:14:48.479481 systemd[1]: disk-uuid.service: Deactivated successfully. May 27 03:14:48.479614 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. May 27 03:14:48.506647 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... May 27 03:14:48.533859 sh[632]: Success May 27 03:14:48.556776 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 27 03:14:48.556858 kernel: device-mapper: uevent: version 1.0.3 May 27 03:14:48.556877 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev May 27 03:14:48.569161 kernel: device-mapper: verity: sha256 using shash "sha256-avx2" May 27 03:14:48.624798 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. May 27 03:14:48.627957 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... May 27 03:14:48.638379 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. May 27 03:14:48.653118 kernel: BTRFS info: 'norecovery' is for compatibility only, recommended to use 'rescue=nologreplay' May 27 03:14:48.653199 kernel: BTRFS: device fsid f0f66fe8-3990-49eb-980e-559a3dfd3522 devid 1 transid 40 /dev/mapper/usr (253:0) scanned by mount (644) May 27 03:14:48.656214 kernel: BTRFS info (device dm-0): first mount of filesystem f0f66fe8-3990-49eb-980e-559a3dfd3522 May 27 03:14:48.656304 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm May 27 03:14:48.656326 kernel: BTRFS info (device dm-0): using free-space-tree May 27 03:14:48.665529 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. May 27 03:14:48.666972 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. May 27 03:14:48.667991 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. May 27 03:14:48.668931 systemd[1]: Starting ignition-setup.service - Ignition (setup)... May 27 03:14:48.673025 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... May 27 03:14:48.705103 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/vda6 (254:6) scanned by mount (674) May 27 03:14:48.708107 kernel: BTRFS info (device vda6): first mount of filesystem fd7bb961-7a0f-4c90-a609-3bffeb956d05 May 27 03:14:48.710260 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 27 03:14:48.710314 kernel: BTRFS info (device vda6): using free-space-tree May 27 03:14:48.722143 kernel: BTRFS info (device vda6): last unmount of filesystem fd7bb961-7a0f-4c90-a609-3bffeb956d05 May 27 03:14:48.723619 systemd[1]: Finished ignition-setup.service - Ignition (setup). May 27 03:14:48.726246 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... May 27 03:14:48.809599 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 27 03:14:48.814257 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 27 03:14:48.863796 systemd-networkd[814]: lo: Link UP May 27 03:14:48.864118 systemd-networkd[814]: lo: Gained carrier May 27 03:14:48.867611 systemd-networkd[814]: Enumeration completed May 27 03:14:48.867774 systemd[1]: Started systemd-networkd.service - Network Configuration. May 27 03:14:48.869783 systemd-networkd[814]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. May 27 03:14:48.869788 systemd-networkd[814]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. May 27 03:14:48.870073 systemd[1]: Reached target network.target - Network. May 27 03:14:48.871758 systemd-networkd[814]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 27 03:14:48.871762 systemd-networkd[814]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. May 27 03:14:48.873105 systemd-networkd[814]: eth0: Link UP May 27 03:14:48.873114 systemd-networkd[814]: eth0: Gained carrier May 27 03:14:48.873128 systemd-networkd[814]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. May 27 03:14:48.881601 systemd-networkd[814]: eth1: Link UP May 27 03:14:48.881614 systemd-networkd[814]: eth1: Gained carrier May 27 03:14:48.881638 systemd-networkd[814]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 27 03:14:48.902281 systemd-networkd[814]: eth1: DHCPv4 address 10.124.0.18/20 acquired from 169.254.169.253 May 27 03:14:48.912171 systemd-networkd[814]: eth0: DHCPv4 address 209.38.65.15/20, gateway 209.38.64.1 acquired from 169.254.169.253 May 27 03:14:48.927210 ignition[716]: Ignition 2.21.0 May 27 03:14:48.927998 ignition[716]: Stage: fetch-offline May 27 03:14:48.928043 ignition[716]: no configs at "/usr/lib/ignition/base.d" May 27 03:14:48.928077 ignition[716]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 27 03:14:48.928171 ignition[716]: parsed url from cmdline: "" May 27 03:14:48.928174 ignition[716]: no config URL provided May 27 03:14:48.928179 ignition[716]: reading system config file "/usr/lib/ignition/user.ign" May 27 03:14:48.931486 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). May 27 03:14:48.928187 ignition[716]: no config at "/usr/lib/ignition/user.ign" May 27 03:14:48.928192 ignition[716]: failed to fetch config: resource requires networking May 27 03:14:48.928391 ignition[716]: Ignition finished successfully May 27 03:14:48.935765 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... May 27 03:14:48.986403 ignition[823]: Ignition 2.21.0 May 27 03:14:48.986426 ignition[823]: Stage: fetch May 27 03:14:48.986647 ignition[823]: no configs at "/usr/lib/ignition/base.d" May 27 03:14:48.986658 ignition[823]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 27 03:14:48.986995 ignition[823]: parsed url from cmdline: "" May 27 03:14:48.987000 ignition[823]: no config URL provided May 27 03:14:48.987010 ignition[823]: reading system config file "/usr/lib/ignition/user.ign" May 27 03:14:48.987026 ignition[823]: no config at "/usr/lib/ignition/user.ign" May 27 03:14:48.987111 ignition[823]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 May 27 03:14:49.017561 ignition[823]: GET result: OK May 27 03:14:49.017715 ignition[823]: parsing config with SHA512: db0af898de3baac145a097455c4d4e2f6b09ce6e5e60b4b49cfd6eb6fe965d45e456463b82c3a5dedd45f68cd5d7898ff25ee9f2b586e92985f647249db04f13 May 27 03:14:49.023882 unknown[823]: fetched base config from "system" May 27 03:14:49.023895 unknown[823]: fetched base config from "system" May 27 03:14:49.024597 ignition[823]: fetch: fetch complete May 27 03:14:49.023902 unknown[823]: fetched user config from "digitalocean" May 27 03:14:49.024606 ignition[823]: fetch: fetch passed May 27 03:14:49.028221 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). May 27 03:14:49.024698 ignition[823]: Ignition finished successfully May 27 03:14:49.032278 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... May 27 03:14:49.085844 ignition[829]: Ignition 2.21.0 May 27 03:14:49.085869 ignition[829]: Stage: kargs May 27 03:14:49.086263 ignition[829]: no configs at "/usr/lib/ignition/base.d" May 27 03:14:49.086283 ignition[829]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 27 03:14:49.089841 ignition[829]: kargs: kargs passed May 27 03:14:49.089927 ignition[829]: Ignition finished successfully May 27 03:14:49.091243 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). May 27 03:14:49.093251 systemd[1]: Starting ignition-disks.service - Ignition (disks)... May 27 03:14:49.129173 ignition[835]: Ignition 2.21.0 May 27 03:14:49.129191 ignition[835]: Stage: disks May 27 03:14:49.129354 ignition[835]: no configs at "/usr/lib/ignition/base.d" May 27 03:14:49.129364 ignition[835]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 27 03:14:49.130806 ignition[835]: disks: disks passed May 27 03:14:49.130881 ignition[835]: Ignition finished successfully May 27 03:14:49.132123 systemd[1]: Finished ignition-disks.service - Ignition (disks). May 27 03:14:49.133576 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. May 27 03:14:49.133997 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. May 27 03:14:49.134764 systemd[1]: Reached target local-fs.target - Local File Systems. May 27 03:14:49.135519 systemd[1]: Reached target sysinit.target - System Initialization. May 27 03:14:49.136237 systemd[1]: Reached target basic.target - Basic System. May 27 03:14:49.138143 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... May 27 03:14:49.170984 systemd-fsck[843]: ROOT: clean, 15/553520 files, 52789/553472 blocks May 27 03:14:49.176959 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. May 27 03:14:49.179385 systemd[1]: Mounting sysroot.mount - /sysroot... May 27 03:14:49.323147 kernel: EXT4-fs (vda9): mounted filesystem 18301365-b380-45d7-9677-e42472a122bc r/w with ordered data mode. Quota mode: none. May 27 03:14:49.325151 systemd[1]: Mounted sysroot.mount - /sysroot. May 27 03:14:49.325665 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. May 27 03:14:49.328089 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 27 03:14:49.329723 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... May 27 03:14:49.332172 systemd[1]: Starting flatcar-afterburn-network.service - Flatcar Afterburn network service... May 27 03:14:49.338266 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... May 27 03:14:49.338741 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 27 03:14:49.338863 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. May 27 03:14:49.350912 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. May 27 03:14:49.362526 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... May 27 03:14:49.368376 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/vda6 (254:6) scanned by mount (851) May 27 03:14:49.376103 kernel: BTRFS info (device vda6): first mount of filesystem fd7bb961-7a0f-4c90-a609-3bffeb956d05 May 27 03:14:49.378231 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 27 03:14:49.378323 kernel: BTRFS info (device vda6): using free-space-tree May 27 03:14:49.424010 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 27 03:14:49.441526 coreos-metadata[853]: May 27 03:14:49.441 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 27 03:14:49.443905 initrd-setup-root[882]: cut: /sysroot/etc/passwd: No such file or directory May 27 03:14:49.453965 initrd-setup-root[889]: cut: /sysroot/etc/group: No such file or directory May 27 03:14:49.455039 coreos-metadata[854]: May 27 03:14:49.454 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 27 03:14:49.457116 coreos-metadata[853]: May 27 03:14:49.455 INFO Fetch successful May 27 03:14:49.463300 initrd-setup-root[896]: cut: /sysroot/etc/shadow: No such file or directory May 27 03:14:49.465181 systemd[1]: flatcar-afterburn-network.service: Deactivated successfully. May 27 03:14:49.465318 systemd[1]: Finished flatcar-afterburn-network.service - Flatcar Afterburn network service. May 27 03:14:49.467973 coreos-metadata[854]: May 27 03:14:49.467 INFO Fetch successful May 27 03:14:49.471336 initrd-setup-root[904]: cut: /sysroot/etc/gshadow: No such file or directory May 27 03:14:49.474210 coreos-metadata[854]: May 27 03:14:49.474 INFO wrote hostname ci-4344.0.0-4-f2dd3e9233 to /sysroot/etc/hostname May 27 03:14:49.475148 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. May 27 03:14:49.588805 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. May 27 03:14:49.591465 systemd[1]: Starting ignition-mount.service - Ignition (mount)... May 27 03:14:49.593015 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... May 27 03:14:49.609075 kernel: BTRFS info (device vda6): last unmount of filesystem fd7bb961-7a0f-4c90-a609-3bffeb956d05 May 27 03:14:49.627134 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. May 27 03:14:49.641736 ignition[977]: INFO : Ignition 2.21.0 May 27 03:14:49.641736 ignition[977]: INFO : Stage: mount May 27 03:14:49.644490 ignition[977]: INFO : no configs at "/usr/lib/ignition/base.d" May 27 03:14:49.644490 ignition[977]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 27 03:14:49.644490 ignition[977]: INFO : mount: mount passed May 27 03:14:49.644490 ignition[977]: INFO : Ignition finished successfully May 27 03:14:49.646286 systemd[1]: Finished ignition-mount.service - Ignition (mount). May 27 03:14:49.648763 systemd[1]: Starting ignition-files.service - Ignition (files)... May 27 03:14:49.651914 systemd[1]: sysroot-oem.mount: Deactivated successfully. May 27 03:14:49.667421 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 27 03:14:49.685094 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/vda6 (254:6) scanned by mount (989) May 27 03:14:49.687561 kernel: BTRFS info (device vda6): first mount of filesystem fd7bb961-7a0f-4c90-a609-3bffeb956d05 May 27 03:14:49.687643 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 27 03:14:49.687663 kernel: BTRFS info (device vda6): using free-space-tree May 27 03:14:49.693150 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 27 03:14:49.731966 ignition[1005]: INFO : Ignition 2.21.0 May 27 03:14:49.731966 ignition[1005]: INFO : Stage: files May 27 03:14:49.731966 ignition[1005]: INFO : no configs at "/usr/lib/ignition/base.d" May 27 03:14:49.731966 ignition[1005]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 27 03:14:49.731966 ignition[1005]: DEBUG : files: compiled without relabeling support, skipping May 27 03:14:49.735730 ignition[1005]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 27 03:14:49.735730 ignition[1005]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 27 03:14:49.738675 ignition[1005]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 27 03:14:49.739729 ignition[1005]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 27 03:14:49.741200 unknown[1005]: wrote ssh authorized keys file for user: core May 27 03:14:49.742129 ignition[1005]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 27 03:14:49.743404 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" May 27 03:14:49.744280 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 May 27 03:14:49.777736 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK May 27 03:14:49.913072 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" May 27 03:14:49.914748 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" May 27 03:14:49.914748 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" May 27 03:14:49.914748 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" May 27 03:14:49.914748 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" May 27 03:14:49.914748 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 27 03:14:49.914748 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 27 03:14:49.914748 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 27 03:14:49.914748 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 27 03:14:49.923875 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" May 27 03:14:49.923875 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 27 03:14:49.923875 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" May 27 03:14:49.923875 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" May 27 03:14:49.923875 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" May 27 03:14:49.923875 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-x86-64.raw: attempt #1 May 27 03:14:50.454667 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK May 27 03:14:50.533246 systemd-networkd[814]: eth0: Gained IPv6LL May 27 03:14:50.661602 systemd-networkd[814]: eth1: Gained IPv6LL May 27 03:14:50.742807 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" May 27 03:14:50.742807 ignition[1005]: INFO : files: op(b): [started] processing unit "prepare-helm.service" May 27 03:14:50.744802 ignition[1005]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 27 03:14:50.745437 ignition[1005]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 27 03:14:50.745437 ignition[1005]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" May 27 03:14:50.745437 ignition[1005]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" May 27 03:14:50.745437 ignition[1005]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" May 27 03:14:50.745437 ignition[1005]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" May 27 03:14:50.750265 ignition[1005]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" May 27 03:14:50.750265 ignition[1005]: INFO : files: files passed May 27 03:14:50.750265 ignition[1005]: INFO : Ignition finished successfully May 27 03:14:50.748559 systemd[1]: Finished ignition-files.service - Ignition (files). May 27 03:14:50.752218 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... May 27 03:14:50.754158 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... May 27 03:14:50.771127 systemd[1]: ignition-quench.service: Deactivated successfully. May 27 03:14:50.771969 systemd[1]: Finished ignition-quench.service - Ignition (record completion). May 27 03:14:50.779041 initrd-setup-root-after-ignition[1035]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 27 03:14:50.779041 initrd-setup-root-after-ignition[1035]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory May 27 03:14:50.781747 initrd-setup-root-after-ignition[1039]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 27 03:14:50.783782 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. May 27 03:14:50.784744 systemd[1]: Reached target ignition-complete.target - Ignition Complete. May 27 03:14:50.786245 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... May 27 03:14:50.842820 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 27 03:14:50.842943 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. May 27 03:14:50.844092 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. May 27 03:14:50.844749 systemd[1]: Reached target initrd.target - Initrd Default Target. May 27 03:14:50.845577 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. May 27 03:14:50.846652 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... May 27 03:14:50.885480 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 27 03:14:50.887452 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... May 27 03:14:50.913972 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. May 27 03:14:50.915038 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. May 27 03:14:50.915576 systemd[1]: Stopped target timers.target - Timer Units. May 27 03:14:50.916453 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 27 03:14:50.916598 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 27 03:14:50.917540 systemd[1]: Stopped target initrd.target - Initrd Default Target. May 27 03:14:50.917972 systemd[1]: Stopped target basic.target - Basic System. May 27 03:14:50.918705 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. May 27 03:14:50.919313 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. May 27 03:14:50.919944 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. May 27 03:14:50.920608 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. May 27 03:14:50.921364 systemd[1]: Stopped target remote-fs.target - Remote File Systems. May 27 03:14:50.922062 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. May 27 03:14:50.922709 systemd[1]: Stopped target sysinit.target - System Initialization. May 27 03:14:50.923341 systemd[1]: Stopped target local-fs.target - Local File Systems. May 27 03:14:50.924076 systemd[1]: Stopped target swap.target - Swaps. May 27 03:14:50.924647 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 27 03:14:50.924847 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. May 27 03:14:50.925518 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. May 27 03:14:50.925971 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 27 03:14:50.926520 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. May 27 03:14:50.926644 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 27 03:14:50.927258 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 27 03:14:50.927417 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. May 27 03:14:50.928264 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 27 03:14:50.928423 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. May 27 03:14:50.929267 systemd[1]: ignition-files.service: Deactivated successfully. May 27 03:14:50.929411 systemd[1]: Stopped ignition-files.service - Ignition (files). May 27 03:14:50.929922 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. May 27 03:14:50.930047 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. May 27 03:14:50.932129 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... May 27 03:14:50.935173 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... May 27 03:14:50.935556 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 27 03:14:50.935713 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. May 27 03:14:50.937282 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 27 03:14:50.937422 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. May 27 03:14:50.943138 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 27 03:14:50.943799 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. May 27 03:14:50.965401 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 27 03:14:50.974230 ignition[1059]: INFO : Ignition 2.21.0 May 27 03:14:50.974230 ignition[1059]: INFO : Stage: umount May 27 03:14:50.974230 ignition[1059]: INFO : no configs at "/usr/lib/ignition/base.d" May 27 03:14:50.974230 ignition[1059]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 27 03:14:50.974230 ignition[1059]: INFO : umount: umount passed May 27 03:14:50.974230 ignition[1059]: INFO : Ignition finished successfully May 27 03:14:50.974499 systemd[1]: ignition-mount.service: Deactivated successfully. May 27 03:14:50.974628 systemd[1]: Stopped ignition-mount.service - Ignition (mount). May 27 03:14:50.986172 systemd[1]: ignition-disks.service: Deactivated successfully. May 27 03:14:50.986298 systemd[1]: Stopped ignition-disks.service - Ignition (disks). May 27 03:14:50.988938 systemd[1]: ignition-kargs.service: Deactivated successfully. May 27 03:14:50.989037 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). May 27 03:14:50.989422 systemd[1]: ignition-fetch.service: Deactivated successfully. May 27 03:14:50.989468 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). May 27 03:14:50.996909 systemd[1]: Stopped target network.target - Network. May 27 03:14:50.997965 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 27 03:14:50.998083 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). May 27 03:14:50.998742 systemd[1]: Stopped target paths.target - Path Units. May 27 03:14:50.999485 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 27 03:14:50.999542 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 27 03:14:51.000165 systemd[1]: Stopped target slices.target - Slice Units. May 27 03:14:51.002257 systemd[1]: Stopped target sockets.target - Socket Units. May 27 03:14:51.002777 systemd[1]: iscsid.socket: Deactivated successfully. May 27 03:14:51.002832 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. May 27 03:14:51.003278 systemd[1]: iscsiuio.socket: Deactivated successfully. May 27 03:14:51.003332 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 27 03:14:51.003711 systemd[1]: ignition-setup.service: Deactivated successfully. May 27 03:14:51.003785 systemd[1]: Stopped ignition-setup.service - Ignition (setup). May 27 03:14:51.004183 systemd[1]: ignition-setup-pre.service: Deactivated successfully. May 27 03:14:51.004228 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. May 27 03:14:51.005285 systemd[1]: Stopping systemd-networkd.service - Network Configuration... May 27 03:14:51.006455 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... May 27 03:14:51.007311 systemd[1]: sysroot-boot.service: Deactivated successfully. May 27 03:14:51.007417 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. May 27 03:14:51.008943 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 27 03:14:51.009539 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. May 27 03:14:51.011887 systemd[1]: systemd-resolved.service: Deactivated successfully. May 27 03:14:51.012046 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. May 27 03:14:51.015575 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. May 27 03:14:51.016212 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. May 27 03:14:51.016333 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. May 27 03:14:51.018535 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. May 27 03:14:51.020231 systemd[1]: systemd-networkd.service: Deactivated successfully. May 27 03:14:51.020338 systemd[1]: Stopped systemd-networkd.service - Network Configuration. May 27 03:14:51.022833 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. May 27 03:14:51.024473 systemd[1]: Stopped target network-pre.target - Preparation for Network. May 27 03:14:51.025469 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 27 03:14:51.025514 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. May 27 03:14:51.027153 systemd[1]: Stopping network-cleanup.service - Network Cleanup... May 27 03:14:51.028604 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 27 03:14:51.028735 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 27 03:14:51.029381 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 27 03:14:51.029427 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 27 03:14:51.031500 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 27 03:14:51.031549 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. May 27 03:14:51.031915 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... May 27 03:14:51.036908 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 27 03:14:51.045469 systemd[1]: systemd-udevd.service: Deactivated successfully. May 27 03:14:51.045627 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. May 27 03:14:51.046615 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 27 03:14:51.046718 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. May 27 03:14:51.047468 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 27 03:14:51.047502 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. May 27 03:14:51.048284 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 27 03:14:51.048335 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. May 27 03:14:51.049641 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 27 03:14:51.049692 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. May 27 03:14:51.050082 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 27 03:14:51.050130 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 27 03:14:51.057258 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... May 27 03:14:51.058070 systemd[1]: systemd-network-generator.service: Deactivated successfully. May 27 03:14:51.058147 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. May 27 03:14:51.059691 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 27 03:14:51.059758 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 27 03:14:51.061244 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. May 27 03:14:51.061300 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 27 03:14:51.062687 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 27 03:14:51.063122 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. May 27 03:14:51.063982 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 27 03:14:51.064028 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 27 03:14:51.065262 systemd[1]: network-cleanup.service: Deactivated successfully. May 27 03:14:51.066220 systemd[1]: Stopped network-cleanup.service - Network Cleanup. May 27 03:14:51.074118 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 27 03:14:51.074229 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. May 27 03:14:51.075677 systemd[1]: Reached target initrd-switch-root.target - Switch Root. May 27 03:14:51.077044 systemd[1]: Starting initrd-switch-root.service - Switch Root... May 27 03:14:51.108261 systemd[1]: Switching root. May 27 03:14:51.154744 systemd-journald[210]: Journal stopped May 27 03:14:52.237020 systemd-journald[210]: Received SIGTERM from PID 1 (systemd). May 27 03:14:52.237194 kernel: SELinux: policy capability network_peer_controls=1 May 27 03:14:52.237236 kernel: SELinux: policy capability open_perms=1 May 27 03:14:52.237257 kernel: SELinux: policy capability extended_socket_class=1 May 27 03:14:52.237277 kernel: SELinux: policy capability always_check_network=0 May 27 03:14:52.237297 kernel: SELinux: policy capability cgroup_seclabel=1 May 27 03:14:52.237333 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 27 03:14:52.237351 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 27 03:14:52.237370 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 27 03:14:52.237395 kernel: SELinux: policy capability userspace_initial_context=0 May 27 03:14:52.237436 kernel: audit: type=1403 audit(1748315691.274:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 May 27 03:14:52.237461 systemd[1]: Successfully loaded SELinux policy in 39.250ms. May 27 03:14:52.237497 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 11.605ms. May 27 03:14:52.237521 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 27 03:14:52.237541 systemd[1]: Detected virtualization kvm. May 27 03:14:52.237561 systemd[1]: Detected architecture x86-64. May 27 03:14:52.237583 systemd[1]: Detected first boot. May 27 03:14:52.237605 systemd[1]: Hostname set to . May 27 03:14:52.237640 systemd[1]: Initializing machine ID from VM UUID. May 27 03:14:52.237663 zram_generator::config[1103]: No configuration found. May 27 03:14:52.237685 kernel: Guest personality initialized and is inactive May 27 03:14:52.237704 kernel: VMCI host device registered (name=vmci, major=10, minor=125) May 27 03:14:52.237725 kernel: Initialized host personality May 27 03:14:52.237747 kernel: NET: Registered PF_VSOCK protocol family May 27 03:14:52.237768 systemd[1]: Populated /etc with preset unit settings. May 27 03:14:52.237792 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. May 27 03:14:52.237834 systemd[1]: initrd-switch-root.service: Deactivated successfully. May 27 03:14:52.237856 systemd[1]: Stopped initrd-switch-root.service - Switch Root. May 27 03:14:52.237877 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. May 27 03:14:52.237905 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. May 27 03:14:52.237927 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. May 27 03:14:52.237948 systemd[1]: Created slice system-getty.slice - Slice /system/getty. May 27 03:14:52.237969 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. May 27 03:14:52.237989 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. May 27 03:14:52.238012 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. May 27 03:14:52.238047 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. May 27 03:14:52.238109 systemd[1]: Created slice user.slice - User and Session Slice. May 27 03:14:52.238131 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 27 03:14:52.238150 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 27 03:14:52.238169 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. May 27 03:14:52.238202 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. May 27 03:14:52.238242 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. May 27 03:14:52.238264 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 27 03:14:52.238284 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... May 27 03:14:52.238304 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 27 03:14:52.238328 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 27 03:14:52.238349 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. May 27 03:14:52.238371 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. May 27 03:14:52.238401 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. May 27 03:14:52.238418 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. May 27 03:14:52.238452 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 27 03:14:52.238473 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 27 03:14:52.238492 systemd[1]: Reached target slices.target - Slice Units. May 27 03:14:52.238511 systemd[1]: Reached target swap.target - Swaps. May 27 03:14:52.238533 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. May 27 03:14:52.238554 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. May 27 03:14:52.238576 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. May 27 03:14:52.238598 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 27 03:14:52.238620 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 27 03:14:52.238655 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 27 03:14:52.238678 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. May 27 03:14:52.238700 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... May 27 03:14:52.238722 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... May 27 03:14:52.238744 systemd[1]: Mounting media.mount - External Media Directory... May 27 03:14:52.238768 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:14:52.238790 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... May 27 03:14:52.238811 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... May 27 03:14:52.238833 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... May 27 03:14:52.238870 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 27 03:14:52.238892 systemd[1]: Reached target machines.target - Containers. May 27 03:14:52.238914 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... May 27 03:14:52.238934 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 27 03:14:52.238956 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 27 03:14:52.238978 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... May 27 03:14:52.239000 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 27 03:14:52.239022 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 27 03:14:52.239119 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 27 03:14:52.239147 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... May 27 03:14:52.239170 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 27 03:14:52.239192 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 27 03:14:52.239213 systemd[1]: systemd-fsck-root.service: Deactivated successfully. May 27 03:14:52.239235 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. May 27 03:14:52.239257 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. May 27 03:14:52.239279 systemd[1]: Stopped systemd-fsck-usr.service. May 27 03:14:52.239302 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 27 03:14:52.239339 systemd[1]: Starting systemd-journald.service - Journal Service... May 27 03:14:52.239361 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 27 03:14:52.239396 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 27 03:14:52.239419 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... May 27 03:14:52.239441 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... May 27 03:14:52.239475 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 27 03:14:52.239497 systemd[1]: verity-setup.service: Deactivated successfully. May 27 03:14:52.239518 systemd[1]: Stopped verity-setup.service. May 27 03:14:52.239560 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:14:52.239584 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. May 27 03:14:52.239618 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. May 27 03:14:52.239640 systemd[1]: Mounted media.mount - External Media Directory. May 27 03:14:52.239663 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. May 27 03:14:52.239685 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. May 27 03:14:52.239708 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. May 27 03:14:52.239730 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 27 03:14:52.239765 kernel: fuse: init (API version 7.41) May 27 03:14:52.239787 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 27 03:14:52.239809 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. May 27 03:14:52.239862 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 27 03:14:52.239936 systemd-journald[1175]: Collecting audit messages is disabled. May 27 03:14:52.239980 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 27 03:14:52.240003 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 27 03:14:52.240025 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 27 03:14:52.240047 systemd-journald[1175]: Journal started May 27 03:14:52.240150 systemd-journald[1175]: Runtime Journal (/run/log/journal/7a8f55516abd40b095f542f0efd4ee1d) is 4.9M, max 39.5M, 34.6M free. May 27 03:14:51.926566 systemd[1]: Queued start job for default target multi-user.target. May 27 03:14:51.951989 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. May 27 03:14:51.952599 systemd[1]: systemd-journald.service: Deactivated successfully. May 27 03:14:52.242089 systemd[1]: Started systemd-journald.service - Journal Service. May 27 03:14:52.246365 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 27 03:14:52.247277 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. May 27 03:14:52.249185 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 27 03:14:52.252440 kernel: loop: module loaded May 27 03:14:52.250924 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. May 27 03:14:52.253601 systemd[1]: modprobe@loop.service: Deactivated successfully. May 27 03:14:52.254606 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 27 03:14:52.301508 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... May 27 03:14:52.307418 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... May 27 03:14:52.310189 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 27 03:14:52.310249 systemd[1]: Reached target local-fs.target - Local File Systems. May 27 03:14:52.315153 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. May 27 03:14:52.322386 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... May 27 03:14:52.324323 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 27 03:14:52.352390 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... May 27 03:14:52.360911 kernel: ACPI: bus type drm_connector registered May 27 03:14:52.360316 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... May 27 03:14:52.361247 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 27 03:14:52.373467 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... May 27 03:14:52.374260 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 27 03:14:52.377424 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 27 03:14:52.388252 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... May 27 03:14:52.398764 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 27 03:14:52.404598 systemd[1]: modprobe@drm.service: Deactivated successfully. May 27 03:14:52.405289 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 27 03:14:52.406831 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 27 03:14:52.407796 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. May 27 03:14:52.412498 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. May 27 03:14:52.420004 systemd[1]: Reached target network-pre.target - Preparation for Network. May 27 03:14:52.462550 kernel: loop0: detected capacity change from 0 to 113872 May 27 03:14:52.472864 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. May 27 03:14:52.489937 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. May 27 03:14:52.506644 systemd-journald[1175]: Time spent on flushing to /var/log/journal/7a8f55516abd40b095f542f0efd4ee1d is 183.280ms for 1008 entries. May 27 03:14:52.506644 systemd-journald[1175]: System Journal (/var/log/journal/7a8f55516abd40b095f542f0efd4ee1d) is 8M, max 195.6M, 187.6M free. May 27 03:14:52.707616 systemd-journald[1175]: Received client request to flush runtime journal. May 27 03:14:52.707830 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 27 03:14:52.707867 kernel: loop1: detected capacity change from 0 to 146240 May 27 03:14:52.707890 kernel: loop2: detected capacity change from 0 to 221472 May 27 03:14:52.500081 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. May 27 03:14:52.508280 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... May 27 03:14:52.575551 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 27 03:14:52.599746 systemd-tmpfiles[1218]: ACLs are not supported, ignoring. May 27 03:14:52.599774 systemd-tmpfiles[1218]: ACLs are not supported, ignoring. May 27 03:14:52.619659 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. May 27 03:14:52.626137 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 27 03:14:52.656182 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 27 03:14:52.658441 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. May 27 03:14:52.666338 systemd[1]: Starting systemd-sysusers.service - Create System Users... May 27 03:14:52.714684 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. May 27 03:14:52.733966 kernel: loop3: detected capacity change from 0 to 8 May 27 03:14:52.747322 systemd[1]: Finished systemd-sysusers.service - Create System Users. May 27 03:14:52.752265 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 27 03:14:52.768683 kernel: loop4: detected capacity change from 0 to 113872 May 27 03:14:52.796109 kernel: loop5: detected capacity change from 0 to 146240 May 27 03:14:52.828081 kernel: loop6: detected capacity change from 0 to 221472 May 27 03:14:52.846092 kernel: loop7: detected capacity change from 0 to 8 May 27 03:14:52.849406 (sd-merge)[1253]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-digitalocean'. May 27 03:14:52.850680 (sd-merge)[1253]: Merged extensions into '/usr'. May 27 03:14:52.858008 systemd-tmpfiles[1252]: ACLs are not supported, ignoring. May 27 03:14:52.858045 systemd-tmpfiles[1252]: ACLs are not supported, ignoring. May 27 03:14:52.860488 systemd[1]: Reload requested from client PID 1217 ('systemd-sysext') (unit systemd-sysext.service)... May 27 03:14:52.860622 systemd[1]: Reloading... May 27 03:14:53.082138 zram_generator::config[1281]: No configuration found. May 27 03:14:53.289915 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 27 03:14:53.413088 ldconfig[1209]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 27 03:14:53.476041 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 27 03:14:53.476772 systemd[1]: Reloading finished in 614 ms. May 27 03:14:53.498030 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. May 27 03:14:53.499955 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 27 03:14:53.501516 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. May 27 03:14:53.517343 systemd[1]: Starting ensure-sysext.service... May 27 03:14:53.522388 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 27 03:14:53.571276 systemd[1]: Reload requested from client PID 1325 ('systemctl') (unit ensure-sysext.service)... May 27 03:14:53.571310 systemd[1]: Reloading... May 27 03:14:53.592220 systemd-tmpfiles[1326]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. May 27 03:14:53.592280 systemd-tmpfiles[1326]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. May 27 03:14:53.592693 systemd-tmpfiles[1326]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 27 03:14:53.593115 systemd-tmpfiles[1326]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. May 27 03:14:53.594477 systemd-tmpfiles[1326]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 27 03:14:53.594916 systemd-tmpfiles[1326]: ACLs are not supported, ignoring. May 27 03:14:53.595021 systemd-tmpfiles[1326]: ACLs are not supported, ignoring. May 27 03:14:53.602166 systemd-tmpfiles[1326]: Detected autofs mount point /boot during canonicalization of boot. May 27 03:14:53.602390 systemd-tmpfiles[1326]: Skipping /boot May 27 03:14:53.628115 systemd-tmpfiles[1326]: Detected autofs mount point /boot during canonicalization of boot. May 27 03:14:53.628306 systemd-tmpfiles[1326]: Skipping /boot May 27 03:14:53.759100 zram_generator::config[1353]: No configuration found. May 27 03:14:53.947024 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 27 03:14:54.106615 systemd[1]: Reloading finished in 534 ms. May 27 03:14:54.126599 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. May 27 03:14:54.133849 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 27 03:14:54.143245 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 27 03:14:54.146644 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... May 27 03:14:54.150367 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... May 27 03:14:54.154714 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 27 03:14:54.165336 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 27 03:14:54.167380 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... May 27 03:14:54.175470 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:14:54.175719 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 27 03:14:54.177380 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 27 03:14:54.183915 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 27 03:14:54.190489 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 27 03:14:54.191085 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 27 03:14:54.191221 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 27 03:14:54.191347 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:14:54.198516 systemd[1]: Starting systemd-userdbd.service - User Database Manager... May 27 03:14:54.201794 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:14:54.201992 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 27 03:14:54.202276 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 27 03:14:54.202381 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 27 03:14:54.202468 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:14:54.206905 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:14:54.208243 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 27 03:14:54.221517 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 27 03:14:54.223318 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 27 03:14:54.223472 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 27 03:14:54.223611 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:14:54.232303 systemd[1]: Finished ensure-sysext.service. May 27 03:14:54.237453 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... May 27 03:14:54.247186 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. May 27 03:14:54.284592 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. May 27 03:14:54.289691 systemd[1]: Starting systemd-update-done.service - Update is Completed... May 27 03:14:54.300494 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. May 27 03:14:54.301891 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 27 03:14:54.308677 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 27 03:14:54.309547 systemd[1]: modprobe@loop.service: Deactivated successfully. May 27 03:14:54.311185 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 27 03:14:54.314752 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 27 03:14:54.315385 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 27 03:14:54.317897 systemd[1]: modprobe@drm.service: Deactivated successfully. May 27 03:14:54.320363 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 27 03:14:54.321755 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 27 03:14:54.321826 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 27 03:14:54.321852 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 27 03:14:54.323331 systemd-udevd[1403]: Using default interface naming scheme 'v255'. May 27 03:14:54.346924 systemd[1]: Finished systemd-update-done.service - Update is Completed. May 27 03:14:54.357097 augenrules[1440]: No rules May 27 03:14:54.358523 systemd[1]: audit-rules.service: Deactivated successfully. May 27 03:14:54.358851 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 27 03:14:54.365424 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 27 03:14:54.371376 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 27 03:14:54.395476 systemd[1]: Started systemd-userdbd.service - User Database Manager. May 27 03:14:54.526693 systemd[1]: Condition check resulted in dev-disk-by\x2dlabel-config\x2d2.device - /dev/disk/by-label/config-2 being skipped. May 27 03:14:54.529190 systemd[1]: Mounting media-configdrive.mount - /media/configdrive... May 27 03:14:54.529724 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:14:54.529971 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 27 03:14:54.532676 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 27 03:14:54.534353 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 27 03:14:54.538371 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 27 03:14:54.538839 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 27 03:14:54.539950 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 27 03:14:54.539989 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 27 03:14:54.540007 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:14:54.587090 kernel: ISO 9660 Extensions: RRIP_1991A May 27 03:14:54.599887 systemd[1]: Mounted media-configdrive.mount - /media/configdrive. May 27 03:14:54.601012 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 27 03:14:54.602224 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 27 03:14:54.602958 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 27 03:14:54.603160 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 27 03:14:54.603793 systemd[1]: modprobe@loop.service: Deactivated successfully. May 27 03:14:54.604102 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 27 03:14:54.609938 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 27 03:14:54.610141 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 27 03:14:54.690218 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. May 27 03:14:54.739229 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 27 03:14:54.742534 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... May 27 03:14:54.800112 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. May 27 03:14:54.835928 systemd-networkd[1448]: lo: Link UP May 27 03:14:54.835940 systemd-networkd[1448]: lo: Gained carrier May 27 03:14:54.839627 systemd-networkd[1448]: Enumeration completed May 27 03:14:54.839800 systemd[1]: Started systemd-networkd.service - Network Configuration. May 27 03:14:54.840168 systemd-networkd[1448]: eth0: Configuring with /run/systemd/network/10-26:dd:92:88:d4:5e.network. May 27 03:14:54.841008 systemd-networkd[1448]: eth1: Configuring with /run/systemd/network/10-4e:82:bf:90:15:90.network. May 27 03:14:54.841597 systemd-networkd[1448]: eth0: Link UP May 27 03:14:54.841790 systemd-networkd[1448]: eth0: Gained carrier May 27 03:14:54.846080 kernel: mousedev: PS/2 mouse device common for all mice May 27 03:14:54.847232 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... May 27 03:14:54.848523 systemd-networkd[1448]: eth1: Link UP May 27 03:14:54.849146 systemd-networkd[1448]: eth1: Gained carrier May 27 03:14:54.851083 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... May 27 03:14:54.924203 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. May 27 03:14:54.927241 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. May 27 03:14:54.928132 systemd[1]: Reached target time-set.target - System Time Set. May 27 03:14:54.934331 systemd-resolved[1402]: Positive Trust Anchors: May 27 03:14:54.934669 systemd-resolved[1402]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 27 03:14:54.934899 systemd-resolved[1402]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 27 03:14:54.941117 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 May 27 03:14:54.942342 systemd-resolved[1402]: Using system hostname 'ci-4344.0.0-4-f2dd3e9233'. May 27 03:14:54.946000 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 27 03:14:54.946731 systemd[1]: Reached target network.target - Network. May 27 03:14:54.951376 systemd-timesyncd[1417]: Contacted time server 65.100.46.164:123 (0.flatcar.pool.ntp.org). May 27 03:14:54.951454 systemd-timesyncd[1417]: Initial clock synchronization to Tue 2025-05-27 03:14:54.951815 UTC. May 27 03:14:54.951819 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 27 03:14:54.952496 systemd[1]: Reached target sysinit.target - System Initialization. May 27 03:14:54.953356 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. May 27 03:14:54.954266 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. May 27 03:14:54.955071 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. May 27 03:14:54.955739 systemd[1]: Started logrotate.timer - Daily rotation of log files. May 27 03:14:54.956825 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. May 27 03:14:54.957338 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. May 27 03:14:54.958008 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 27 03:14:54.958046 systemd[1]: Reached target paths.target - Path Units. May 27 03:14:54.958968 systemd[1]: Reached target timers.target - Timer Units. May 27 03:14:54.960975 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. May 27 03:14:54.964030 systemd[1]: Starting docker.socket - Docker Socket for the API... May 27 03:14:54.971665 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). May 27 03:14:54.973190 kernel: ACPI: button: Power Button [PWRF] May 27 03:14:54.973372 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). May 27 03:14:54.973960 systemd[1]: Reached target ssh-access.target - SSH Access Available. May 27 03:14:54.983622 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. May 27 03:14:54.985272 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. May 27 03:14:54.987446 systemd[1]: Listening on docker.socket - Docker Socket for the API. May 27 03:14:54.990018 systemd[1]: Reached target sockets.target - Socket Units. May 27 03:14:54.991194 systemd[1]: Reached target basic.target - Basic System. May 27 03:14:54.992044 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. May 27 03:14:54.992106 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. May 27 03:14:54.994462 systemd[1]: Starting containerd.service - containerd container runtime... May 27 03:14:54.998399 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... May 27 03:14:55.003398 systemd[1]: Starting dbus.service - D-Bus System Message Bus... May 27 03:14:55.009470 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... May 27 03:14:55.015098 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 May 27 03:14:55.017505 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... May 27 03:14:55.021083 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD May 27 03:14:55.024481 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... May 27 03:14:55.025089 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). May 27 03:14:55.030400 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... May 27 03:14:55.040558 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... May 27 03:14:55.052219 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... May 27 03:14:55.066650 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... May 27 03:14:55.077461 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... May 27 03:14:55.090378 systemd[1]: Starting systemd-logind.service - User Login Management... May 27 03:14:55.092803 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 27 03:14:55.093573 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. May 27 03:14:55.101331 systemd[1]: Starting update-engine.service - Update Engine... May 27 03:14:55.106707 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... May 27 03:14:55.112260 jq[1513]: false May 27 03:14:55.121208 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. May 27 03:14:55.122859 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 27 03:14:55.131385 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. May 27 03:14:55.132989 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 27 03:14:55.133400 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. May 27 03:14:55.153104 google_oslogin_nss_cache[1515]: oslogin_cache_refresh[1515]: Refreshing passwd entry cache May 27 03:14:55.150162 oslogin_cache_refresh[1515]: Refreshing passwd entry cache May 27 03:14:55.177159 google_oslogin_nss_cache[1515]: oslogin_cache_refresh[1515]: Failure getting users, quitting May 27 03:14:55.177159 google_oslogin_nss_cache[1515]: oslogin_cache_refresh[1515]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. May 27 03:14:55.177159 google_oslogin_nss_cache[1515]: oslogin_cache_refresh[1515]: Refreshing group entry cache May 27 03:14:55.175764 oslogin_cache_refresh[1515]: Failure getting users, quitting May 27 03:14:55.175787 oslogin_cache_refresh[1515]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. May 27 03:14:55.175854 oslogin_cache_refresh[1515]: Refreshing group entry cache May 27 03:14:55.187274 google_oslogin_nss_cache[1515]: oslogin_cache_refresh[1515]: Failure getting groups, quitting May 27 03:14:55.187379 extend-filesystems[1514]: Found loop4 May 27 03:14:55.187379 extend-filesystems[1514]: Found loop5 May 27 03:14:55.187379 extend-filesystems[1514]: Found loop6 May 27 03:14:55.187379 extend-filesystems[1514]: Found loop7 May 27 03:14:55.187379 extend-filesystems[1514]: Found vda May 27 03:14:55.187379 extend-filesystems[1514]: Found vda1 May 27 03:14:55.187379 extend-filesystems[1514]: Found vda2 May 27 03:14:55.187379 extend-filesystems[1514]: Found vda3 May 27 03:14:55.187379 extend-filesystems[1514]: Found usr May 27 03:14:55.187379 extend-filesystems[1514]: Found vda4 May 27 03:14:55.187379 extend-filesystems[1514]: Found vda6 May 27 03:14:55.187379 extend-filesystems[1514]: Found vda7 May 27 03:14:55.187379 extend-filesystems[1514]: Found vda9 May 27 03:14:55.187379 extend-filesystems[1514]: Checking size of /dev/vda9 May 27 03:14:55.193953 systemd[1]: google-oslogin-cache.service: Deactivated successfully. May 27 03:14:55.191569 oslogin_cache_refresh[1515]: Failure getting groups, quitting May 27 03:14:55.241583 google_oslogin_nss_cache[1515]: oslogin_cache_refresh[1515]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. May 27 03:14:55.241623 jq[1525]: true May 27 03:14:55.197265 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. May 27 03:14:55.191603 oslogin_cache_refresh[1515]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. May 27 03:14:55.206768 (ntainerd)[1545]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR May 27 03:14:55.271755 update_engine[1524]: I20250527 03:14:55.269970 1524 main.cc:92] Flatcar Update Engine starting May 27 03:14:55.265591 systemd[1]: motdgen.service: Deactivated successfully. May 27 03:14:55.267179 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. May 27 03:14:55.284266 tar[1533]: linux-amd64/helm May 27 03:14:55.302148 extend-filesystems[1514]: Resized partition /dev/vda9 May 27 03:14:55.304364 dbus-daemon[1511]: [system] SELinux support is enabled May 27 03:14:55.304581 systemd[1]: Started dbus.service - D-Bus System Message Bus. May 27 03:14:55.309875 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 27 03:14:55.314469 coreos-metadata[1510]: May 27 03:14:55.314 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 27 03:14:55.311358 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. May 27 03:14:55.311842 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 27 03:14:55.311923 systemd[1]: user-configdrive.service - Load cloud-config from /media/configdrive was skipped because of an unmet condition check (ConditionKernelCommandLine=!flatcar.oem.id=digitalocean). May 27 03:14:55.311942 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. May 27 03:14:55.326629 extend-filesystems[1561]: resize2fs 1.47.2 (1-Jan-2025) May 27 03:14:55.337504 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 15121403 blocks May 27 03:14:55.346111 jq[1555]: true May 27 03:14:55.355691 systemd[1]: Started update-engine.service - Update Engine. May 27 03:14:55.358938 update_engine[1524]: I20250527 03:14:55.358722 1524 update_check_scheduler.cc:74] Next update check in 4m20s May 27 03:14:55.365942 coreos-metadata[1510]: May 27 03:14:55.365 INFO Fetch successful May 27 03:14:55.397689 systemd[1]: Started locksmithd.service - Cluster reboot manager. May 27 03:14:55.513542 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. May 27 03:14:55.516045 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. May 27 03:14:55.618229 bash[1587]: Updated "/home/core/.ssh/authorized_keys" May 27 03:14:55.623762 kernel: EXT4-fs (vda9): resized filesystem to 15121403 May 27 03:14:55.621368 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. May 27 03:14:55.632449 systemd[1]: Starting sshkeys.service... May 27 03:14:55.659196 extend-filesystems[1561]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required May 27 03:14:55.659196 extend-filesystems[1561]: old_desc_blocks = 1, new_desc_blocks = 8 May 27 03:14:55.659196 extend-filesystems[1561]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. May 27 03:14:55.663782 extend-filesystems[1514]: Resized filesystem in /dev/vda9 May 27 03:14:55.663782 extend-filesystems[1514]: Found vdb May 27 03:14:55.661748 systemd[1]: extend-filesystems.service: Deactivated successfully. May 27 03:14:55.661999 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. May 27 03:14:55.680510 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. May 27 03:14:55.684444 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... May 27 03:14:55.687768 systemd-logind[1522]: New seat seat0. May 27 03:14:55.697525 systemd[1]: Started systemd-logind.service - User Login Management. May 27 03:14:55.792239 coreos-metadata[1596]: May 27 03:14:55.791 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 27 03:14:55.811104 coreos-metadata[1596]: May 27 03:14:55.810 INFO Fetch successful May 27 03:14:55.835068 unknown[1596]: wrote ssh authorized keys file for user: core May 27 03:14:55.892532 update-ssh-keys[1604]: Updated "/home/core/.ssh/authorized_keys" May 27 03:14:55.889620 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). May 27 03:14:55.892177 systemd[1]: Finished sshkeys.service. May 27 03:14:55.905687 containerd[1545]: time="2025-05-27T03:14:55Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 May 27 03:14:55.909080 containerd[1545]: time="2025-05-27T03:14:55.908708752Z" level=info msg="starting containerd" revision=06b99ca80cdbfbc6cc8bd567021738c9af2b36ce version=v2.0.4 May 27 03:14:55.919167 locksmithd[1564]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 27 03:14:55.932597 containerd[1545]: time="2025-05-27T03:14:55.932545922Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="9.539µs" May 27 03:14:55.934075 containerd[1545]: time="2025-05-27T03:14:55.932734522Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 May 27 03:14:55.934075 containerd[1545]: time="2025-05-27T03:14:55.932764343Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 May 27 03:14:55.934075 containerd[1545]: time="2025-05-27T03:14:55.932947876Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 May 27 03:14:55.934075 containerd[1545]: time="2025-05-27T03:14:55.932966242Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 May 27 03:14:55.934075 containerd[1545]: time="2025-05-27T03:14:55.932991872Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 27 03:14:55.934075 containerd[1545]: time="2025-05-27T03:14:55.933045679Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 27 03:14:55.934075 containerd[1545]: time="2025-05-27T03:14:55.933071942Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 27 03:14:55.934075 containerd[1545]: time="2025-05-27T03:14:55.933323058Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 27 03:14:55.934075 containerd[1545]: time="2025-05-27T03:14:55.933340649Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 27 03:14:55.934075 containerd[1545]: time="2025-05-27T03:14:55.933352086Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 27 03:14:55.934075 containerd[1545]: time="2025-05-27T03:14:55.933359936Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 May 27 03:14:55.934075 containerd[1545]: time="2025-05-27T03:14:55.933453997Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 May 27 03:14:55.934514 containerd[1545]: time="2025-05-27T03:14:55.933660525Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 27 03:14:55.934514 containerd[1545]: time="2025-05-27T03:14:55.933688704Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 27 03:14:55.934514 containerd[1545]: time="2025-05-27T03:14:55.933698489Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 May 27 03:14:55.934514 containerd[1545]: time="2025-05-27T03:14:55.933755056Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 May 27 03:14:55.937589 containerd[1545]: time="2025-05-27T03:14:55.937288838Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 May 27 03:14:55.937589 containerd[1545]: time="2025-05-27T03:14:55.937422617Z" level=info msg="metadata content store policy set" policy=shared May 27 03:14:55.941047 containerd[1545]: time="2025-05-27T03:14:55.939517651Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 May 27 03:14:55.941047 containerd[1545]: time="2025-05-27T03:14:55.939570495Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 May 27 03:14:55.941047 containerd[1545]: time="2025-05-27T03:14:55.939585001Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 May 27 03:14:55.941047 containerd[1545]: time="2025-05-27T03:14:55.939599751Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 May 27 03:14:55.941047 containerd[1545]: time="2025-05-27T03:14:55.939611674Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 May 27 03:14:55.941047 containerd[1545]: time="2025-05-27T03:14:55.939634687Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 May 27 03:14:55.941047 containerd[1545]: time="2025-05-27T03:14:55.939659686Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 May 27 03:14:55.941047 containerd[1545]: time="2025-05-27T03:14:55.939672794Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 May 27 03:14:55.941047 containerd[1545]: time="2025-05-27T03:14:55.939711168Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 May 27 03:14:55.941047 containerd[1545]: time="2025-05-27T03:14:55.939734279Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 May 27 03:14:55.941047 containerd[1545]: time="2025-05-27T03:14:55.939745862Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 May 27 03:14:55.941047 containerd[1545]: time="2025-05-27T03:14:55.939758851Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 May 27 03:14:55.941047 containerd[1545]: time="2025-05-27T03:14:55.939893050Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 May 27 03:14:55.941047 containerd[1545]: time="2025-05-27T03:14:55.939912158Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 May 27 03:14:55.941526 containerd[1545]: time="2025-05-27T03:14:55.939926402Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 May 27 03:14:55.941526 containerd[1545]: time="2025-05-27T03:14:55.939937999Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 May 27 03:14:55.941526 containerd[1545]: time="2025-05-27T03:14:55.939947845Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 May 27 03:14:55.941526 containerd[1545]: time="2025-05-27T03:14:55.939959150Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 May 27 03:14:55.941526 containerd[1545]: time="2025-05-27T03:14:55.939970308Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 May 27 03:14:55.941526 containerd[1545]: time="2025-05-27T03:14:55.939979482Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 May 27 03:14:55.941526 containerd[1545]: time="2025-05-27T03:14:55.939990971Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 May 27 03:14:55.941526 containerd[1545]: time="2025-05-27T03:14:55.940028201Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 May 27 03:14:55.941526 containerd[1545]: time="2025-05-27T03:14:55.940066302Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 May 27 03:14:55.941526 containerd[1545]: time="2025-05-27T03:14:55.940136938Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" May 27 03:14:55.941526 containerd[1545]: time="2025-05-27T03:14:55.940149901Z" level=info msg="Start snapshots syncer" May 27 03:14:55.941526 containerd[1545]: time="2025-05-27T03:14:55.940317085Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 May 27 03:14:55.941858 containerd[1545]: time="2025-05-27T03:14:55.940623356Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" May 27 03:14:55.941858 containerd[1545]: time="2025-05-27T03:14:55.940770117Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 May 27 03:14:55.944071 containerd[1545]: time="2025-05-27T03:14:55.944017923Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 May 27 03:14:55.948076 containerd[1545]: time="2025-05-27T03:14:55.946251975Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 May 27 03:14:55.948076 containerd[1545]: time="2025-05-27T03:14:55.946298173Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 May 27 03:14:55.948076 containerd[1545]: time="2025-05-27T03:14:55.946315343Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 May 27 03:14:55.948076 containerd[1545]: time="2025-05-27T03:14:55.946328066Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 May 27 03:14:55.948076 containerd[1545]: time="2025-05-27T03:14:55.946342591Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 May 27 03:14:55.948076 containerd[1545]: time="2025-05-27T03:14:55.946353859Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 May 27 03:14:55.948076 containerd[1545]: time="2025-05-27T03:14:55.946367924Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 May 27 03:14:55.948076 containerd[1545]: time="2025-05-27T03:14:55.946398097Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 May 27 03:14:55.948076 containerd[1545]: time="2025-05-27T03:14:55.946408439Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 May 27 03:14:55.948076 containerd[1545]: time="2025-05-27T03:14:55.946420233Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 May 27 03:14:55.948076 containerd[1545]: time="2025-05-27T03:14:55.946457670Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 27 03:14:55.948076 containerd[1545]: time="2025-05-27T03:14:55.946477807Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 27 03:14:55.948076 containerd[1545]: time="2025-05-27T03:14:55.946490662Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 27 03:14:55.948585 containerd[1545]: time="2025-05-27T03:14:55.946503439Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 27 03:14:55.948585 containerd[1545]: time="2025-05-27T03:14:55.946514075Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 May 27 03:14:55.948585 containerd[1545]: time="2025-05-27T03:14:55.946524005Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 May 27 03:14:55.948585 containerd[1545]: time="2025-05-27T03:14:55.946534096Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 May 27 03:14:55.948585 containerd[1545]: time="2025-05-27T03:14:55.946552540Z" level=info msg="runtime interface created" May 27 03:14:55.948585 containerd[1545]: time="2025-05-27T03:14:55.946558299Z" level=info msg="created NRI interface" May 27 03:14:55.948585 containerd[1545]: time="2025-05-27T03:14:55.946566494Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 May 27 03:14:55.948585 containerd[1545]: time="2025-05-27T03:14:55.946581881Z" level=info msg="Connect containerd service" May 27 03:14:55.948585 containerd[1545]: time="2025-05-27T03:14:55.946615690Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" May 27 03:14:55.948585 containerd[1545]: time="2025-05-27T03:14:55.947863316Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 27 03:14:56.223331 containerd[1545]: time="2025-05-27T03:14:56.223214980Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 27 03:14:56.223548 containerd[1545]: time="2025-05-27T03:14:56.223526447Z" level=info msg=serving... address=/run/containerd/containerd.sock May 27 03:14:56.223633 containerd[1545]: time="2025-05-27T03:14:56.223619904Z" level=info msg="Start subscribing containerd event" May 27 03:14:56.223725 containerd[1545]: time="2025-05-27T03:14:56.223709326Z" level=info msg="Start recovering state" May 27 03:14:56.223889 containerd[1545]: time="2025-05-27T03:14:56.223872915Z" level=info msg="Start event monitor" May 27 03:14:56.223996 containerd[1545]: time="2025-05-27T03:14:56.223980537Z" level=info msg="Start cni network conf syncer for default" May 27 03:14:56.224126 containerd[1545]: time="2025-05-27T03:14:56.224109013Z" level=info msg="Start streaming server" May 27 03:14:56.224198 containerd[1545]: time="2025-05-27T03:14:56.224185011Z" level=info msg="Registered namespace \"k8s.io\" with NRI" May 27 03:14:56.224262 containerd[1545]: time="2025-05-27T03:14:56.224249541Z" level=info msg="runtime interface starting up..." May 27 03:14:56.224322 containerd[1545]: time="2025-05-27T03:14:56.224310792Z" level=info msg="starting plugins..." May 27 03:14:56.224399 containerd[1545]: time="2025-05-27T03:14:56.224387808Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" May 27 03:14:56.224664 containerd[1545]: time="2025-05-27T03:14:56.224639579Z" level=info msg="containerd successfully booted in 0.319403s" May 27 03:14:56.224954 systemd[1]: Started containerd.service - containerd container runtime. May 27 03:14:56.231178 systemd-networkd[1448]: eth0: Gained IPv6LL May 27 03:14:56.235524 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. May 27 03:14:56.240772 systemd[1]: Reached target network-online.target - Network is Online. May 27 03:14:56.247552 systemd-logind[1522]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) May 27 03:14:56.251151 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 27 03:14:56.258381 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... May 27 03:14:56.269938 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 27 03:14:56.282518 kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 May 27 03:14:56.282599 kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console May 27 03:14:56.286374 kernel: Console: switching to colour dummy device 80x25 May 27 03:14:56.287125 kernel: [drm] features: -virgl +edid -resource_blob -host_visible May 27 03:14:56.287149 kernel: [drm] features: -context_init May 27 03:14:56.288143 kernel: [drm] number of scanouts: 1 May 27 03:14:56.288176 kernel: [drm] number of cap sets: 0 May 27 03:14:56.289156 kernel: [drm] Initialized virtio_gpu 0.1.0 for 0000:00:02.0 on minor 0 May 27 03:14:56.442805 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. May 27 03:14:56.486175 systemd-networkd[1448]: eth1: Gained IPv6LL May 27 03:14:56.525557 systemd-logind[1522]: Watching system buttons on /dev/input/event2 (Power Button) May 27 03:14:56.534848 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 27 03:14:56.599441 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 27 03:14:56.599768 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 27 03:14:56.601005 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... May 27 03:14:56.605415 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 27 03:14:56.608224 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 27 03:14:56.733622 kernel: EDAC MC: Ver: 3.0.0 May 27 03:14:56.753714 (udev-worker)[1467]: could not read from '/sys/module/sb_edac/initstate': No such device May 27 03:14:56.771209 sshd_keygen[1540]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 27 03:14:56.799736 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 27 03:14:56.857647 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. May 27 03:14:56.864462 systemd[1]: Starting issuegen.service - Generate /run/issue... May 27 03:14:56.909848 systemd[1]: issuegen.service: Deactivated successfully. May 27 03:14:56.910659 systemd[1]: Finished issuegen.service - Generate /run/issue. May 27 03:14:56.914819 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... May 27 03:14:56.934806 tar[1533]: linux-amd64/LICENSE May 27 03:14:56.936615 tar[1533]: linux-amd64/README.md May 27 03:14:56.960264 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. May 27 03:14:56.963499 systemd[1]: Started getty@tty1.service - Getty on tty1. May 27 03:14:56.965612 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. May 27 03:14:56.966358 systemd[1]: Reached target getty.target - Login Prompts. May 27 03:14:56.971176 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. May 27 03:14:57.839133 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 27 03:14:57.840967 systemd[1]: Reached target multi-user.target - Multi-User System. May 27 03:14:57.842022 systemd[1]: Startup finished in 3.388s (kernel) + 5.637s (initrd) + 6.605s (userspace) = 15.631s. May 27 03:14:57.847776 (kubelet)[1676]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 27 03:14:58.451378 kubelet[1676]: E0527 03:14:58.451259 1676 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 27 03:14:58.454813 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 27 03:14:58.455010 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 27 03:14:58.455512 systemd[1]: kubelet.service: Consumed 1.276s CPU time, 263.4M memory peak. May 27 03:14:59.439408 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. May 27 03:14:59.441242 systemd[1]: Started sshd@0-209.38.65.15:22-139.178.68.195:41536.service - OpenSSH per-connection server daemon (139.178.68.195:41536). May 27 03:14:59.541560 sshd[1688]: Accepted publickey for core from 139.178.68.195 port 41536 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:14:59.544689 sshd-session[1688]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:14:59.553583 systemd[1]: Created slice user-500.slice - User Slice of UID 500. May 27 03:14:59.555085 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... May 27 03:14:59.572599 systemd-logind[1522]: New session 1 of user core. May 27 03:14:59.589777 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. May 27 03:14:59.594070 systemd[1]: Starting user@500.service - User Manager for UID 500... May 27 03:14:59.617011 (systemd)[1692]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 27 03:14:59.621260 systemd-logind[1522]: New session c1 of user core. May 27 03:14:59.804828 systemd[1692]: Queued start job for default target default.target. May 27 03:14:59.812946 systemd[1692]: Created slice app.slice - User Application Slice. May 27 03:14:59.812985 systemd[1692]: Reached target paths.target - Paths. May 27 03:14:59.813032 systemd[1692]: Reached target timers.target - Timers. May 27 03:14:59.814596 systemd[1692]: Starting dbus.socket - D-Bus User Message Bus Socket... May 27 03:14:59.829340 systemd[1692]: Listening on dbus.socket - D-Bus User Message Bus Socket. May 27 03:14:59.829470 systemd[1692]: Reached target sockets.target - Sockets. May 27 03:14:59.829523 systemd[1692]: Reached target basic.target - Basic System. May 27 03:14:59.829563 systemd[1692]: Reached target default.target - Main User Target. May 27 03:14:59.829597 systemd[1692]: Startup finished in 193ms. May 27 03:14:59.829769 systemd[1]: Started user@500.service - User Manager for UID 500. May 27 03:14:59.840443 systemd[1]: Started session-1.scope - Session 1 of User core. May 27 03:14:59.909337 systemd[1]: Started sshd@1-209.38.65.15:22-139.178.68.195:41552.service - OpenSSH per-connection server daemon (139.178.68.195:41552). May 27 03:14:59.973981 sshd[1703]: Accepted publickey for core from 139.178.68.195 port 41552 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:14:59.976612 sshd-session[1703]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:14:59.983571 systemd-logind[1522]: New session 2 of user core. May 27 03:14:59.991398 systemd[1]: Started session-2.scope - Session 2 of User core. May 27 03:15:00.059885 sshd[1705]: Connection closed by 139.178.68.195 port 41552 May 27 03:15:00.060421 sshd-session[1703]: pam_unix(sshd:session): session closed for user core May 27 03:15:00.076006 systemd[1]: sshd@1-209.38.65.15:22-139.178.68.195:41552.service: Deactivated successfully. May 27 03:15:00.080119 systemd[1]: session-2.scope: Deactivated successfully. May 27 03:15:00.082187 systemd-logind[1522]: Session 2 logged out. Waiting for processes to exit. May 27 03:15:00.088435 systemd[1]: Started sshd@2-209.38.65.15:22-139.178.68.195:41556.service - OpenSSH per-connection server daemon (139.178.68.195:41556). May 27 03:15:00.091223 systemd-logind[1522]: Removed session 2. May 27 03:15:00.160039 sshd[1711]: Accepted publickey for core from 139.178.68.195 port 41556 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:15:00.162396 sshd-session[1711]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:15:00.171227 systemd-logind[1522]: New session 3 of user core. May 27 03:15:00.178470 systemd[1]: Started session-3.scope - Session 3 of User core. May 27 03:15:00.243124 sshd[1713]: Connection closed by 139.178.68.195 port 41556 May 27 03:15:00.243025 sshd-session[1711]: pam_unix(sshd:session): session closed for user core May 27 03:15:00.257386 systemd[1]: sshd@2-209.38.65.15:22-139.178.68.195:41556.service: Deactivated successfully. May 27 03:15:00.260951 systemd[1]: session-3.scope: Deactivated successfully. May 27 03:15:00.262461 systemd-logind[1522]: Session 3 logged out. Waiting for processes to exit. May 27 03:15:00.268409 systemd[1]: Started sshd@3-209.38.65.15:22-139.178.68.195:41572.service - OpenSSH per-connection server daemon (139.178.68.195:41572). May 27 03:15:00.270448 systemd-logind[1522]: Removed session 3. May 27 03:15:00.334120 sshd[1719]: Accepted publickey for core from 139.178.68.195 port 41572 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:15:00.336164 sshd-session[1719]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:15:00.343697 systemd-logind[1522]: New session 4 of user core. May 27 03:15:00.354366 systemd[1]: Started session-4.scope - Session 4 of User core. May 27 03:15:00.417089 sshd[1721]: Connection closed by 139.178.68.195 port 41572 May 27 03:15:00.417772 sshd-session[1719]: pam_unix(sshd:session): session closed for user core May 27 03:15:00.433369 systemd[1]: sshd@3-209.38.65.15:22-139.178.68.195:41572.service: Deactivated successfully. May 27 03:15:00.435761 systemd[1]: session-4.scope: Deactivated successfully. May 27 03:15:00.437154 systemd-logind[1522]: Session 4 logged out. Waiting for processes to exit. May 27 03:15:00.440463 systemd-logind[1522]: Removed session 4. May 27 03:15:00.442570 systemd[1]: Started sshd@4-209.38.65.15:22-139.178.68.195:41576.service - OpenSSH per-connection server daemon (139.178.68.195:41576). May 27 03:15:00.511578 sshd[1727]: Accepted publickey for core from 139.178.68.195 port 41576 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:15:00.513625 sshd-session[1727]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:15:00.521769 systemd-logind[1522]: New session 5 of user core. May 27 03:15:00.528381 systemd[1]: Started session-5.scope - Session 5 of User core. May 27 03:15:00.601214 sudo[1730]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 May 27 03:15:00.601620 sudo[1730]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 27 03:15:00.617100 sudo[1730]: pam_unix(sudo:session): session closed for user root May 27 03:15:00.622096 sshd[1729]: Connection closed by 139.178.68.195 port 41576 May 27 03:15:00.621877 sshd-session[1727]: pam_unix(sshd:session): session closed for user core May 27 03:15:00.635565 systemd[1]: sshd@4-209.38.65.15:22-139.178.68.195:41576.service: Deactivated successfully. May 27 03:15:00.638578 systemd[1]: session-5.scope: Deactivated successfully. May 27 03:15:00.639829 systemd-logind[1522]: Session 5 logged out. Waiting for processes to exit. May 27 03:15:00.647479 systemd[1]: Started sshd@5-209.38.65.15:22-139.178.68.195:41592.service - OpenSSH per-connection server daemon (139.178.68.195:41592). May 27 03:15:00.650361 systemd-logind[1522]: Removed session 5. May 27 03:15:00.716707 sshd[1736]: Accepted publickey for core from 139.178.68.195 port 41592 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:15:00.718784 sshd-session[1736]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:15:00.726133 systemd-logind[1522]: New session 6 of user core. May 27 03:15:00.734488 systemd[1]: Started session-6.scope - Session 6 of User core. May 27 03:15:00.796630 sudo[1740]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules May 27 03:15:00.796971 sudo[1740]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 27 03:15:00.807382 sudo[1740]: pam_unix(sudo:session): session closed for user root May 27 03:15:00.814787 sudo[1739]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules May 27 03:15:00.815168 sudo[1739]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 27 03:15:00.830165 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 27 03:15:00.890805 augenrules[1762]: No rules May 27 03:15:00.892572 systemd[1]: audit-rules.service: Deactivated successfully. May 27 03:15:00.892919 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 27 03:15:00.894357 sudo[1739]: pam_unix(sudo:session): session closed for user root May 27 03:15:00.897513 sshd[1738]: Connection closed by 139.178.68.195 port 41592 May 27 03:15:00.898102 sshd-session[1736]: pam_unix(sshd:session): session closed for user core May 27 03:15:00.902850 systemd[1]: sshd@5-209.38.65.15:22-139.178.68.195:41592.service: Deactivated successfully. May 27 03:15:00.905012 systemd[1]: session-6.scope: Deactivated successfully. May 27 03:15:00.922074 systemd-logind[1522]: Session 6 logged out. Waiting for processes to exit. May 27 03:15:00.925520 systemd[1]: Started sshd@6-209.38.65.15:22-139.178.68.195:41608.service - OpenSSH per-connection server daemon (139.178.68.195:41608). May 27 03:15:00.927566 systemd-logind[1522]: Removed session 6. May 27 03:15:00.986502 sshd[1771]: Accepted publickey for core from 139.178.68.195 port 41608 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:15:00.988619 sshd-session[1771]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:15:00.996328 systemd-logind[1522]: New session 7 of user core. May 27 03:15:01.011480 systemd[1]: Started session-7.scope - Session 7 of User core. May 27 03:15:01.071655 sudo[1774]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 27 03:15:01.071996 sudo[1774]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 27 03:15:01.775809 systemd[1]: Starting docker.service - Docker Application Container Engine... May 27 03:15:01.805171 (dockerd)[1793]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU May 27 03:15:02.384326 dockerd[1793]: time="2025-05-27T03:15:02.383434469Z" level=info msg="Starting up" May 27 03:15:02.387360 dockerd[1793]: time="2025-05-27T03:15:02.387311134Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" May 27 03:15:02.459730 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport3969122689-merged.mount: Deactivated successfully. May 27 03:15:02.479918 systemd[1]: var-lib-docker-metacopy\x2dcheck795760041-merged.mount: Deactivated successfully. May 27 03:15:02.516737 dockerd[1793]: time="2025-05-27T03:15:02.516575569Z" level=info msg="Loading containers: start." May 27 03:15:02.542295 kernel: Initializing XFRM netlink socket May 27 03:15:03.157294 systemd-networkd[1448]: docker0: Link UP May 27 03:15:03.164438 dockerd[1793]: time="2025-05-27T03:15:03.164270436Z" level=info msg="Loading containers: done." May 27 03:15:03.195044 dockerd[1793]: time="2025-05-27T03:15:03.194528441Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 27 03:15:03.195044 dockerd[1793]: time="2025-05-27T03:15:03.194645694Z" level=info msg="Docker daemon" commit=bbd0a17ccc67e48d4a69393287b7fcc4f0578683 containerd-snapshotter=false storage-driver=overlay2 version=28.0.1 May 27 03:15:03.195044 dockerd[1793]: time="2025-05-27T03:15:03.194822371Z" level=info msg="Initializing buildkit" May 27 03:15:03.277969 dockerd[1793]: time="2025-05-27T03:15:03.277858592Z" level=info msg="Completed buildkit initialization" May 27 03:15:03.299225 dockerd[1793]: time="2025-05-27T03:15:03.299123670Z" level=info msg="Daemon has completed initialization" May 27 03:15:03.299518 systemd[1]: Started docker.service - Docker Application Container Engine. May 27 03:15:03.301727 dockerd[1793]: time="2025-05-27T03:15:03.299485241Z" level=info msg="API listen on /run/docker.sock" May 27 03:15:03.450983 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2949904887-merged.mount: Deactivated successfully. May 27 03:15:04.396101 containerd[1545]: time="2025-05-27T03:15:04.395907186Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.9\"" May 27 03:15:04.981653 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount682820708.mount: Deactivated successfully. May 27 03:15:06.332528 containerd[1545]: time="2025-05-27T03:15:06.332451014Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:06.334585 containerd[1545]: time="2025-05-27T03:15:06.333964998Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.9: active requests=0, bytes read=28078845" May 27 03:15:06.334585 containerd[1545]: time="2025-05-27T03:15:06.334163510Z" level=info msg="ImageCreate event name:\"sha256:0c19e0eafbdfffa1317cf99a16478265a4cd746ef677de27b0be6a8b515f36b1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:06.337895 containerd[1545]: time="2025-05-27T03:15:06.337829608Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:5b68f0df22013422dc8fb9ddfcff513eb6fc92f9dbf8aae41555c895efef5a20\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:06.339560 containerd[1545]: time="2025-05-27T03:15:06.339497982Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.9\" with image id \"sha256:0c19e0eafbdfffa1317cf99a16478265a4cd746ef677de27b0be6a8b515f36b1\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:5b68f0df22013422dc8fb9ddfcff513eb6fc92f9dbf8aae41555c895efef5a20\", size \"28075645\" in 1.942641516s" May 27 03:15:06.339560 containerd[1545]: time="2025-05-27T03:15:06.339562918Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.9\" returns image reference \"sha256:0c19e0eafbdfffa1317cf99a16478265a4cd746ef677de27b0be6a8b515f36b1\"" May 27 03:15:06.340810 containerd[1545]: time="2025-05-27T03:15:06.340730028Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.9\"" May 27 03:15:07.729826 containerd[1545]: time="2025-05-27T03:15:07.729749737Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:07.730843 containerd[1545]: time="2025-05-27T03:15:07.730790983Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.9: active requests=0, bytes read=24713522" May 27 03:15:07.732187 containerd[1545]: time="2025-05-27T03:15:07.732134851Z" level=info msg="ImageCreate event name:\"sha256:6aa3d581404ae6ae5dc355cb750aaedec843d2c99263d28fce50277e8e2a6ec2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:07.734896 containerd[1545]: time="2025-05-27T03:15:07.734850062Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:be9e7987d323b38a12e28436cff6d6ec6fc31ffdd3ea11eaa9d74852e9d31248\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:07.736321 containerd[1545]: time="2025-05-27T03:15:07.736260664Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.9\" with image id \"sha256:6aa3d581404ae6ae5dc355cb750aaedec843d2c99263d28fce50277e8e2a6ec2\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:be9e7987d323b38a12e28436cff6d6ec6fc31ffdd3ea11eaa9d74852e9d31248\", size \"26315362\" in 1.395444863s" May 27 03:15:07.736538 containerd[1545]: time="2025-05-27T03:15:07.736513117Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.9\" returns image reference \"sha256:6aa3d581404ae6ae5dc355cb750aaedec843d2c99263d28fce50277e8e2a6ec2\"" May 27 03:15:07.737517 containerd[1545]: time="2025-05-27T03:15:07.737462908Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.9\"" May 27 03:15:08.705756 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 27 03:15:08.710071 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 27 03:15:08.924752 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 27 03:15:08.941938 (kubelet)[2074]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 27 03:15:09.042333 kubelet[2074]: E0527 03:15:09.042109 2074 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 27 03:15:09.052256 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 27 03:15:09.052434 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 27 03:15:09.053630 systemd[1]: kubelet.service: Consumed 244ms CPU time, 109.1M memory peak. May 27 03:15:09.170283 containerd[1545]: time="2025-05-27T03:15:09.170191210Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:09.171422 containerd[1545]: time="2025-05-27T03:15:09.171337234Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.9: active requests=0, bytes read=18784311" May 27 03:15:09.172434 containerd[1545]: time="2025-05-27T03:15:09.172047611Z" level=info msg="ImageCreate event name:\"sha256:737ed3eafaf27a28ea9e13b736011bfed5bd349785ac6bc220b34eaf4adc51e3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:09.176213 containerd[1545]: time="2025-05-27T03:15:09.176155290Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:eb358c7346bb17ab2c639c3ff8ab76a147dec7ae609f5c0c2800233e42253ed1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:09.177800 containerd[1545]: time="2025-05-27T03:15:09.177747024Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.9\" with image id \"sha256:737ed3eafaf27a28ea9e13b736011bfed5bd349785ac6bc220b34eaf4adc51e3\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:eb358c7346bb17ab2c639c3ff8ab76a147dec7ae609f5c0c2800233e42253ed1\", size \"20386169\" in 1.440066399s" May 27 03:15:09.177975 containerd[1545]: time="2025-05-27T03:15:09.177958206Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.9\" returns image reference \"sha256:737ed3eafaf27a28ea9e13b736011bfed5bd349785ac6bc220b34eaf4adc51e3\"" May 27 03:15:09.178906 containerd[1545]: time="2025-05-27T03:15:09.178812151Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.9\"" May 27 03:15:10.198418 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3301278764.mount: Deactivated successfully. May 27 03:15:10.687021 containerd[1545]: time="2025-05-27T03:15:10.686864820Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:10.688382 containerd[1545]: time="2025-05-27T03:15:10.688332179Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.9: active requests=0, bytes read=30355623" May 27 03:15:10.689725 containerd[1545]: time="2025-05-27T03:15:10.689131731Z" level=info msg="ImageCreate event name:\"sha256:11a47a71ed3ecf643e15a11990daed3b656279449ba9344db0b54652c4723578\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:10.690657 containerd[1545]: time="2025-05-27T03:15:10.690617771Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:fdf026cf2434537e499e9c739d189ca8fc57101d929ac5ccd8e24f979a9738c1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:10.691295 containerd[1545]: time="2025-05-27T03:15:10.691259703Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.9\" with image id \"sha256:11a47a71ed3ecf643e15a11990daed3b656279449ba9344db0b54652c4723578\", repo tag \"registry.k8s.io/kube-proxy:v1.31.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:fdf026cf2434537e499e9c739d189ca8fc57101d929ac5ccd8e24f979a9738c1\", size \"30354642\" in 1.512398395s" May 27 03:15:10.691295 containerd[1545]: time="2025-05-27T03:15:10.691294246Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.9\" returns image reference \"sha256:11a47a71ed3ecf643e15a11990daed3b656279449ba9344db0b54652c4723578\"" May 27 03:15:10.691845 containerd[1545]: time="2025-05-27T03:15:10.691734958Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" May 27 03:15:10.692947 systemd-resolved[1402]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.3. May 27 03:15:11.170970 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3135380232.mount: Deactivated successfully. May 27 03:15:12.256914 containerd[1545]: time="2025-05-27T03:15:12.256845608Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:12.258027 containerd[1545]: time="2025-05-27T03:15:12.257987157Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" May 27 03:15:12.259233 containerd[1545]: time="2025-05-27T03:15:12.259144670Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:12.263709 containerd[1545]: time="2025-05-27T03:15:12.263105106Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:12.264436 containerd[1545]: time="2025-05-27T03:15:12.264383394Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.572617635s" May 27 03:15:12.264554 containerd[1545]: time="2025-05-27T03:15:12.264438798Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" May 27 03:15:12.265278 containerd[1545]: time="2025-05-27T03:15:12.265236140Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" May 27 03:15:12.664133 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1520072023.mount: Deactivated successfully. May 27 03:15:12.669114 containerd[1545]: time="2025-05-27T03:15:12.668480465Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 27 03:15:12.670420 containerd[1545]: time="2025-05-27T03:15:12.670384091Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" May 27 03:15:12.671105 containerd[1545]: time="2025-05-27T03:15:12.671076217Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 27 03:15:12.675027 containerd[1545]: time="2025-05-27T03:15:12.674967015Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 27 03:15:12.675906 containerd[1545]: time="2025-05-27T03:15:12.675851432Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 410.568967ms" May 27 03:15:12.675906 containerd[1545]: time="2025-05-27T03:15:12.675907536Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" May 27 03:15:12.677362 containerd[1545]: time="2025-05-27T03:15:12.677314405Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" May 27 03:15:13.208671 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount209312080.mount: Deactivated successfully. May 27 03:15:13.766425 systemd-resolved[1402]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.2. May 27 03:15:15.140406 containerd[1545]: time="2025-05-27T03:15:15.140335698Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:15.141460 containerd[1545]: time="2025-05-27T03:15:15.141422181Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=56780013" May 27 03:15:15.144084 containerd[1545]: time="2025-05-27T03:15:15.141946716Z" level=info msg="ImageCreate event name:\"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:15.145239 containerd[1545]: time="2025-05-27T03:15:15.145190940Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:15.146520 containerd[1545]: time="2025-05-27T03:15:15.146479050Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"56909194\" in 2.469114359s" May 27 03:15:15.146672 containerd[1545]: time="2025-05-27T03:15:15.146650849Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" May 27 03:15:18.236364 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 27 03:15:18.236536 systemd[1]: kubelet.service: Consumed 244ms CPU time, 109.1M memory peak. May 27 03:15:18.239256 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 27 03:15:18.273821 systemd[1]: Reload requested from client PID 2227 ('systemctl') (unit session-7.scope)... May 27 03:15:18.274023 systemd[1]: Reloading... May 27 03:15:18.433154 zram_generator::config[2271]: No configuration found. May 27 03:15:18.553646 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 27 03:15:18.697859 systemd[1]: Reloading finished in 423 ms. May 27 03:15:18.753797 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM May 27 03:15:18.753948 systemd[1]: kubelet.service: Failed with result 'signal'. May 27 03:15:18.754335 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 27 03:15:18.754393 systemd[1]: kubelet.service: Consumed 121ms CPU time, 97.6M memory peak. May 27 03:15:18.756381 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 27 03:15:18.940145 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 27 03:15:18.954220 (kubelet)[2322]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 27 03:15:19.023446 kubelet[2322]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 27 03:15:19.023446 kubelet[2322]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 27 03:15:19.023446 kubelet[2322]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 27 03:15:19.023972 kubelet[2322]: I0527 03:15:19.023597 2322 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 27 03:15:19.748219 kubelet[2322]: I0527 03:15:19.748151 2322 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" May 27 03:15:19.748493 kubelet[2322]: I0527 03:15:19.748478 2322 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 27 03:15:19.748912 kubelet[2322]: I0527 03:15:19.748885 2322 server.go:934] "Client rotation is on, will bootstrap in background" May 27 03:15:19.773288 kubelet[2322]: E0527 03:15:19.773215 2322 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://209.38.65.15:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 209.38.65.15:6443: connect: connection refused" logger="UnhandledError" May 27 03:15:19.775956 kubelet[2322]: I0527 03:15:19.775895 2322 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 27 03:15:19.784844 kubelet[2322]: I0527 03:15:19.784789 2322 server.go:1431] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" May 27 03:15:19.791949 kubelet[2322]: I0527 03:15:19.791877 2322 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 27 03:15:19.794020 kubelet[2322]: I0527 03:15:19.792991 2322 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" May 27 03:15:19.794020 kubelet[2322]: I0527 03:15:19.793306 2322 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 27 03:15:19.794020 kubelet[2322]: I0527 03:15:19.793339 2322 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4344.0.0-4-f2dd3e9233","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} May 27 03:15:19.794020 kubelet[2322]: I0527 03:15:19.793646 2322 topology_manager.go:138] "Creating topology manager with none policy" May 27 03:15:19.794356 kubelet[2322]: I0527 03:15:19.793659 2322 container_manager_linux.go:300] "Creating device plugin manager" May 27 03:15:19.794356 kubelet[2322]: I0527 03:15:19.793809 2322 state_mem.go:36] "Initialized new in-memory state store" May 27 03:15:19.799100 kubelet[2322]: I0527 03:15:19.799013 2322 kubelet.go:408] "Attempting to sync node with API server" May 27 03:15:19.799362 kubelet[2322]: I0527 03:15:19.799338 2322 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" May 27 03:15:19.799495 kubelet[2322]: I0527 03:15:19.799484 2322 kubelet.go:314] "Adding apiserver pod source" May 27 03:15:19.799569 kubelet[2322]: I0527 03:15:19.799560 2322 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 27 03:15:19.804025 kubelet[2322]: W0527 03:15:19.803911 2322 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://209.38.65.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4344.0.0-4-f2dd3e9233&limit=500&resourceVersion=0": dial tcp 209.38.65.15:6443: connect: connection refused May 27 03:15:19.804453 kubelet[2322]: E0527 03:15:19.804042 2322 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://209.38.65.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4344.0.0-4-f2dd3e9233&limit=500&resourceVersion=0\": dial tcp 209.38.65.15:6443: connect: connection refused" logger="UnhandledError" May 27 03:15:19.804621 kubelet[2322]: I0527 03:15:19.804600 2322 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" May 27 03:15:19.808544 kubelet[2322]: I0527 03:15:19.808491 2322 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 27 03:15:19.809377 kubelet[2322]: W0527 03:15:19.809330 2322 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 27 03:15:19.811565 kubelet[2322]: I0527 03:15:19.811527 2322 server.go:1274] "Started kubelet" May 27 03:15:19.811882 kubelet[2322]: W0527 03:15:19.811837 2322 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://209.38.65.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 209.38.65.15:6443: connect: connection refused May 27 03:15:19.812004 kubelet[2322]: E0527 03:15:19.811985 2322 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://209.38.65.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 209.38.65.15:6443: connect: connection refused" logger="UnhandledError" May 27 03:15:19.812222 kubelet[2322]: I0527 03:15:19.812181 2322 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 27 03:15:19.815314 kubelet[2322]: I0527 03:15:19.815281 2322 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 27 03:15:19.815667 kubelet[2322]: I0527 03:15:19.815576 2322 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 27 03:15:19.816297 kubelet[2322]: I0527 03:15:19.816268 2322 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 27 03:15:19.820810 kubelet[2322]: E0527 03:15:19.819252 2322 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://209.38.65.15:6443/api/v1/namespaces/default/events\": dial tcp 209.38.65.15:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4344.0.0-4-f2dd3e9233.184343e25a986592 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4344.0.0-4-f2dd3e9233,UID:ci-4344.0.0-4-f2dd3e9233,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4344.0.0-4-f2dd3e9233,},FirstTimestamp:2025-05-27 03:15:19.811491218 +0000 UTC m=+0.851288005,LastTimestamp:2025-05-27 03:15:19.811491218 +0000 UTC m=+0.851288005,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4344.0.0-4-f2dd3e9233,}" May 27 03:15:19.822085 kubelet[2322]: I0527 03:15:19.821937 2322 server.go:449] "Adding debug handlers to kubelet server" May 27 03:15:19.825729 kubelet[2322]: I0527 03:15:19.825676 2322 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 27 03:15:19.827805 kubelet[2322]: I0527 03:15:19.827713 2322 volume_manager.go:289] "Starting Kubelet Volume Manager" May 27 03:15:19.828123 kubelet[2322]: E0527 03:15:19.828093 2322 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4344.0.0-4-f2dd3e9233\" not found" May 27 03:15:19.828588 kubelet[2322]: I0527 03:15:19.828550 2322 desired_state_of_world_populator.go:147] "Desired state populator starts to run" May 27 03:15:19.828674 kubelet[2322]: I0527 03:15:19.828639 2322 reconciler.go:26] "Reconciler: start to sync state" May 27 03:15:19.829641 kubelet[2322]: W0527 03:15:19.829478 2322 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://209.38.65.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 209.38.65.15:6443: connect: connection refused May 27 03:15:19.829641 kubelet[2322]: E0527 03:15:19.829571 2322 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://209.38.65.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 209.38.65.15:6443: connect: connection refused" logger="UnhandledError" May 27 03:15:19.829918 kubelet[2322]: E0527 03:15:19.829878 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://209.38.65.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4344.0.0-4-f2dd3e9233?timeout=10s\": dial tcp 209.38.65.15:6443: connect: connection refused" interval="200ms" May 27 03:15:19.831880 kubelet[2322]: I0527 03:15:19.831285 2322 factory.go:221] Registration of the systemd container factory successfully May 27 03:15:19.831880 kubelet[2322]: I0527 03:15:19.831407 2322 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 27 03:15:19.836913 kubelet[2322]: I0527 03:15:19.836878 2322 factory.go:221] Registration of the containerd container factory successfully May 27 03:15:19.856229 kubelet[2322]: I0527 03:15:19.856134 2322 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 27 03:15:19.858984 kubelet[2322]: I0527 03:15:19.858945 2322 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 27 03:15:19.859620 kubelet[2322]: I0527 03:15:19.859090 2322 status_manager.go:217] "Starting to sync pod status with apiserver" May 27 03:15:19.859620 kubelet[2322]: I0527 03:15:19.859119 2322 kubelet.go:2321] "Starting kubelet main sync loop" May 27 03:15:19.859620 kubelet[2322]: E0527 03:15:19.859176 2322 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 27 03:15:19.866507 kubelet[2322]: E0527 03:15:19.866460 2322 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 27 03:15:19.871943 kubelet[2322]: W0527 03:15:19.871860 2322 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://209.38.65.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 209.38.65.15:6443: connect: connection refused May 27 03:15:19.872150 kubelet[2322]: E0527 03:15:19.872131 2322 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://209.38.65.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 209.38.65.15:6443: connect: connection refused" logger="UnhandledError" May 27 03:15:19.876159 kubelet[2322]: I0527 03:15:19.876127 2322 cpu_manager.go:214] "Starting CPU manager" policy="none" May 27 03:15:19.876335 kubelet[2322]: I0527 03:15:19.876325 2322 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 27 03:15:19.876567 kubelet[2322]: I0527 03:15:19.876537 2322 state_mem.go:36] "Initialized new in-memory state store" May 27 03:15:19.879087 kubelet[2322]: I0527 03:15:19.879022 2322 policy_none.go:49] "None policy: Start" May 27 03:15:19.880266 kubelet[2322]: I0527 03:15:19.880242 2322 memory_manager.go:170] "Starting memorymanager" policy="None" May 27 03:15:19.880372 kubelet[2322]: I0527 03:15:19.880335 2322 state_mem.go:35] "Initializing new in-memory state store" May 27 03:15:19.888872 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. May 27 03:15:19.905162 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. May 27 03:15:19.909307 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. May 27 03:15:19.922449 kubelet[2322]: I0527 03:15:19.922414 2322 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 27 03:15:19.923313 kubelet[2322]: I0527 03:15:19.923258 2322 eviction_manager.go:189] "Eviction manager: starting control loop" May 27 03:15:19.923497 kubelet[2322]: I0527 03:15:19.923281 2322 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 27 03:15:19.924006 kubelet[2322]: I0527 03:15:19.923979 2322 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 27 03:15:19.927088 kubelet[2322]: E0527 03:15:19.927003 2322 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4344.0.0-4-f2dd3e9233\" not found" May 27 03:15:19.975791 systemd[1]: Created slice kubepods-burstable-pod521952ad3f5ca49f5ec8b6537f02b8c4.slice - libcontainer container kubepods-burstable-pod521952ad3f5ca49f5ec8b6537f02b8c4.slice. May 27 03:15:20.004904 systemd[1]: Created slice kubepods-burstable-pod6cd9fb9d21ceda5ec57176e817db2b13.slice - libcontainer container kubepods-burstable-pod6cd9fb9d21ceda5ec57176e817db2b13.slice. May 27 03:15:20.023444 systemd[1]: Created slice kubepods-burstable-pod8cc95706643904ad5085246ea89deaac.slice - libcontainer container kubepods-burstable-pod8cc95706643904ad5085246ea89deaac.slice. May 27 03:15:20.026521 kubelet[2322]: I0527 03:15:20.026475 2322 kubelet_node_status.go:72] "Attempting to register node" node="ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.027200 kubelet[2322]: E0527 03:15:20.026829 2322 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://209.38.65.15:6443/api/v1/nodes\": dial tcp 209.38.65.15:6443: connect: connection refused" node="ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.029520 kubelet[2322]: I0527 03:15:20.029484 2322 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/521952ad3f5ca49f5ec8b6537f02b8c4-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4344.0.0-4-f2dd3e9233\" (UID: \"521952ad3f5ca49f5ec8b6537f02b8c4\") " pod="kube-system/kube-apiserver-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.029520 kubelet[2322]: I0527 03:15:20.029517 2322 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/6cd9fb9d21ceda5ec57176e817db2b13-ca-certs\") pod \"kube-controller-manager-ci-4344.0.0-4-f2dd3e9233\" (UID: \"6cd9fb9d21ceda5ec57176e817db2b13\") " pod="kube-system/kube-controller-manager-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.029520 kubelet[2322]: I0527 03:15:20.029537 2322 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/6cd9fb9d21ceda5ec57176e817db2b13-flexvolume-dir\") pod \"kube-controller-manager-ci-4344.0.0-4-f2dd3e9233\" (UID: \"6cd9fb9d21ceda5ec57176e817db2b13\") " pod="kube-system/kube-controller-manager-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.029520 kubelet[2322]: I0527 03:15:20.029553 2322 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/6cd9fb9d21ceda5ec57176e817db2b13-k8s-certs\") pod \"kube-controller-manager-ci-4344.0.0-4-f2dd3e9233\" (UID: \"6cd9fb9d21ceda5ec57176e817db2b13\") " pod="kube-system/kube-controller-manager-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.029520 kubelet[2322]: I0527 03:15:20.029570 2322 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6cd9fb9d21ceda5ec57176e817db2b13-kubeconfig\") pod \"kube-controller-manager-ci-4344.0.0-4-f2dd3e9233\" (UID: \"6cd9fb9d21ceda5ec57176e817db2b13\") " pod="kube-system/kube-controller-manager-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.029852 kubelet[2322]: I0527 03:15:20.029585 2322 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/6cd9fb9d21ceda5ec57176e817db2b13-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4344.0.0-4-f2dd3e9233\" (UID: \"6cd9fb9d21ceda5ec57176e817db2b13\") " pod="kube-system/kube-controller-manager-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.029852 kubelet[2322]: I0527 03:15:20.029603 2322 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8cc95706643904ad5085246ea89deaac-kubeconfig\") pod \"kube-scheduler-ci-4344.0.0-4-f2dd3e9233\" (UID: \"8cc95706643904ad5085246ea89deaac\") " pod="kube-system/kube-scheduler-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.029852 kubelet[2322]: I0527 03:15:20.029631 2322 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/521952ad3f5ca49f5ec8b6537f02b8c4-ca-certs\") pod \"kube-apiserver-ci-4344.0.0-4-f2dd3e9233\" (UID: \"521952ad3f5ca49f5ec8b6537f02b8c4\") " pod="kube-system/kube-apiserver-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.029852 kubelet[2322]: I0527 03:15:20.029647 2322 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/521952ad3f5ca49f5ec8b6537f02b8c4-k8s-certs\") pod \"kube-apiserver-ci-4344.0.0-4-f2dd3e9233\" (UID: \"521952ad3f5ca49f5ec8b6537f02b8c4\") " pod="kube-system/kube-apiserver-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.030711 kubelet[2322]: E0527 03:15:20.030653 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://209.38.65.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4344.0.0-4-f2dd3e9233?timeout=10s\": dial tcp 209.38.65.15:6443: connect: connection refused" interval="400ms" May 27 03:15:20.228772 kubelet[2322]: I0527 03:15:20.228704 2322 kubelet_node_status.go:72] "Attempting to register node" node="ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.229822 kubelet[2322]: E0527 03:15:20.229779 2322 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://209.38.65.15:6443/api/v1/nodes\": dial tcp 209.38.65.15:6443: connect: connection refused" node="ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.300851 kubelet[2322]: E0527 03:15:20.300646 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:20.301712 containerd[1545]: time="2025-05-27T03:15:20.301664995Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4344.0.0-4-f2dd3e9233,Uid:521952ad3f5ca49f5ec8b6537f02b8c4,Namespace:kube-system,Attempt:0,}" May 27 03:15:20.321939 kubelet[2322]: E0527 03:15:20.321633 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:20.322508 containerd[1545]: time="2025-05-27T03:15:20.322457727Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4344.0.0-4-f2dd3e9233,Uid:6cd9fb9d21ceda5ec57176e817db2b13,Namespace:kube-system,Attempt:0,}" May 27 03:15:20.329664 kubelet[2322]: E0527 03:15:20.329606 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:20.343118 containerd[1545]: time="2025-05-27T03:15:20.342388982Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4344.0.0-4-f2dd3e9233,Uid:8cc95706643904ad5085246ea89deaac,Namespace:kube-system,Attempt:0,}" May 27 03:15:20.432031 kubelet[2322]: E0527 03:15:20.431969 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://209.38.65.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4344.0.0-4-f2dd3e9233?timeout=10s\": dial tcp 209.38.65.15:6443: connect: connection refused" interval="800ms" May 27 03:15:20.439440 containerd[1545]: time="2025-05-27T03:15:20.439102128Z" level=info msg="connecting to shim a9b656b7f5f8a8f65ed482bae4d36decca9b88c4fdde3dd28ebc2d4f52ed8e94" address="unix:///run/containerd/s/12d640b9bc676db9f36e889826ad6cd470680be3d1b1d49e94deee8ab85e96d7" namespace=k8s.io protocol=ttrpc version=3 May 27 03:15:20.445343 containerd[1545]: time="2025-05-27T03:15:20.445267508Z" level=info msg="connecting to shim e36d91a56cd0dcb87b1194de67e397f98b7019bb25f8fd7b54716bad51a36682" address="unix:///run/containerd/s/08a7fed621db82970d47c094aa40b6eda2fc72e50be410689d41b8a656af2b2c" namespace=k8s.io protocol=ttrpc version=3 May 27 03:15:20.453634 containerd[1545]: time="2025-05-27T03:15:20.453566661Z" level=info msg="connecting to shim a2467e77aa6dadb97de89b260ca325c43fa6ec3d681fff3136447df06b3048aa" address="unix:///run/containerd/s/77016c6a27add994e815a4a0beee132b4cc5176939789c72ccdf23d13e40c44f" namespace=k8s.io protocol=ttrpc version=3 May 27 03:15:20.588823 systemd[1]: Started cri-containerd-a2467e77aa6dadb97de89b260ca325c43fa6ec3d681fff3136447df06b3048aa.scope - libcontainer container a2467e77aa6dadb97de89b260ca325c43fa6ec3d681fff3136447df06b3048aa. May 27 03:15:20.593279 systemd[1]: Started cri-containerd-a9b656b7f5f8a8f65ed482bae4d36decca9b88c4fdde3dd28ebc2d4f52ed8e94.scope - libcontainer container a9b656b7f5f8a8f65ed482bae4d36decca9b88c4fdde3dd28ebc2d4f52ed8e94. May 27 03:15:20.598450 systemd[1]: Started cri-containerd-e36d91a56cd0dcb87b1194de67e397f98b7019bb25f8fd7b54716bad51a36682.scope - libcontainer container e36d91a56cd0dcb87b1194de67e397f98b7019bb25f8fd7b54716bad51a36682. May 27 03:15:20.633100 kubelet[2322]: I0527 03:15:20.632905 2322 kubelet_node_status.go:72] "Attempting to register node" node="ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.637332 kubelet[2322]: E0527 03:15:20.637275 2322 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://209.38.65.15:6443/api/v1/nodes\": dial tcp 209.38.65.15:6443: connect: connection refused" node="ci-4344.0.0-4-f2dd3e9233" May 27 03:15:20.646924 kubelet[2322]: W0527 03:15:20.646714 2322 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://209.38.65.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 209.38.65.15:6443: connect: connection refused May 27 03:15:20.646924 kubelet[2322]: E0527 03:15:20.646818 2322 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://209.38.65.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 209.38.65.15:6443: connect: connection refused" logger="UnhandledError" May 27 03:15:20.707950 containerd[1545]: time="2025-05-27T03:15:20.707906701Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4344.0.0-4-f2dd3e9233,Uid:6cd9fb9d21ceda5ec57176e817db2b13,Namespace:kube-system,Attempt:0,} returns sandbox id \"a2467e77aa6dadb97de89b260ca325c43fa6ec3d681fff3136447df06b3048aa\"" May 27 03:15:20.722562 kubelet[2322]: E0527 03:15:20.722378 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:20.737971 containerd[1545]: time="2025-05-27T03:15:20.737855098Z" level=info msg="CreateContainer within sandbox \"a2467e77aa6dadb97de89b260ca325c43fa6ec3d681fff3136447df06b3048aa\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 27 03:15:20.739028 containerd[1545]: time="2025-05-27T03:15:20.738949369Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4344.0.0-4-f2dd3e9233,Uid:521952ad3f5ca49f5ec8b6537f02b8c4,Namespace:kube-system,Attempt:0,} returns sandbox id \"a9b656b7f5f8a8f65ed482bae4d36decca9b88c4fdde3dd28ebc2d4f52ed8e94\"" May 27 03:15:20.740288 kubelet[2322]: E0527 03:15:20.740247 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:20.744855 containerd[1545]: time="2025-05-27T03:15:20.744804809Z" level=info msg="CreateContainer within sandbox \"a9b656b7f5f8a8f65ed482bae4d36decca9b88c4fdde3dd28ebc2d4f52ed8e94\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 27 03:15:20.761638 containerd[1545]: time="2025-05-27T03:15:20.761585564Z" level=info msg="Container 866bb033abc83073bb6a64a3305bde658c3fb639e13d648d9a1c044ef6a26953: CDI devices from CRI Config.CDIDevices: []" May 27 03:15:20.781650 containerd[1545]: time="2025-05-27T03:15:20.781398844Z" level=info msg="CreateContainer within sandbox \"a9b656b7f5f8a8f65ed482bae4d36decca9b88c4fdde3dd28ebc2d4f52ed8e94\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"866bb033abc83073bb6a64a3305bde658c3fb639e13d648d9a1c044ef6a26953\"" May 27 03:15:20.784141 containerd[1545]: time="2025-05-27T03:15:20.784087487Z" level=info msg="StartContainer for \"866bb033abc83073bb6a64a3305bde658c3fb639e13d648d9a1c044ef6a26953\"" May 27 03:15:20.787553 containerd[1545]: time="2025-05-27T03:15:20.786571467Z" level=info msg="Container f534c08a5f8a13484695d9a539a72879145a845c3d84b923759b2c47ede373fe: CDI devices from CRI Config.CDIDevices: []" May 27 03:15:20.787553 containerd[1545]: time="2025-05-27T03:15:20.787307611Z" level=info msg="connecting to shim 866bb033abc83073bb6a64a3305bde658c3fb639e13d648d9a1c044ef6a26953" address="unix:///run/containerd/s/12d640b9bc676db9f36e889826ad6cd470680be3d1b1d49e94deee8ab85e96d7" protocol=ttrpc version=3 May 27 03:15:20.796335 containerd[1545]: time="2025-05-27T03:15:20.796269810Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4344.0.0-4-f2dd3e9233,Uid:8cc95706643904ad5085246ea89deaac,Namespace:kube-system,Attempt:0,} returns sandbox id \"e36d91a56cd0dcb87b1194de67e397f98b7019bb25f8fd7b54716bad51a36682\"" May 27 03:15:20.798891 kubelet[2322]: E0527 03:15:20.798767 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:20.802329 containerd[1545]: time="2025-05-27T03:15:20.802201226Z" level=info msg="CreateContainer within sandbox \"a2467e77aa6dadb97de89b260ca325c43fa6ec3d681fff3136447df06b3048aa\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"f534c08a5f8a13484695d9a539a72879145a845c3d84b923759b2c47ede373fe\"" May 27 03:15:20.804669 containerd[1545]: time="2025-05-27T03:15:20.803129911Z" level=info msg="CreateContainer within sandbox \"e36d91a56cd0dcb87b1194de67e397f98b7019bb25f8fd7b54716bad51a36682\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 27 03:15:20.804669 containerd[1545]: time="2025-05-27T03:15:20.803279857Z" level=info msg="StartContainer for \"f534c08a5f8a13484695d9a539a72879145a845c3d84b923759b2c47ede373fe\"" May 27 03:15:20.807141 containerd[1545]: time="2025-05-27T03:15:20.807079803Z" level=info msg="connecting to shim f534c08a5f8a13484695d9a539a72879145a845c3d84b923759b2c47ede373fe" address="unix:///run/containerd/s/77016c6a27add994e815a4a0beee132b4cc5176939789c72ccdf23d13e40c44f" protocol=ttrpc version=3 May 27 03:15:20.814537 containerd[1545]: time="2025-05-27T03:15:20.814483567Z" level=info msg="Container 64faa938e27df8400f4fcaa8ca20144e02cbad4a8eb03e0687dfd923ab654775: CDI devices from CRI Config.CDIDevices: []" May 27 03:15:20.824322 containerd[1545]: time="2025-05-27T03:15:20.824274245Z" level=info msg="CreateContainer within sandbox \"e36d91a56cd0dcb87b1194de67e397f98b7019bb25f8fd7b54716bad51a36682\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"64faa938e27df8400f4fcaa8ca20144e02cbad4a8eb03e0687dfd923ab654775\"" May 27 03:15:20.825245 containerd[1545]: time="2025-05-27T03:15:20.825196268Z" level=info msg="StartContainer for \"64faa938e27df8400f4fcaa8ca20144e02cbad4a8eb03e0687dfd923ab654775\"" May 27 03:15:20.826984 containerd[1545]: time="2025-05-27T03:15:20.826944774Z" level=info msg="connecting to shim 64faa938e27df8400f4fcaa8ca20144e02cbad4a8eb03e0687dfd923ab654775" address="unix:///run/containerd/s/08a7fed621db82970d47c094aa40b6eda2fc72e50be410689d41b8a656af2b2c" protocol=ttrpc version=3 May 27 03:15:20.837480 systemd[1]: Started cri-containerd-866bb033abc83073bb6a64a3305bde658c3fb639e13d648d9a1c044ef6a26953.scope - libcontainer container 866bb033abc83073bb6a64a3305bde658c3fb639e13d648d9a1c044ef6a26953. May 27 03:15:20.859382 systemd[1]: Started cri-containerd-f534c08a5f8a13484695d9a539a72879145a845c3d84b923759b2c47ede373fe.scope - libcontainer container f534c08a5f8a13484695d9a539a72879145a845c3d84b923759b2c47ede373fe. May 27 03:15:20.886377 systemd[1]: Started cri-containerd-64faa938e27df8400f4fcaa8ca20144e02cbad4a8eb03e0687dfd923ab654775.scope - libcontainer container 64faa938e27df8400f4fcaa8ca20144e02cbad4a8eb03e0687dfd923ab654775. May 27 03:15:20.956319 containerd[1545]: time="2025-05-27T03:15:20.956263219Z" level=info msg="StartContainer for \"866bb033abc83073bb6a64a3305bde658c3fb639e13d648d9a1c044ef6a26953\" returns successfully" May 27 03:15:21.019610 containerd[1545]: time="2025-05-27T03:15:21.019528620Z" level=info msg="StartContainer for \"f534c08a5f8a13484695d9a539a72879145a845c3d84b923759b2c47ede373fe\" returns successfully" May 27 03:15:21.026070 kubelet[2322]: W0527 03:15:21.026007 2322 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://209.38.65.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 209.38.65.15:6443: connect: connection refused May 27 03:15:21.026238 kubelet[2322]: E0527 03:15:21.026102 2322 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://209.38.65.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 209.38.65.15:6443: connect: connection refused" logger="UnhandledError" May 27 03:15:21.109498 containerd[1545]: time="2025-05-27T03:15:21.109445652Z" level=info msg="StartContainer for \"64faa938e27df8400f4fcaa8ca20144e02cbad4a8eb03e0687dfd923ab654775\" returns successfully" May 27 03:15:21.439299 kubelet[2322]: I0527 03:15:21.439254 2322 kubelet_node_status.go:72] "Attempting to register node" node="ci-4344.0.0-4-f2dd3e9233" May 27 03:15:21.919587 kubelet[2322]: E0527 03:15:21.919475 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:21.924770 kubelet[2322]: E0527 03:15:21.924046 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:21.928141 kubelet[2322]: E0527 03:15:21.928099 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:22.933955 kubelet[2322]: E0527 03:15:22.933390 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:22.933955 kubelet[2322]: E0527 03:15:22.933887 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:22.934498 kubelet[2322]: E0527 03:15:22.933649 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:23.014795 kubelet[2322]: E0527 03:15:23.014751 2322 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4344.0.0-4-f2dd3e9233\" not found" node="ci-4344.0.0-4-f2dd3e9233" May 27 03:15:23.093243 kubelet[2322]: I0527 03:15:23.093156 2322 kubelet_node_status.go:75] "Successfully registered node" node="ci-4344.0.0-4-f2dd3e9233" May 27 03:15:23.813323 kubelet[2322]: I0527 03:15:23.813243 2322 apiserver.go:52] "Watching apiserver" May 27 03:15:23.829674 kubelet[2322]: I0527 03:15:23.829623 2322 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" May 27 03:15:23.959992 kubelet[2322]: W0527 03:15:23.959942 2322 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 27 03:15:23.961519 kubelet[2322]: E0527 03:15:23.961437 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:23.964083 kubelet[2322]: W0527 03:15:23.962488 2322 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 27 03:15:23.964083 kubelet[2322]: E0527 03:15:23.962943 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:24.935169 kubelet[2322]: E0527 03:15:24.935032 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:24.946589 kubelet[2322]: W0527 03:15:24.946426 2322 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 27 03:15:24.947066 kubelet[2322]: E0527 03:15:24.947022 2322 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4344.0.0-4-f2dd3e9233\" already exists" pod="kube-system/kube-apiserver-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:24.948294 kubelet[2322]: E0527 03:15:24.948256 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:24.952924 kubelet[2322]: W0527 03:15:24.952886 2322 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 27 03:15:24.953323 kubelet[2322]: E0527 03:15:24.953300 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:25.418964 systemd[1]: Reload requested from client PID 2594 ('systemctl') (unit session-7.scope)... May 27 03:15:25.419146 systemd[1]: Reloading... May 27 03:15:25.584107 zram_generator::config[2637]: No configuration found. May 27 03:15:25.756756 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 27 03:15:25.937369 kubelet[2322]: E0527 03:15:25.937329 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:25.939360 kubelet[2322]: E0527 03:15:25.937638 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:26.004484 systemd[1]: Reloading finished in 584 ms. May 27 03:15:26.042149 kubelet[2322]: I0527 03:15:26.041361 2322 dynamic_cafile_content.go:174] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 27 03:15:26.041915 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 27 03:15:26.063171 systemd[1]: kubelet.service: Deactivated successfully. May 27 03:15:26.063547 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 27 03:15:26.063648 systemd[1]: kubelet.service: Consumed 1.351s CPU time, 124.9M memory peak. May 27 03:15:26.066937 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 27 03:15:26.274319 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 27 03:15:26.290842 (kubelet)[2688]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 27 03:15:26.358837 kubelet[2688]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 27 03:15:26.358837 kubelet[2688]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 27 03:15:26.358837 kubelet[2688]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 27 03:15:26.361870 kubelet[2688]: I0527 03:15:26.359003 2688 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 27 03:15:26.373967 kubelet[2688]: I0527 03:15:26.373919 2688 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" May 27 03:15:26.374261 kubelet[2688]: I0527 03:15:26.374240 2688 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 27 03:15:26.374781 kubelet[2688]: I0527 03:15:26.374748 2688 server.go:934] "Client rotation is on, will bootstrap in background" May 27 03:15:26.376904 kubelet[2688]: I0527 03:15:26.376867 2688 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". May 27 03:15:26.397622 kubelet[2688]: I0527 03:15:26.396963 2688 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 27 03:15:26.409825 kubelet[2688]: I0527 03:15:26.409753 2688 server.go:1431] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" May 27 03:15:26.419232 kubelet[2688]: I0527 03:15:26.419164 2688 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 27 03:15:26.419977 kubelet[2688]: I0527 03:15:26.419868 2688 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" May 27 03:15:26.420401 kubelet[2688]: I0527 03:15:26.420346 2688 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 27 03:15:26.420822 kubelet[2688]: I0527 03:15:26.420523 2688 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4344.0.0-4-f2dd3e9233","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} May 27 03:15:26.421087 kubelet[2688]: I0527 03:15:26.421029 2688 topology_manager.go:138] "Creating topology manager with none policy" May 27 03:15:26.422139 kubelet[2688]: I0527 03:15:26.421164 2688 container_manager_linux.go:300] "Creating device plugin manager" May 27 03:15:26.422139 kubelet[2688]: I0527 03:15:26.421227 2688 state_mem.go:36] "Initialized new in-memory state store" May 27 03:15:26.422139 kubelet[2688]: I0527 03:15:26.421423 2688 kubelet.go:408] "Attempting to sync node with API server" May 27 03:15:26.422139 kubelet[2688]: I0527 03:15:26.421446 2688 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" May 27 03:15:26.422139 kubelet[2688]: I0527 03:15:26.421491 2688 kubelet.go:314] "Adding apiserver pod source" May 27 03:15:26.422139 kubelet[2688]: I0527 03:15:26.422016 2688 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 27 03:15:26.428331 kubelet[2688]: I0527 03:15:26.428281 2688 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" May 27 03:15:26.429298 kubelet[2688]: I0527 03:15:26.429262 2688 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 27 03:15:26.431391 kubelet[2688]: I0527 03:15:26.430359 2688 server.go:1274] "Started kubelet" May 27 03:15:26.435470 kubelet[2688]: I0527 03:15:26.435432 2688 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 27 03:15:26.456771 kubelet[2688]: I0527 03:15:26.455019 2688 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 27 03:15:26.472848 kubelet[2688]: I0527 03:15:26.472446 2688 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 27 03:15:26.479714 kubelet[2688]: I0527 03:15:26.479038 2688 volume_manager.go:289] "Starting Kubelet Volume Manager" May 27 03:15:26.483685 kubelet[2688]: I0527 03:15:26.483558 2688 desired_state_of_world_populator.go:147] "Desired state populator starts to run" May 27 03:15:26.483685 kubelet[2688]: I0527 03:15:26.483697 2688 reconciler.go:26] "Reconciler: start to sync state" May 27 03:15:26.486698 kubelet[2688]: I0527 03:15:26.486353 2688 server.go:449] "Adding debug handlers to kubelet server" May 27 03:15:26.488309 kubelet[2688]: I0527 03:15:26.471036 2688 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 27 03:15:26.488309 kubelet[2688]: I0527 03:15:26.487730 2688 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 27 03:15:26.489199 kubelet[2688]: I0527 03:15:26.488473 2688 factory.go:221] Registration of the systemd container factory successfully May 27 03:15:26.489199 kubelet[2688]: I0527 03:15:26.488587 2688 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 27 03:15:26.493337 kubelet[2688]: I0527 03:15:26.493269 2688 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 27 03:15:26.501113 kubelet[2688]: I0527 03:15:26.500255 2688 factory.go:221] Registration of the containerd container factory successfully May 27 03:15:26.502724 kubelet[2688]: I0527 03:15:26.501761 2688 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 27 03:15:26.502724 kubelet[2688]: I0527 03:15:26.501793 2688 status_manager.go:217] "Starting to sync pod status with apiserver" May 27 03:15:26.502724 kubelet[2688]: I0527 03:15:26.501817 2688 kubelet.go:2321] "Starting kubelet main sync loop" May 27 03:15:26.502724 kubelet[2688]: E0527 03:15:26.501866 2688 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 27 03:15:26.527766 kubelet[2688]: E0527 03:15:26.526261 2688 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 27 03:15:26.599087 kubelet[2688]: I0527 03:15:26.598875 2688 cpu_manager.go:214] "Starting CPU manager" policy="none" May 27 03:15:26.599087 kubelet[2688]: I0527 03:15:26.598896 2688 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 27 03:15:26.599087 kubelet[2688]: I0527 03:15:26.598922 2688 state_mem.go:36] "Initialized new in-memory state store" May 27 03:15:26.599666 kubelet[2688]: I0527 03:15:26.599524 2688 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 27 03:15:26.599666 kubelet[2688]: I0527 03:15:26.599546 2688 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 27 03:15:26.599666 kubelet[2688]: I0527 03:15:26.599569 2688 policy_none.go:49] "None policy: Start" May 27 03:15:26.601360 kubelet[2688]: I0527 03:15:26.601179 2688 memory_manager.go:170] "Starting memorymanager" policy="None" May 27 03:15:26.601360 kubelet[2688]: I0527 03:15:26.601223 2688 state_mem.go:35] "Initializing new in-memory state store" May 27 03:15:26.601802 kubelet[2688]: I0527 03:15:26.601703 2688 state_mem.go:75] "Updated machine memory state" May 27 03:15:26.601954 kubelet[2688]: E0527 03:15:26.601919 2688 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 27 03:15:26.608219 kubelet[2688]: I0527 03:15:26.607678 2688 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 27 03:15:26.608219 kubelet[2688]: I0527 03:15:26.607920 2688 eviction_manager.go:189] "Eviction manager: starting control loop" May 27 03:15:26.608219 kubelet[2688]: I0527 03:15:26.607933 2688 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 27 03:15:26.611566 kubelet[2688]: I0527 03:15:26.610765 2688 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 27 03:15:26.726716 kubelet[2688]: I0527 03:15:26.725463 2688 kubelet_node_status.go:72] "Attempting to register node" node="ci-4344.0.0-4-f2dd3e9233" May 27 03:15:26.734690 kubelet[2688]: I0527 03:15:26.734660 2688 kubelet_node_status.go:111] "Node was previously registered" node="ci-4344.0.0-4-f2dd3e9233" May 27 03:15:26.735185 kubelet[2688]: I0527 03:15:26.735116 2688 kubelet_node_status.go:75] "Successfully registered node" node="ci-4344.0.0-4-f2dd3e9233" May 27 03:15:26.816408 kubelet[2688]: W0527 03:15:26.816366 2688 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 27 03:15:26.816408 kubelet[2688]: E0527 03:15:26.816447 2688 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-ci-4344.0.0-4-f2dd3e9233\" already exists" pod="kube-system/kube-controller-manager-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:26.816717 kubelet[2688]: W0527 03:15:26.816366 2688 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 27 03:15:26.816717 kubelet[2688]: E0527 03:15:26.816541 2688 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4344.0.0-4-f2dd3e9233\" already exists" pod="kube-system/kube-apiserver-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:26.817768 kubelet[2688]: W0527 03:15:26.817703 2688 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 27 03:15:26.817964 kubelet[2688]: E0527 03:15:26.817916 2688 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ci-4344.0.0-4-f2dd3e9233\" already exists" pod="kube-system/kube-scheduler-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:26.887199 kubelet[2688]: I0527 03:15:26.885745 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/6cd9fb9d21ceda5ec57176e817db2b13-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4344.0.0-4-f2dd3e9233\" (UID: \"6cd9fb9d21ceda5ec57176e817db2b13\") " pod="kube-system/kube-controller-manager-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:26.887199 kubelet[2688]: I0527 03:15:26.885821 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8cc95706643904ad5085246ea89deaac-kubeconfig\") pod \"kube-scheduler-ci-4344.0.0-4-f2dd3e9233\" (UID: \"8cc95706643904ad5085246ea89deaac\") " pod="kube-system/kube-scheduler-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:26.887199 kubelet[2688]: I0527 03:15:26.885852 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/521952ad3f5ca49f5ec8b6537f02b8c4-ca-certs\") pod \"kube-apiserver-ci-4344.0.0-4-f2dd3e9233\" (UID: \"521952ad3f5ca49f5ec8b6537f02b8c4\") " pod="kube-system/kube-apiserver-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:26.887199 kubelet[2688]: I0527 03:15:26.885887 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/6cd9fb9d21ceda5ec57176e817db2b13-flexvolume-dir\") pod \"kube-controller-manager-ci-4344.0.0-4-f2dd3e9233\" (UID: \"6cd9fb9d21ceda5ec57176e817db2b13\") " pod="kube-system/kube-controller-manager-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:26.887199 kubelet[2688]: I0527 03:15:26.885926 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/6cd9fb9d21ceda5ec57176e817db2b13-k8s-certs\") pod \"kube-controller-manager-ci-4344.0.0-4-f2dd3e9233\" (UID: \"6cd9fb9d21ceda5ec57176e817db2b13\") " pod="kube-system/kube-controller-manager-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:26.887444 kubelet[2688]: I0527 03:15:26.885960 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6cd9fb9d21ceda5ec57176e817db2b13-kubeconfig\") pod \"kube-controller-manager-ci-4344.0.0-4-f2dd3e9233\" (UID: \"6cd9fb9d21ceda5ec57176e817db2b13\") " pod="kube-system/kube-controller-manager-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:26.887444 kubelet[2688]: I0527 03:15:26.885986 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/521952ad3f5ca49f5ec8b6537f02b8c4-k8s-certs\") pod \"kube-apiserver-ci-4344.0.0-4-f2dd3e9233\" (UID: \"521952ad3f5ca49f5ec8b6537f02b8c4\") " pod="kube-system/kube-apiserver-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:26.887444 kubelet[2688]: I0527 03:15:26.886018 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/521952ad3f5ca49f5ec8b6537f02b8c4-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4344.0.0-4-f2dd3e9233\" (UID: \"521952ad3f5ca49f5ec8b6537f02b8c4\") " pod="kube-system/kube-apiserver-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:26.887837 kubelet[2688]: I0527 03:15:26.886047 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/6cd9fb9d21ceda5ec57176e817db2b13-ca-certs\") pod \"kube-controller-manager-ci-4344.0.0-4-f2dd3e9233\" (UID: \"6cd9fb9d21ceda5ec57176e817db2b13\") " pod="kube-system/kube-controller-manager-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:27.117951 kubelet[2688]: E0527 03:15:27.117886 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:27.118500 kubelet[2688]: E0527 03:15:27.118469 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:27.120480 kubelet[2688]: E0527 03:15:27.120433 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:27.425360 kubelet[2688]: I0527 03:15:27.425308 2688 apiserver.go:52] "Watching apiserver" May 27 03:15:27.483869 kubelet[2688]: I0527 03:15:27.483824 2688 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" May 27 03:15:27.568520 kubelet[2688]: E0527 03:15:27.568464 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:27.569746 kubelet[2688]: E0527 03:15:27.569642 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:27.580138 kubelet[2688]: W0527 03:15:27.580028 2688 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 27 03:15:27.580138 kubelet[2688]: E0527 03:15:27.580135 2688 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4344.0.0-4-f2dd3e9233\" already exists" pod="kube-system/kube-apiserver-ci-4344.0.0-4-f2dd3e9233" May 27 03:15:27.580479 kubelet[2688]: E0527 03:15:27.580372 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:27.606627 kubelet[2688]: I0527 03:15:27.606527 2688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4344.0.0-4-f2dd3e9233" podStartSLOduration=3.606467006 podStartE2EDuration="3.606467006s" podCreationTimestamp="2025-05-27 03:15:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-27 03:15:27.606365904 +0000 UTC m=+1.307978220" watchObservedRunningTime="2025-05-27 03:15:27.606467006 +0000 UTC m=+1.308079312" May 27 03:15:27.630852 kubelet[2688]: I0527 03:15:27.630771 2688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4344.0.0-4-f2dd3e9233" podStartSLOduration=4.630635482 podStartE2EDuration="4.630635482s" podCreationTimestamp="2025-05-27 03:15:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-27 03:15:27.618405237 +0000 UTC m=+1.320017550" watchObservedRunningTime="2025-05-27 03:15:27.630635482 +0000 UTC m=+1.332247800" May 27 03:15:27.631722 kubelet[2688]: I0527 03:15:27.631476 2688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4344.0.0-4-f2dd3e9233" podStartSLOduration=4.6314621030000005 podStartE2EDuration="4.631462103s" podCreationTimestamp="2025-05-27 03:15:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-27 03:15:27.62977246 +0000 UTC m=+1.331384776" watchObservedRunningTime="2025-05-27 03:15:27.631462103 +0000 UTC m=+1.333074414" May 27 03:15:28.571193 kubelet[2688]: E0527 03:15:28.571140 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:30.166089 kubelet[2688]: I0527 03:15:30.165781 2688 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 27 03:15:30.167831 containerd[1545]: time="2025-05-27T03:15:30.166722662Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 27 03:15:30.168573 kubelet[2688]: I0527 03:15:30.167124 2688 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 27 03:15:31.100391 systemd[1]: Created slice kubepods-besteffort-pod12c8709d_6a56_4c40_849a_a7a6ef7e76da.slice - libcontainer container kubepods-besteffort-pod12c8709d_6a56_4c40_849a_a7a6ef7e76da.slice. May 27 03:15:31.117877 kubelet[2688]: I0527 03:15:31.117649 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/12c8709d-6a56-4c40-849a-a7a6ef7e76da-kube-proxy\") pod \"kube-proxy-ppsbg\" (UID: \"12c8709d-6a56-4c40-849a-a7a6ef7e76da\") " pod="kube-system/kube-proxy-ppsbg" May 27 03:15:31.117877 kubelet[2688]: I0527 03:15:31.117718 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/12c8709d-6a56-4c40-849a-a7a6ef7e76da-lib-modules\") pod \"kube-proxy-ppsbg\" (UID: \"12c8709d-6a56-4c40-849a-a7a6ef7e76da\") " pod="kube-system/kube-proxy-ppsbg" May 27 03:15:31.117877 kubelet[2688]: I0527 03:15:31.117746 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmpnz\" (UniqueName: \"kubernetes.io/projected/12c8709d-6a56-4c40-849a-a7a6ef7e76da-kube-api-access-zmpnz\") pod \"kube-proxy-ppsbg\" (UID: \"12c8709d-6a56-4c40-849a-a7a6ef7e76da\") " pod="kube-system/kube-proxy-ppsbg" May 27 03:15:31.117877 kubelet[2688]: I0527 03:15:31.117792 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/12c8709d-6a56-4c40-849a-a7a6ef7e76da-xtables-lock\") pod \"kube-proxy-ppsbg\" (UID: \"12c8709d-6a56-4c40-849a-a7a6ef7e76da\") " pod="kube-system/kube-proxy-ppsbg" May 27 03:15:31.233867 systemd[1]: Created slice kubepods-besteffort-pod8cd3722f_4887_4b0c_b6b3_420c2485f64f.slice - libcontainer container kubepods-besteffort-pod8cd3722f_4887_4b0c_b6b3_420c2485f64f.slice. May 27 03:15:31.319519 kubelet[2688]: I0527 03:15:31.319452 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/8cd3722f-4887-4b0c-b6b3-420c2485f64f-var-lib-calico\") pod \"tigera-operator-7c5755cdcb-6p46c\" (UID: \"8cd3722f-4887-4b0c-b6b3-420c2485f64f\") " pod="tigera-operator/tigera-operator-7c5755cdcb-6p46c" May 27 03:15:31.319519 kubelet[2688]: I0527 03:15:31.319528 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmhbj\" (UniqueName: \"kubernetes.io/projected/8cd3722f-4887-4b0c-b6b3-420c2485f64f-kube-api-access-vmhbj\") pod \"tigera-operator-7c5755cdcb-6p46c\" (UID: \"8cd3722f-4887-4b0c-b6b3-420c2485f64f\") " pod="tigera-operator/tigera-operator-7c5755cdcb-6p46c" May 27 03:15:31.410580 kubelet[2688]: E0527 03:15:31.409697 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:31.411462 containerd[1545]: time="2025-05-27T03:15:31.411388277Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-ppsbg,Uid:12c8709d-6a56-4c40-849a-a7a6ef7e76da,Namespace:kube-system,Attempt:0,}" May 27 03:15:31.455481 containerd[1545]: time="2025-05-27T03:15:31.455433612Z" level=info msg="connecting to shim d0f4d8eea823235d3aed1956b29575bfd4524fa3f9e0b006c09db4f374ff06d4" address="unix:///run/containerd/s/614c0e49afdcaf6e5245c023847c78a67547f0f2a669943095620831e7097216" namespace=k8s.io protocol=ttrpc version=3 May 27 03:15:31.505435 systemd[1]: Started cri-containerd-d0f4d8eea823235d3aed1956b29575bfd4524fa3f9e0b006c09db4f374ff06d4.scope - libcontainer container d0f4d8eea823235d3aed1956b29575bfd4524fa3f9e0b006c09db4f374ff06d4. May 27 03:15:31.541945 containerd[1545]: time="2025-05-27T03:15:31.541822951Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7c5755cdcb-6p46c,Uid:8cd3722f-4887-4b0c-b6b3-420c2485f64f,Namespace:tigera-operator,Attempt:0,}" May 27 03:15:31.543910 containerd[1545]: time="2025-05-27T03:15:31.543856315Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-ppsbg,Uid:12c8709d-6a56-4c40-849a-a7a6ef7e76da,Namespace:kube-system,Attempt:0,} returns sandbox id \"d0f4d8eea823235d3aed1956b29575bfd4524fa3f9e0b006c09db4f374ff06d4\"" May 27 03:15:31.546873 kubelet[2688]: E0527 03:15:31.545958 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:31.552346 containerd[1545]: time="2025-05-27T03:15:31.552304537Z" level=info msg="CreateContainer within sandbox \"d0f4d8eea823235d3aed1956b29575bfd4524fa3f9e0b006c09db4f374ff06d4\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 27 03:15:31.573493 containerd[1545]: time="2025-05-27T03:15:31.572821308Z" level=info msg="Container f3e6da65be7fef22443495a017551aff100b992dcbe6a7139ea7a89330ccd3f9: CDI devices from CRI Config.CDIDevices: []" May 27 03:15:31.577497 containerd[1545]: time="2025-05-27T03:15:31.577435475Z" level=info msg="connecting to shim b393392d6eade1a33f44134bdff84a430e11d65d93eec0d7398332fcf86dae9b" address="unix:///run/containerd/s/565ca70a3e73e73cf0dbb4035ff5e3569043b39a2349b325b64f5b6f7e256b76" namespace=k8s.io protocol=ttrpc version=3 May 27 03:15:31.582230 containerd[1545]: time="2025-05-27T03:15:31.582174083Z" level=info msg="CreateContainer within sandbox \"d0f4d8eea823235d3aed1956b29575bfd4524fa3f9e0b006c09db4f374ff06d4\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"f3e6da65be7fef22443495a017551aff100b992dcbe6a7139ea7a89330ccd3f9\"" May 27 03:15:31.582991 containerd[1545]: time="2025-05-27T03:15:31.582849290Z" level=info msg="StartContainer for \"f3e6da65be7fef22443495a017551aff100b992dcbe6a7139ea7a89330ccd3f9\"" May 27 03:15:31.589306 containerd[1545]: time="2025-05-27T03:15:31.589265102Z" level=info msg="connecting to shim f3e6da65be7fef22443495a017551aff100b992dcbe6a7139ea7a89330ccd3f9" address="unix:///run/containerd/s/614c0e49afdcaf6e5245c023847c78a67547f0f2a669943095620831e7097216" protocol=ttrpc version=3 May 27 03:15:31.615465 systemd[1]: Started cri-containerd-f3e6da65be7fef22443495a017551aff100b992dcbe6a7139ea7a89330ccd3f9.scope - libcontainer container f3e6da65be7fef22443495a017551aff100b992dcbe6a7139ea7a89330ccd3f9. May 27 03:15:31.630342 systemd[1]: Started cri-containerd-b393392d6eade1a33f44134bdff84a430e11d65d93eec0d7398332fcf86dae9b.scope - libcontainer container b393392d6eade1a33f44134bdff84a430e11d65d93eec0d7398332fcf86dae9b. May 27 03:15:31.696360 containerd[1545]: time="2025-05-27T03:15:31.696049243Z" level=info msg="StartContainer for \"f3e6da65be7fef22443495a017551aff100b992dcbe6a7139ea7a89330ccd3f9\" returns successfully" May 27 03:15:31.723902 containerd[1545]: time="2025-05-27T03:15:31.723671564Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7c5755cdcb-6p46c,Uid:8cd3722f-4887-4b0c-b6b3-420c2485f64f,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"b393392d6eade1a33f44134bdff84a430e11d65d93eec0d7398332fcf86dae9b\"" May 27 03:15:31.731790 containerd[1545]: time="2025-05-27T03:15:31.731666488Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.0\"" May 27 03:15:31.735099 systemd-resolved[1402]: Using degraded feature set TCP instead of UDP for DNS server 67.207.67.2. May 27 03:15:32.591290 kubelet[2688]: E0527 03:15:32.591097 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:32.604179 kubelet[2688]: I0527 03:15:32.604106 2688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-ppsbg" podStartSLOduration=1.6040838609999999 podStartE2EDuration="1.604083861s" podCreationTimestamp="2025-05-27 03:15:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-27 03:15:32.602760475 +0000 UTC m=+6.304372789" watchObservedRunningTime="2025-05-27 03:15:32.604083861 +0000 UTC m=+6.305696175" May 27 03:15:33.018013 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1235248690.mount: Deactivated successfully. May 27 03:15:33.301370 kubelet[2688]: E0527 03:15:33.300913 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:33.592098 kubelet[2688]: E0527 03:15:33.591094 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:33.592098 kubelet[2688]: E0527 03:15:33.591607 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:33.695397 kubelet[2688]: E0527 03:15:33.695225 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:33.828994 containerd[1545]: time="2025-05-27T03:15:33.827863778Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:33.828994 containerd[1545]: time="2025-05-27T03:15:33.828565306Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.0: active requests=0, bytes read=25055451" May 27 03:15:33.828994 containerd[1545]: time="2025-05-27T03:15:33.828875820Z" level=info msg="ImageCreate event name:\"sha256:5e43c1322619406528ff596056dfeb70cb8d20c5c00439feb752a7725302e033\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:33.831033 containerd[1545]: time="2025-05-27T03:15:33.830988886Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:33.831640 containerd[1545]: time="2025-05-27T03:15:33.831601886Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.0\" with image id \"sha256:5e43c1322619406528ff596056dfeb70cb8d20c5c00439feb752a7725302e033\", repo tag \"quay.io/tigera/operator:v1.38.0\", repo digest \"quay.io/tigera/operator@sha256:e0a34b265aebce1a2db906d8dad99190706e8bf3910cae626b9c2eb6bbb21775\", size \"25051446\" in 2.099458044s" May 27 03:15:33.831640 containerd[1545]: time="2025-05-27T03:15:33.831637292Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.0\" returns image reference \"sha256:5e43c1322619406528ff596056dfeb70cb8d20c5c00439feb752a7725302e033\"" May 27 03:15:33.834545 containerd[1545]: time="2025-05-27T03:15:33.834508312Z" level=info msg="CreateContainer within sandbox \"b393392d6eade1a33f44134bdff84a430e11d65d93eec0d7398332fcf86dae9b\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" May 27 03:15:33.839243 containerd[1545]: time="2025-05-27T03:15:33.839187855Z" level=info msg="Container a1878a94f6242155f88db01bbb4d2c7baf3a086a0c9bcec9e4a1b4d31fbc293f: CDI devices from CRI Config.CDIDevices: []" May 27 03:15:33.854745 containerd[1545]: time="2025-05-27T03:15:33.854228085Z" level=info msg="CreateContainer within sandbox \"b393392d6eade1a33f44134bdff84a430e11d65d93eec0d7398332fcf86dae9b\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"a1878a94f6242155f88db01bbb4d2c7baf3a086a0c9bcec9e4a1b4d31fbc293f\"" May 27 03:15:33.856116 containerd[1545]: time="2025-05-27T03:15:33.855348907Z" level=info msg="StartContainer for \"a1878a94f6242155f88db01bbb4d2c7baf3a086a0c9bcec9e4a1b4d31fbc293f\"" May 27 03:15:33.856469 containerd[1545]: time="2025-05-27T03:15:33.856444671Z" level=info msg="connecting to shim a1878a94f6242155f88db01bbb4d2c7baf3a086a0c9bcec9e4a1b4d31fbc293f" address="unix:///run/containerd/s/565ca70a3e73e73cf0dbb4035ff5e3569043b39a2349b325b64f5b6f7e256b76" protocol=ttrpc version=3 May 27 03:15:33.882303 systemd[1]: Started cri-containerd-a1878a94f6242155f88db01bbb4d2c7baf3a086a0c9bcec9e4a1b4d31fbc293f.scope - libcontainer container a1878a94f6242155f88db01bbb4d2c7baf3a086a0c9bcec9e4a1b4d31fbc293f. May 27 03:15:33.918304 containerd[1545]: time="2025-05-27T03:15:33.918237332Z" level=info msg="StartContainer for \"a1878a94f6242155f88db01bbb4d2c7baf3a086a0c9bcec9e4a1b4d31fbc293f\" returns successfully" May 27 03:15:34.232351 kubelet[2688]: E0527 03:15:34.232214 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:34.596098 kubelet[2688]: E0527 03:15:34.595864 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:34.597263 kubelet[2688]: E0527 03:15:34.597236 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:34.608757 kubelet[2688]: I0527 03:15:34.608456 2688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7c5755cdcb-6p46c" podStartSLOduration=1.501571123 podStartE2EDuration="3.608436469s" podCreationTimestamp="2025-05-27 03:15:31 +0000 UTC" firstStartedPulling="2025-05-27 03:15:31.725949704 +0000 UTC m=+5.427561997" lastFinishedPulling="2025-05-27 03:15:33.83281505 +0000 UTC m=+7.534427343" observedRunningTime="2025-05-27 03:15:34.608430961 +0000 UTC m=+8.310043277" watchObservedRunningTime="2025-05-27 03:15:34.608436469 +0000 UTC m=+8.310048783" May 27 03:15:37.710258 systemd[1]: cri-containerd-a1878a94f6242155f88db01bbb4d2c7baf3a086a0c9bcec9e4a1b4d31fbc293f.scope: Deactivated successfully. May 27 03:15:37.724396 containerd[1545]: time="2025-05-27T03:15:37.724221245Z" level=info msg="received exit event container_id:\"a1878a94f6242155f88db01bbb4d2c7baf3a086a0c9bcec9e4a1b4d31fbc293f\" id:\"a1878a94f6242155f88db01bbb4d2c7baf3a086a0c9bcec9e4a1b4d31fbc293f\" pid:3002 exit_status:1 exited_at:{seconds:1748315737 nanos:723513675}" May 27 03:15:37.726267 containerd[1545]: time="2025-05-27T03:15:37.726205090Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a1878a94f6242155f88db01bbb4d2c7baf3a086a0c9bcec9e4a1b4d31fbc293f\" id:\"a1878a94f6242155f88db01bbb4d2c7baf3a086a0c9bcec9e4a1b4d31fbc293f\" pid:3002 exit_status:1 exited_at:{seconds:1748315737 nanos:723513675}" May 27 03:15:37.770246 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a1878a94f6242155f88db01bbb4d2c7baf3a086a0c9bcec9e4a1b4d31fbc293f-rootfs.mount: Deactivated successfully. May 27 03:15:38.613491 kubelet[2688]: I0527 03:15:38.613452 2688 scope.go:117] "RemoveContainer" containerID="a1878a94f6242155f88db01bbb4d2c7baf3a086a0c9bcec9e4a1b4d31fbc293f" May 27 03:15:38.617455 containerd[1545]: time="2025-05-27T03:15:38.617386373Z" level=info msg="CreateContainer within sandbox \"b393392d6eade1a33f44134bdff84a430e11d65d93eec0d7398332fcf86dae9b\" for container &ContainerMetadata{Name:tigera-operator,Attempt:1,}" May 27 03:15:38.625638 containerd[1545]: time="2025-05-27T03:15:38.625426011Z" level=info msg="Container 821dc3247113cf94599068f922474152a2e0a4cac653217ece13ab79cd872b9f: CDI devices from CRI Config.CDIDevices: []" May 27 03:15:38.637946 containerd[1545]: time="2025-05-27T03:15:38.637875910Z" level=info msg="CreateContainer within sandbox \"b393392d6eade1a33f44134bdff84a430e11d65d93eec0d7398332fcf86dae9b\" for &ContainerMetadata{Name:tigera-operator,Attempt:1,} returns container id \"821dc3247113cf94599068f922474152a2e0a4cac653217ece13ab79cd872b9f\"" May 27 03:15:38.639150 containerd[1545]: time="2025-05-27T03:15:38.639015167Z" level=info msg="StartContainer for \"821dc3247113cf94599068f922474152a2e0a4cac653217ece13ab79cd872b9f\"" May 27 03:15:38.642903 containerd[1545]: time="2025-05-27T03:15:38.642821536Z" level=info msg="connecting to shim 821dc3247113cf94599068f922474152a2e0a4cac653217ece13ab79cd872b9f" address="unix:///run/containerd/s/565ca70a3e73e73cf0dbb4035ff5e3569043b39a2349b325b64f5b6f7e256b76" protocol=ttrpc version=3 May 27 03:15:38.697489 systemd[1]: Started cri-containerd-821dc3247113cf94599068f922474152a2e0a4cac653217ece13ab79cd872b9f.scope - libcontainer container 821dc3247113cf94599068f922474152a2e0a4cac653217ece13ab79cd872b9f. May 27 03:15:38.739298 containerd[1545]: time="2025-05-27T03:15:38.739237927Z" level=info msg="StartContainer for \"821dc3247113cf94599068f922474152a2e0a4cac653217ece13ab79cd872b9f\" returns successfully" May 27 03:15:40.212888 update_engine[1524]: I20250527 03:15:40.211114 1524 update_attempter.cc:509] Updating boot flags... May 27 03:15:40.971261 sudo[1774]: pam_unix(sudo:session): session closed for user root May 27 03:15:40.975600 sshd[1773]: Connection closed by 139.178.68.195 port 41608 May 27 03:15:40.976654 sshd-session[1771]: pam_unix(sshd:session): session closed for user core May 27 03:15:40.981985 systemd[1]: sshd@6-209.38.65.15:22-139.178.68.195:41608.service: Deactivated successfully. May 27 03:15:40.987025 systemd[1]: session-7.scope: Deactivated successfully. May 27 03:15:40.987695 systemd[1]: session-7.scope: Consumed 5.875s CPU time, 164.2M memory peak. May 27 03:15:40.990023 systemd-logind[1522]: Session 7 logged out. Waiting for processes to exit. May 27 03:15:40.995605 systemd-logind[1522]: Removed session 7. May 27 03:15:47.439986 systemd[1]: Created slice kubepods-besteffort-pod5293fd53_9cd5_420e_90dd_7c4374228794.slice - libcontainer container kubepods-besteffort-pod5293fd53_9cd5_420e_90dd_7c4374228794.slice. May 27 03:15:47.527097 kubelet[2688]: I0527 03:15:47.525769 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5293fd53-9cd5-420e-90dd-7c4374228794-tigera-ca-bundle\") pod \"calico-typha-7d9f759884-9q2bq\" (UID: \"5293fd53-9cd5-420e-90dd-7c4374228794\") " pod="calico-system/calico-typha-7d9f759884-9q2bq" May 27 03:15:47.527097 kubelet[2688]: I0527 03:15:47.525816 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/5293fd53-9cd5-420e-90dd-7c4374228794-typha-certs\") pod \"calico-typha-7d9f759884-9q2bq\" (UID: \"5293fd53-9cd5-420e-90dd-7c4374228794\") " pod="calico-system/calico-typha-7d9f759884-9q2bq" May 27 03:15:47.527097 kubelet[2688]: I0527 03:15:47.525838 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2q7f\" (UniqueName: \"kubernetes.io/projected/5293fd53-9cd5-420e-90dd-7c4374228794-kube-api-access-j2q7f\") pod \"calico-typha-7d9f759884-9q2bq\" (UID: \"5293fd53-9cd5-420e-90dd-7c4374228794\") " pod="calico-system/calico-typha-7d9f759884-9q2bq" May 27 03:15:47.611344 systemd[1]: Created slice kubepods-besteffort-pod025145e5_a803_4f04_a919_91b66f1f8f64.slice - libcontainer container kubepods-besteffort-pod025145e5_a803_4f04_a919_91b66f1f8f64.slice. May 27 03:15:47.626769 kubelet[2688]: I0527 03:15:47.626701 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/025145e5-a803-4f04-a919-91b66f1f8f64-policysync\") pod \"calico-node-2z8x7\" (UID: \"025145e5-a803-4f04-a919-91b66f1f8f64\") " pod="calico-system/calico-node-2z8x7" May 27 03:15:47.626769 kubelet[2688]: I0527 03:15:47.626769 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/025145e5-a803-4f04-a919-91b66f1f8f64-tigera-ca-bundle\") pod \"calico-node-2z8x7\" (UID: \"025145e5-a803-4f04-a919-91b66f1f8f64\") " pod="calico-system/calico-node-2z8x7" May 27 03:15:47.627036 kubelet[2688]: I0527 03:15:47.626795 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/025145e5-a803-4f04-a919-91b66f1f8f64-lib-modules\") pod \"calico-node-2z8x7\" (UID: \"025145e5-a803-4f04-a919-91b66f1f8f64\") " pod="calico-system/calico-node-2z8x7" May 27 03:15:47.627036 kubelet[2688]: I0527 03:15:47.626819 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/025145e5-a803-4f04-a919-91b66f1f8f64-xtables-lock\") pod \"calico-node-2z8x7\" (UID: \"025145e5-a803-4f04-a919-91b66f1f8f64\") " pod="calico-system/calico-node-2z8x7" May 27 03:15:47.627036 kubelet[2688]: I0527 03:15:47.626841 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/025145e5-a803-4f04-a919-91b66f1f8f64-cni-log-dir\") pod \"calico-node-2z8x7\" (UID: \"025145e5-a803-4f04-a919-91b66f1f8f64\") " pod="calico-system/calico-node-2z8x7" May 27 03:15:47.627036 kubelet[2688]: I0527 03:15:47.626865 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzpz7\" (UniqueName: \"kubernetes.io/projected/025145e5-a803-4f04-a919-91b66f1f8f64-kube-api-access-mzpz7\") pod \"calico-node-2z8x7\" (UID: \"025145e5-a803-4f04-a919-91b66f1f8f64\") " pod="calico-system/calico-node-2z8x7" May 27 03:15:47.627036 kubelet[2688]: I0527 03:15:47.626908 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/025145e5-a803-4f04-a919-91b66f1f8f64-var-run-calico\") pod \"calico-node-2z8x7\" (UID: \"025145e5-a803-4f04-a919-91b66f1f8f64\") " pod="calico-system/calico-node-2z8x7" May 27 03:15:47.628150 kubelet[2688]: I0527 03:15:47.626927 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/025145e5-a803-4f04-a919-91b66f1f8f64-var-lib-calico\") pod \"calico-node-2z8x7\" (UID: \"025145e5-a803-4f04-a919-91b66f1f8f64\") " pod="calico-system/calico-node-2z8x7" May 27 03:15:47.628150 kubelet[2688]: I0527 03:15:47.626947 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/025145e5-a803-4f04-a919-91b66f1f8f64-cni-bin-dir\") pod \"calico-node-2z8x7\" (UID: \"025145e5-a803-4f04-a919-91b66f1f8f64\") " pod="calico-system/calico-node-2z8x7" May 27 03:15:47.628150 kubelet[2688]: I0527 03:15:47.626960 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/025145e5-a803-4f04-a919-91b66f1f8f64-cni-net-dir\") pod \"calico-node-2z8x7\" (UID: \"025145e5-a803-4f04-a919-91b66f1f8f64\") " pod="calico-system/calico-node-2z8x7" May 27 03:15:47.628150 kubelet[2688]: I0527 03:15:47.626977 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/025145e5-a803-4f04-a919-91b66f1f8f64-node-certs\") pod \"calico-node-2z8x7\" (UID: \"025145e5-a803-4f04-a919-91b66f1f8f64\") " pod="calico-system/calico-node-2z8x7" May 27 03:15:47.628150 kubelet[2688]: I0527 03:15:47.626995 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/025145e5-a803-4f04-a919-91b66f1f8f64-flexvol-driver-host\") pod \"calico-node-2z8x7\" (UID: \"025145e5-a803-4f04-a919-91b66f1f8f64\") " pod="calico-system/calico-node-2z8x7" May 27 03:15:47.732646 kubelet[2688]: E0527 03:15:47.731140 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.732646 kubelet[2688]: W0527 03:15:47.731203 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.732646 kubelet[2688]: E0527 03:15:47.731240 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.732646 kubelet[2688]: E0527 03:15:47.731609 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.732646 kubelet[2688]: W0527 03:15:47.731626 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.732646 kubelet[2688]: E0527 03:15:47.731660 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.732646 kubelet[2688]: E0527 03:15:47.731918 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.732646 kubelet[2688]: W0527 03:15:47.731931 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.732646 kubelet[2688]: E0527 03:15:47.731946 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.732646 kubelet[2688]: E0527 03:15:47.732377 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.733662 kubelet[2688]: W0527 03:15:47.732407 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.733662 kubelet[2688]: E0527 03:15:47.732427 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.744974 kubelet[2688]: E0527 03:15:47.744697 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.744974 kubelet[2688]: W0527 03:15:47.744736 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.744974 kubelet[2688]: E0527 03:15:47.744774 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.748733 kubelet[2688]: E0527 03:15:47.747850 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:47.750075 containerd[1545]: time="2025-05-27T03:15:47.750010997Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7d9f759884-9q2bq,Uid:5293fd53-9cd5-420e-90dd-7c4374228794,Namespace:calico-system,Attempt:0,}" May 27 03:15:47.768350 kubelet[2688]: E0527 03:15:47.768221 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.768350 kubelet[2688]: W0527 03:15:47.768255 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.768350 kubelet[2688]: E0527 03:15:47.768285 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.795400 containerd[1545]: time="2025-05-27T03:15:47.795223227Z" level=info msg="connecting to shim ee55f0a62775158f1c34d1b6d2ee6627522c319b0bff44d0eb27e9f096f850af" address="unix:///run/containerd/s/94d8c9516c56870c0bd771902b40e061d724c086f1c08e28a731e1b173a7d6b3" namespace=k8s.io protocol=ttrpc version=3 May 27 03:15:47.849298 systemd[1]: Started cri-containerd-ee55f0a62775158f1c34d1b6d2ee6627522c319b0bff44d0eb27e9f096f850af.scope - libcontainer container ee55f0a62775158f1c34d1b6d2ee6627522c319b0bff44d0eb27e9f096f850af. May 27 03:15:47.867453 kubelet[2688]: E0527 03:15:47.867250 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-drsns" podUID="851b262b-7562-4011-b67e-5e2de773f0d3" May 27 03:15:47.920311 containerd[1545]: time="2025-05-27T03:15:47.920226946Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-2z8x7,Uid:025145e5-a803-4f04-a919-91b66f1f8f64,Namespace:calico-system,Attempt:0,}" May 27 03:15:47.934479 kubelet[2688]: E0527 03:15:47.932617 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.934479 kubelet[2688]: W0527 03:15:47.932652 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.934479 kubelet[2688]: E0527 03:15:47.932696 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.934479 kubelet[2688]: E0527 03:15:47.933542 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.934479 kubelet[2688]: W0527 03:15:47.933563 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.934479 kubelet[2688]: E0527 03:15:47.933587 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.934479 kubelet[2688]: E0527 03:15:47.934123 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.934479 kubelet[2688]: W0527 03:15:47.934135 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.934479 kubelet[2688]: E0527 03:15:47.934151 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.934479 kubelet[2688]: E0527 03:15:47.934493 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.935838 kubelet[2688]: W0527 03:15:47.934503 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.935838 kubelet[2688]: E0527 03:15:47.934517 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.935838 kubelet[2688]: E0527 03:15:47.934775 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.935838 kubelet[2688]: W0527 03:15:47.934783 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.935838 kubelet[2688]: E0527 03:15:47.934805 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.935838 kubelet[2688]: E0527 03:15:47.935070 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.935838 kubelet[2688]: W0527 03:15:47.935081 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.935838 kubelet[2688]: E0527 03:15:47.935097 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.936219 kubelet[2688]: E0527 03:15:47.935988 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.936219 kubelet[2688]: W0527 03:15:47.935999 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.936219 kubelet[2688]: E0527 03:15:47.936019 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.936346 kubelet[2688]: E0527 03:15:47.936291 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.936346 kubelet[2688]: W0527 03:15:47.936315 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.936346 kubelet[2688]: E0527 03:15:47.936330 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.939190 kubelet[2688]: E0527 03:15:47.936658 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.939190 kubelet[2688]: W0527 03:15:47.936673 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.939190 kubelet[2688]: E0527 03:15:47.936687 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.939190 kubelet[2688]: E0527 03:15:47.937439 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.939190 kubelet[2688]: W0527 03:15:47.937457 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.939190 kubelet[2688]: E0527 03:15:47.937474 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.939190 kubelet[2688]: E0527 03:15:47.937897 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.939190 kubelet[2688]: W0527 03:15:47.937909 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.939190 kubelet[2688]: E0527 03:15:47.937956 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.939190 kubelet[2688]: E0527 03:15:47.938510 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.939575 kubelet[2688]: W0527 03:15:47.938521 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.939575 kubelet[2688]: E0527 03:15:47.938538 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.944387 kubelet[2688]: E0527 03:15:47.944319 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.944387 kubelet[2688]: W0527 03:15:47.944365 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.944711 kubelet[2688]: E0527 03:15:47.944402 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.949711 kubelet[2688]: E0527 03:15:47.949394 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.949711 kubelet[2688]: W0527 03:15:47.949431 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.949711 kubelet[2688]: E0527 03:15:47.949468 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.952031 kubelet[2688]: E0527 03:15:47.951988 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.952031 kubelet[2688]: W0527 03:15:47.952025 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.952233 kubelet[2688]: E0527 03:15:47.952078 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.952444 kubelet[2688]: E0527 03:15:47.952411 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.952444 kubelet[2688]: W0527 03:15:47.952429 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.952523 kubelet[2688]: E0527 03:15:47.952447 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.952716 kubelet[2688]: E0527 03:15:47.952700 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.952752 kubelet[2688]: W0527 03:15:47.952716 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.952752 kubelet[2688]: E0527 03:15:47.952732 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.953223 kubelet[2688]: E0527 03:15:47.953197 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.953223 kubelet[2688]: W0527 03:15:47.953220 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.953374 kubelet[2688]: E0527 03:15:47.953242 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.961604 kubelet[2688]: E0527 03:15:47.961534 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.961947 kubelet[2688]: W0527 03:15:47.961760 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.961947 kubelet[2688]: E0527 03:15:47.961797 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:47.964343 kubelet[2688]: E0527 03:15:47.964311 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:47.964558 kubelet[2688]: W0527 03:15:47.964492 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:47.964558 kubelet[2688]: E0527 03:15:47.964525 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.011474 containerd[1545]: time="2025-05-27T03:15:48.011287172Z" level=info msg="connecting to shim 6c251db89611d929b5891e7f51c221935e5f85890e558310106fd72966d833af" address="unix:///run/containerd/s/604f30d5500ea97da73323276f3fde0f21d6cc0527a8e9361235c40e7e69416a" namespace=k8s.io protocol=ttrpc version=3 May 27 03:15:48.032324 kubelet[2688]: E0527 03:15:48.032138 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.033261 kubelet[2688]: W0527 03:15:48.033202 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.034932 kubelet[2688]: E0527 03:15:48.034103 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.035707 kubelet[2688]: I0527 03:15:48.034465 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2b9k\" (UniqueName: \"kubernetes.io/projected/851b262b-7562-4011-b67e-5e2de773f0d3-kube-api-access-z2b9k\") pod \"csi-node-driver-drsns\" (UID: \"851b262b-7562-4011-b67e-5e2de773f0d3\") " pod="calico-system/csi-node-driver-drsns" May 27 03:15:48.036014 kubelet[2688]: E0527 03:15:48.035971 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.036444 kubelet[2688]: W0527 03:15:48.035994 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.036761 kubelet[2688]: E0527 03:15:48.036359 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.037414 kubelet[2688]: E0527 03:15:48.036788 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.037414 kubelet[2688]: W0527 03:15:48.036813 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.037414 kubelet[2688]: E0527 03:15:48.036843 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.037414 kubelet[2688]: I0527 03:15:48.036884 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/851b262b-7562-4011-b67e-5e2de773f0d3-kubelet-dir\") pod \"csi-node-driver-drsns\" (UID: \"851b262b-7562-4011-b67e-5e2de773f0d3\") " pod="calico-system/csi-node-driver-drsns" May 27 03:15:48.038189 kubelet[2688]: E0527 03:15:48.037536 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.038189 kubelet[2688]: W0527 03:15:48.037556 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.038189 kubelet[2688]: E0527 03:15:48.037579 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.038189 kubelet[2688]: E0527 03:15:48.037903 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.038189 kubelet[2688]: W0527 03:15:48.037922 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.038189 kubelet[2688]: E0527 03:15:48.037938 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.039755 kubelet[2688]: E0527 03:15:48.038572 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.039755 kubelet[2688]: W0527 03:15:48.038588 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.039755 kubelet[2688]: E0527 03:15:48.038614 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.039755 kubelet[2688]: E0527 03:15:48.039149 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.039755 kubelet[2688]: W0527 03:15:48.039164 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.039755 kubelet[2688]: E0527 03:15:48.039183 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.039755 kubelet[2688]: I0527 03:15:48.039225 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/851b262b-7562-4011-b67e-5e2de773f0d3-registration-dir\") pod \"csi-node-driver-drsns\" (UID: \"851b262b-7562-4011-b67e-5e2de773f0d3\") " pod="calico-system/csi-node-driver-drsns" May 27 03:15:48.040444 kubelet[2688]: E0527 03:15:48.039777 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.040444 kubelet[2688]: W0527 03:15:48.039794 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.040444 kubelet[2688]: E0527 03:15:48.039819 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.040444 kubelet[2688]: I0527 03:15:48.039853 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/851b262b-7562-4011-b67e-5e2de773f0d3-socket-dir\") pod \"csi-node-driver-drsns\" (UID: \"851b262b-7562-4011-b67e-5e2de773f0d3\") " pod="calico-system/csi-node-driver-drsns" May 27 03:15:48.042737 kubelet[2688]: E0527 03:15:48.042416 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.043569 kubelet[2688]: W0527 03:15:48.042633 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.044255 kubelet[2688]: E0527 03:15:48.043018 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.045378 kubelet[2688]: E0527 03:15:48.045234 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.045378 kubelet[2688]: W0527 03:15:48.045295 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.046122 kubelet[2688]: E0527 03:15:48.045619 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.046122 kubelet[2688]: I0527 03:15:48.045701 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/851b262b-7562-4011-b67e-5e2de773f0d3-varrun\") pod \"csi-node-driver-drsns\" (UID: \"851b262b-7562-4011-b67e-5e2de773f0d3\") " pod="calico-system/csi-node-driver-drsns" May 27 03:15:48.046122 kubelet[2688]: E0527 03:15:48.045826 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.046122 kubelet[2688]: W0527 03:15:48.045847 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.046122 kubelet[2688]: E0527 03:15:48.045872 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.046930 kubelet[2688]: E0527 03:15:48.046629 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.046930 kubelet[2688]: W0527 03:15:48.046651 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.046930 kubelet[2688]: E0527 03:15:48.046680 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.047275 kubelet[2688]: E0527 03:15:48.047004 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.047275 kubelet[2688]: W0527 03:15:48.047017 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.047275 kubelet[2688]: E0527 03:15:48.047037 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.048299 kubelet[2688]: E0527 03:15:48.048192 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.048299 kubelet[2688]: W0527 03:15:48.048210 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.048299 kubelet[2688]: E0527 03:15:48.048231 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.049229 kubelet[2688]: E0527 03:15:48.048483 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.049229 kubelet[2688]: W0527 03:15:48.048495 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.049229 kubelet[2688]: E0527 03:15:48.048508 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.065001 containerd[1545]: time="2025-05-27T03:15:48.064877851Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7d9f759884-9q2bq,Uid:5293fd53-9cd5-420e-90dd-7c4374228794,Namespace:calico-system,Attempt:0,} returns sandbox id \"ee55f0a62775158f1c34d1b6d2ee6627522c319b0bff44d0eb27e9f096f850af\"" May 27 03:15:48.069729 kubelet[2688]: E0527 03:15:48.069290 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:48.074510 containerd[1545]: time="2025-05-27T03:15:48.074028827Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.0\"" May 27 03:15:48.112929 systemd[1]: Started cri-containerd-6c251db89611d929b5891e7f51c221935e5f85890e558310106fd72966d833af.scope - libcontainer container 6c251db89611d929b5891e7f51c221935e5f85890e558310106fd72966d833af. May 27 03:15:48.148302 kubelet[2688]: E0527 03:15:48.148247 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.148591 kubelet[2688]: W0527 03:15:48.148537 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.148855 kubelet[2688]: E0527 03:15:48.148798 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.150304 kubelet[2688]: E0527 03:15:48.150254 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.150304 kubelet[2688]: W0527 03:15:48.150286 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.150720 kubelet[2688]: E0527 03:15:48.150330 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.151231 kubelet[2688]: E0527 03:15:48.151199 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.151231 kubelet[2688]: W0527 03:15:48.151225 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.153163 kubelet[2688]: E0527 03:15:48.151342 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.153261 kubelet[2688]: E0527 03:15:48.153221 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.153261 kubelet[2688]: W0527 03:15:48.153246 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.153524 kubelet[2688]: E0527 03:15:48.153294 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.153656 kubelet[2688]: E0527 03:15:48.153643 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.153755 kubelet[2688]: W0527 03:15:48.153656 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.153885 kubelet[2688]: E0527 03:15:48.153706 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.154714 kubelet[2688]: E0527 03:15:48.154682 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.154714 kubelet[2688]: W0527 03:15:48.154707 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.154994 kubelet[2688]: E0527 03:15:48.154774 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.154994 kubelet[2688]: E0527 03:15:48.154974 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.154994 kubelet[2688]: W0527 03:15:48.154985 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.155452 kubelet[2688]: E0527 03:15:48.155211 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.155782 kubelet[2688]: E0527 03:15:48.155754 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.155782 kubelet[2688]: W0527 03:15:48.155776 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.155989 kubelet[2688]: E0527 03:15:48.155819 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.156574 kubelet[2688]: E0527 03:15:48.156551 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.156574 kubelet[2688]: W0527 03:15:48.156571 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.156776 kubelet[2688]: E0527 03:15:48.156654 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.157343 kubelet[2688]: E0527 03:15:48.157317 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.157343 kubelet[2688]: W0527 03:15:48.157339 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.157652 kubelet[2688]: E0527 03:15:48.157446 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.158151 kubelet[2688]: E0527 03:15:48.158023 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.158249 kubelet[2688]: W0527 03:15:48.158043 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.158496 kubelet[2688]: E0527 03:15:48.158401 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.159315 kubelet[2688]: E0527 03:15:48.159289 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.159315 kubelet[2688]: W0527 03:15:48.159310 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.159763 kubelet[2688]: E0527 03:15:48.159726 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.159763 kubelet[2688]: W0527 03:15:48.159741 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.160108 kubelet[2688]: E0527 03:15:48.159405 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.160295 kubelet[2688]: E0527 03:15:48.160238 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.161373 kubelet[2688]: E0527 03:15:48.161341 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.161373 kubelet[2688]: W0527 03:15:48.161365 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.161767 kubelet[2688]: E0527 03:15:48.161420 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.161767 kubelet[2688]: E0527 03:15:48.161598 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.161767 kubelet[2688]: W0527 03:15:48.161609 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.162865 kubelet[2688]: E0527 03:15:48.161791 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.162865 kubelet[2688]: W0527 03:15:48.161801 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.162865 kubelet[2688]: E0527 03:15:48.162097 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.162865 kubelet[2688]: W0527 03:15:48.162111 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.162865 kubelet[2688]: E0527 03:15:48.162557 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.162865 kubelet[2688]: E0527 03:15:48.162581 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.162865 kubelet[2688]: E0527 03:15:48.162591 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.162865 kubelet[2688]: E0527 03:15:48.162649 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.162865 kubelet[2688]: W0527 03:15:48.162664 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.162865 kubelet[2688]: E0527 03:15:48.162681 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.163452 kubelet[2688]: E0527 03:15:48.163083 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.163452 kubelet[2688]: W0527 03:15:48.163098 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.163452 kubelet[2688]: E0527 03:15:48.163233 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.163908 kubelet[2688]: E0527 03:15:48.163880 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.163908 kubelet[2688]: W0527 03:15:48.163900 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.164255 kubelet[2688]: E0527 03:15:48.163926 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.164469 kubelet[2688]: E0527 03:15:48.164439 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.164469 kubelet[2688]: W0527 03:15:48.164460 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.164936 kubelet[2688]: E0527 03:15:48.164505 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.167097 kubelet[2688]: E0527 03:15:48.165483 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.167382 kubelet[2688]: W0527 03:15:48.167338 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.167433 kubelet[2688]: E0527 03:15:48.167401 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.167867 kubelet[2688]: E0527 03:15:48.167821 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.167867 kubelet[2688]: W0527 03:15:48.167864 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.167974 kubelet[2688]: E0527 03:15:48.167958 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.168341 kubelet[2688]: E0527 03:15:48.168316 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.168341 kubelet[2688]: W0527 03:15:48.168337 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.168434 kubelet[2688]: E0527 03:15:48.168359 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.168754 kubelet[2688]: E0527 03:15:48.168729 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.168754 kubelet[2688]: W0527 03:15:48.168750 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.168857 kubelet[2688]: E0527 03:15:48.168766 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.179435 kubelet[2688]: E0527 03:15:48.179310 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:48.179435 kubelet[2688]: W0527 03:15:48.179343 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:48.179435 kubelet[2688]: E0527 03:15:48.179374 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:48.264987 containerd[1545]: time="2025-05-27T03:15:48.264278268Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-2z8x7,Uid:025145e5-a803-4f04-a919-91b66f1f8f64,Namespace:calico-system,Attempt:0,} returns sandbox id \"6c251db89611d929b5891e7f51c221935e5f85890e558310106fd72966d833af\"" May 27 03:15:49.330386 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3247619021.mount: Deactivated successfully. May 27 03:15:49.502839 kubelet[2688]: E0527 03:15:49.502772 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-drsns" podUID="851b262b-7562-4011-b67e-5e2de773f0d3" May 27 03:15:50.410092 containerd[1545]: time="2025-05-27T03:15:50.409992672Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:50.411009 containerd[1545]: time="2025-05-27T03:15:50.410826838Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.0: active requests=0, bytes read=35158669" May 27 03:15:50.411526 containerd[1545]: time="2025-05-27T03:15:50.411490305Z" level=info msg="ImageCreate event name:\"sha256:71be0570e8645ac646675719e0da6ac33a05810991b31aecc303e7add70933be\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:50.414002 containerd[1545]: time="2025-05-27T03:15:50.413899443Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:d282f6c773c4631b9dc8379eb093c54ca34c7728d55d6509cb45da5e1f5baf8f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:50.414592 containerd[1545]: time="2025-05-27T03:15:50.414399255Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.0\" with image id \"sha256:71be0570e8645ac646675719e0da6ac33a05810991b31aecc303e7add70933be\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.0\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:d282f6c773c4631b9dc8379eb093c54ca34c7728d55d6509cb45da5e1f5baf8f\", size \"35158523\" in 2.340100084s" May 27 03:15:50.414592 containerd[1545]: time="2025-05-27T03:15:50.414542052Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.0\" returns image reference \"sha256:71be0570e8645ac646675719e0da6ac33a05810991b31aecc303e7add70933be\"" May 27 03:15:50.417132 containerd[1545]: time="2025-05-27T03:15:50.416374544Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.0\"" May 27 03:15:50.446717 containerd[1545]: time="2025-05-27T03:15:50.446667689Z" level=info msg="CreateContainer within sandbox \"ee55f0a62775158f1c34d1b6d2ee6627522c319b0bff44d0eb27e9f096f850af\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" May 27 03:15:50.477628 containerd[1545]: time="2025-05-27T03:15:50.477577605Z" level=info msg="Container c231d273fec318bbd731a160835cfa28edc40b71538fa4fb00cfcb82a02b74cd: CDI devices from CRI Config.CDIDevices: []" May 27 03:15:50.487886 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount335538099.mount: Deactivated successfully. May 27 03:15:50.491232 containerd[1545]: time="2025-05-27T03:15:50.491186722Z" level=info msg="CreateContainer within sandbox \"ee55f0a62775158f1c34d1b6d2ee6627522c319b0bff44d0eb27e9f096f850af\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"c231d273fec318bbd731a160835cfa28edc40b71538fa4fb00cfcb82a02b74cd\"" May 27 03:15:50.492865 containerd[1545]: time="2025-05-27T03:15:50.492758396Z" level=info msg="StartContainer for \"c231d273fec318bbd731a160835cfa28edc40b71538fa4fb00cfcb82a02b74cd\"" May 27 03:15:50.495996 containerd[1545]: time="2025-05-27T03:15:50.495854791Z" level=info msg="connecting to shim c231d273fec318bbd731a160835cfa28edc40b71538fa4fb00cfcb82a02b74cd" address="unix:///run/containerd/s/94d8c9516c56870c0bd771902b40e061d724c086f1c08e28a731e1b173a7d6b3" protocol=ttrpc version=3 May 27 03:15:50.544791 systemd[1]: Started cri-containerd-c231d273fec318bbd731a160835cfa28edc40b71538fa4fb00cfcb82a02b74cd.scope - libcontainer container c231d273fec318bbd731a160835cfa28edc40b71538fa4fb00cfcb82a02b74cd. May 27 03:15:50.617746 containerd[1545]: time="2025-05-27T03:15:50.617179747Z" level=info msg="StartContainer for \"c231d273fec318bbd731a160835cfa28edc40b71538fa4fb00cfcb82a02b74cd\" returns successfully" May 27 03:15:50.681161 kubelet[2688]: E0527 03:15:50.680541 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:50.689992 kubelet[2688]: E0527 03:15:50.689831 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.689992 kubelet[2688]: W0527 03:15:50.689866 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.690687 kubelet[2688]: E0527 03:15:50.690349 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.691040 kubelet[2688]: E0527 03:15:50.690984 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.691040 kubelet[2688]: W0527 03:15:50.691006 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.691778 kubelet[2688]: E0527 03:15:50.691649 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.692139 kubelet[2688]: E0527 03:15:50.692049 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.692139 kubelet[2688]: W0527 03:15:50.692081 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.692139 kubelet[2688]: E0527 03:15:50.692099 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.692607 kubelet[2688]: E0527 03:15:50.692592 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.692765 kubelet[2688]: W0527 03:15:50.692697 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.692765 kubelet[2688]: E0527 03:15:50.692717 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.693089 kubelet[2688]: E0527 03:15:50.693014 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.693089 kubelet[2688]: W0527 03:15:50.693026 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.693274 kubelet[2688]: E0527 03:15:50.693209 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.693509 kubelet[2688]: E0527 03:15:50.693477 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.693509 kubelet[2688]: W0527 03:15:50.693489 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.693685 kubelet[2688]: E0527 03:15:50.693614 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.693894 kubelet[2688]: E0527 03:15:50.693864 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.693894 kubelet[2688]: W0527 03:15:50.693876 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.694062 kubelet[2688]: E0527 03:15:50.693982 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.694393 kubelet[2688]: E0527 03:15:50.694340 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.694393 kubelet[2688]: W0527 03:15:50.694354 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.694393 kubelet[2688]: E0527 03:15:50.694366 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.694788 kubelet[2688]: E0527 03:15:50.694698 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.694788 kubelet[2688]: W0527 03:15:50.694709 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.694788 kubelet[2688]: E0527 03:15:50.694720 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.694975 kubelet[2688]: E0527 03:15:50.694965 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.695144 kubelet[2688]: W0527 03:15:50.695019 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.695144 kubelet[2688]: E0527 03:15:50.695033 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.695340 kubelet[2688]: E0527 03:15:50.695329 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.695481 kubelet[2688]: W0527 03:15:50.695388 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.695481 kubelet[2688]: E0527 03:15:50.695402 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.695654 kubelet[2688]: E0527 03:15:50.695643 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.695709 kubelet[2688]: W0527 03:15:50.695701 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.695763 kubelet[2688]: E0527 03:15:50.695751 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.696061 kubelet[2688]: E0527 03:15:50.695990 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.696061 kubelet[2688]: W0527 03:15:50.696001 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.696061 kubelet[2688]: E0527 03:15:50.696011 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.696485 kubelet[2688]: E0527 03:15:50.696376 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.696485 kubelet[2688]: W0527 03:15:50.696387 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.696485 kubelet[2688]: E0527 03:15:50.696398 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.696815 kubelet[2688]: E0527 03:15:50.696770 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.696989 kubelet[2688]: W0527 03:15:50.696902 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.696989 kubelet[2688]: E0527 03:15:50.696921 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.781369 kubelet[2688]: E0527 03:15:50.781328 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.781882 kubelet[2688]: W0527 03:15:50.781557 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.781882 kubelet[2688]: E0527 03:15:50.781592 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.782318 kubelet[2688]: E0527 03:15:50.782279 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.783166 kubelet[2688]: W0527 03:15:50.782301 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.783436 kubelet[2688]: E0527 03:15:50.783149 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.783931 kubelet[2688]: E0527 03:15:50.783792 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.783931 kubelet[2688]: W0527 03:15:50.783813 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.783931 kubelet[2688]: E0527 03:15:50.783853 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.784224 kubelet[2688]: E0527 03:15:50.784180 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.784224 kubelet[2688]: W0527 03:15:50.784199 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.784469 kubelet[2688]: E0527 03:15:50.784428 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.785284 kubelet[2688]: E0527 03:15:50.785220 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.786193 kubelet[2688]: W0527 03:15:50.786117 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.786376 kubelet[2688]: E0527 03:15:50.786307 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.789309 kubelet[2688]: E0527 03:15:50.789280 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.789309 kubelet[2688]: W0527 03:15:50.789303 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.789791 kubelet[2688]: E0527 03:15:50.789337 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.789791 kubelet[2688]: E0527 03:15:50.789674 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.789791 kubelet[2688]: W0527 03:15:50.789688 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.790419 kubelet[2688]: E0527 03:15:50.790291 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.790419 kubelet[2688]: W0527 03:15:50.790304 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.790419 kubelet[2688]: E0527 03:15:50.790352 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.790419 kubelet[2688]: E0527 03:15:50.790381 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.790665 kubelet[2688]: E0527 03:15:50.790486 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.790665 kubelet[2688]: W0527 03:15:50.790495 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.790665 kubelet[2688]: E0527 03:15:50.790519 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.792281 kubelet[2688]: E0527 03:15:50.792257 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.792281 kubelet[2688]: W0527 03:15:50.792279 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.792559 kubelet[2688]: E0527 03:15:50.792305 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.793369 kubelet[2688]: E0527 03:15:50.793335 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.793369 kubelet[2688]: W0527 03:15:50.793365 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.793610 kubelet[2688]: E0527 03:15:50.793471 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.794114 kubelet[2688]: E0527 03:15:50.794089 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.794114 kubelet[2688]: W0527 03:15:50.794110 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.794358 kubelet[2688]: E0527 03:15:50.794187 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.795178 kubelet[2688]: E0527 03:15:50.795158 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.795178 kubelet[2688]: W0527 03:15:50.795174 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.795750 kubelet[2688]: E0527 03:15:50.795691 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.796233 kubelet[2688]: E0527 03:15:50.796215 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.796233 kubelet[2688]: W0527 03:15:50.796229 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.797292 kubelet[2688]: E0527 03:15:50.797267 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.798437 kubelet[2688]: E0527 03:15:50.798413 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.798437 kubelet[2688]: W0527 03:15:50.798431 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.798754 kubelet[2688]: E0527 03:15:50.798538 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.798754 kubelet[2688]: E0527 03:15:50.798706 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.798754 kubelet[2688]: W0527 03:15:50.798716 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.798754 kubelet[2688]: E0527 03:15:50.798729 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.799675 kubelet[2688]: E0527 03:15:50.799212 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.799675 kubelet[2688]: W0527 03:15:50.799228 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.799675 kubelet[2688]: E0527 03:15:50.799241 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:50.800164 kubelet[2688]: E0527 03:15:50.800146 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:50.800257 kubelet[2688]: W0527 03:15:50.800245 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:50.800324 kubelet[2688]: E0527 03:15:50.800313 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.503035 kubelet[2688]: E0527 03:15:51.502964 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-drsns" podUID="851b262b-7562-4011-b67e-5e2de773f0d3" May 27 03:15:51.682286 kubelet[2688]: I0527 03:15:51.682251 2688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 27 03:15:51.682816 kubelet[2688]: E0527 03:15:51.682796 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:51.704607 kubelet[2688]: E0527 03:15:51.704559 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.704607 kubelet[2688]: W0527 03:15:51.704590 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.704848 kubelet[2688]: E0527 03:15:51.704631 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.704996 kubelet[2688]: E0527 03:15:51.704977 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.705134 kubelet[2688]: W0527 03:15:51.704994 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.705134 kubelet[2688]: E0527 03:15:51.705011 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.705500 kubelet[2688]: E0527 03:15:51.705481 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.705500 kubelet[2688]: W0527 03:15:51.705497 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.705601 kubelet[2688]: E0527 03:15:51.705511 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.705779 kubelet[2688]: E0527 03:15:51.705763 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.705818 kubelet[2688]: W0527 03:15:51.705781 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.705818 kubelet[2688]: E0527 03:15:51.705796 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.706081 kubelet[2688]: E0527 03:15:51.706045 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.706081 kubelet[2688]: W0527 03:15:51.706080 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.706189 kubelet[2688]: E0527 03:15:51.706095 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.706328 kubelet[2688]: E0527 03:15:51.706315 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.706328 kubelet[2688]: W0527 03:15:51.706326 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.706435 kubelet[2688]: E0527 03:15:51.706336 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.706631 kubelet[2688]: E0527 03:15:51.706611 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.706631 kubelet[2688]: W0527 03:15:51.706628 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.706729 kubelet[2688]: E0527 03:15:51.706640 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.706812 kubelet[2688]: E0527 03:15:51.706800 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.706812 kubelet[2688]: W0527 03:15:51.706810 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.706885 kubelet[2688]: E0527 03:15:51.706817 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.707087 kubelet[2688]: E0527 03:15:51.707071 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.707087 kubelet[2688]: W0527 03:15:51.707084 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.707202 kubelet[2688]: E0527 03:15:51.707095 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.707265 kubelet[2688]: E0527 03:15:51.707251 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.707265 kubelet[2688]: W0527 03:15:51.707261 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.707367 kubelet[2688]: E0527 03:15:51.707270 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.707562 kubelet[2688]: E0527 03:15:51.707544 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.707562 kubelet[2688]: W0527 03:15:51.707557 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.707680 kubelet[2688]: E0527 03:15:51.707567 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.707765 kubelet[2688]: E0527 03:15:51.707753 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.707765 kubelet[2688]: W0527 03:15:51.707765 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.707845 kubelet[2688]: E0527 03:15:51.707774 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.707931 kubelet[2688]: E0527 03:15:51.707918 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.707931 kubelet[2688]: W0527 03:15:51.707927 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.708031 kubelet[2688]: E0527 03:15:51.707935 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.708117 kubelet[2688]: E0527 03:15:51.708102 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.708117 kubelet[2688]: W0527 03:15:51.708109 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.708117 kubelet[2688]: E0527 03:15:51.708116 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.708330 kubelet[2688]: E0527 03:15:51.708316 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.708330 kubelet[2688]: W0527 03:15:51.708327 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.708426 kubelet[2688]: E0527 03:15:51.708335 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.793454 kubelet[2688]: E0527 03:15:51.793177 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.793454 kubelet[2688]: W0527 03:15:51.793206 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.793454 kubelet[2688]: E0527 03:15:51.793231 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.794620 kubelet[2688]: E0527 03:15:51.794137 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.794620 kubelet[2688]: W0527 03:15:51.794154 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.794620 kubelet[2688]: E0527 03:15:51.794181 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.794620 kubelet[2688]: E0527 03:15:51.794493 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.794620 kubelet[2688]: W0527 03:15:51.794508 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.794620 kubelet[2688]: E0527 03:15:51.794532 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.795364 kubelet[2688]: E0527 03:15:51.795337 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.795364 kubelet[2688]: W0527 03:15:51.795358 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.795520 kubelet[2688]: E0527 03:15:51.795379 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.796790 kubelet[2688]: E0527 03:15:51.796677 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.796790 kubelet[2688]: W0527 03:15:51.796696 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.796790 kubelet[2688]: E0527 03:15:51.796735 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.796950 kubelet[2688]: E0527 03:15:51.796868 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.796950 kubelet[2688]: W0527 03:15:51.796875 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.797021 kubelet[2688]: E0527 03:15:51.796995 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.797417 kubelet[2688]: E0527 03:15:51.797154 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.797417 kubelet[2688]: W0527 03:15:51.797176 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.797417 kubelet[2688]: E0527 03:15:51.797270 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.797591 kubelet[2688]: E0527 03:15:51.797468 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.797591 kubelet[2688]: W0527 03:15:51.797477 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.797591 kubelet[2688]: E0527 03:15:51.797499 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.797822 kubelet[2688]: E0527 03:15:51.797799 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.797822 kubelet[2688]: W0527 03:15:51.797809 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.797917 kubelet[2688]: E0527 03:15:51.797827 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.798276 kubelet[2688]: E0527 03:15:51.798261 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.798276 kubelet[2688]: W0527 03:15:51.798276 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.798477 kubelet[2688]: E0527 03:15:51.798290 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.798477 kubelet[2688]: E0527 03:15:51.798465 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.798477 kubelet[2688]: W0527 03:15:51.798474 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.798689 kubelet[2688]: E0527 03:15:51.798501 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.798689 kubelet[2688]: E0527 03:15:51.798621 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.798689 kubelet[2688]: W0527 03:15:51.798628 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.798783 kubelet[2688]: E0527 03:15:51.798708 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.798905 kubelet[2688]: E0527 03:15:51.798837 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.798905 kubelet[2688]: W0527 03:15:51.798848 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.798905 kubelet[2688]: E0527 03:15:51.798860 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.799444 kubelet[2688]: E0527 03:15:51.799090 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.799444 kubelet[2688]: W0527 03:15:51.799099 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.799444 kubelet[2688]: E0527 03:15:51.799115 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.799792 kubelet[2688]: E0527 03:15:51.799635 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.799792 kubelet[2688]: W0527 03:15:51.799650 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.799792 kubelet[2688]: E0527 03:15:51.799671 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.800205 kubelet[2688]: E0527 03:15:51.800187 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.800365 kubelet[2688]: W0527 03:15:51.800343 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.800934 kubelet[2688]: E0527 03:15:51.800440 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.800934 kubelet[2688]: E0527 03:15:51.800667 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.800934 kubelet[2688]: W0527 03:15:51.800679 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.800934 kubelet[2688]: E0527 03:15:51.800697 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.802323 kubelet[2688]: E0527 03:15:51.801945 2688 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 27 03:15:51.802323 kubelet[2688]: W0527 03:15:51.801965 2688 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 27 03:15:51.802323 kubelet[2688]: E0527 03:15:51.801985 2688 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 27 03:15:51.870009 containerd[1545]: time="2025-05-27T03:15:51.869947132Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:51.871210 containerd[1545]: time="2025-05-27T03:15:51.871167654Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.0: active requests=0, bytes read=4441619" May 27 03:15:51.873400 containerd[1545]: time="2025-05-27T03:15:51.873349096Z" level=info msg="ImageCreate event name:\"sha256:c53606cea03e59dcbfa981dc43a55dff05952895f72576b8389fa00be09ab676\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:51.876353 containerd[1545]: time="2025-05-27T03:15:51.876256418Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:ce76dd87f11d3fd0054c35ad2e0e9f833748d007f77a9bfe859d0ddcb66fcb2c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:51.877098 containerd[1545]: time="2025-05-27T03:15:51.876702537Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.0\" with image id \"sha256:c53606cea03e59dcbfa981dc43a55dff05952895f72576b8389fa00be09ab676\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.0\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:ce76dd87f11d3fd0054c35ad2e0e9f833748d007f77a9bfe859d0ddcb66fcb2c\", size \"5934282\" in 1.459272447s" May 27 03:15:51.877098 containerd[1545]: time="2025-05-27T03:15:51.876743412Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.0\" returns image reference \"sha256:c53606cea03e59dcbfa981dc43a55dff05952895f72576b8389fa00be09ab676\"" May 27 03:15:51.882034 containerd[1545]: time="2025-05-27T03:15:51.881953857Z" level=info msg="CreateContainer within sandbox \"6c251db89611d929b5891e7f51c221935e5f85890e558310106fd72966d833af\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" May 27 03:15:51.896086 containerd[1545]: time="2025-05-27T03:15:51.894499205Z" level=info msg="Container fb2cd4b4c76cd6163dd2b3cf90619f2391317f6fa89ddeb42b91f2dba75e8343: CDI devices from CRI Config.CDIDevices: []" May 27 03:15:51.904020 containerd[1545]: time="2025-05-27T03:15:51.903957815Z" level=info msg="CreateContainer within sandbox \"6c251db89611d929b5891e7f51c221935e5f85890e558310106fd72966d833af\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"fb2cd4b4c76cd6163dd2b3cf90619f2391317f6fa89ddeb42b91f2dba75e8343\"" May 27 03:15:51.905256 containerd[1545]: time="2025-05-27T03:15:51.905214767Z" level=info msg="StartContainer for \"fb2cd4b4c76cd6163dd2b3cf90619f2391317f6fa89ddeb42b91f2dba75e8343\"" May 27 03:15:51.908214 containerd[1545]: time="2025-05-27T03:15:51.908043535Z" level=info msg="connecting to shim fb2cd4b4c76cd6163dd2b3cf90619f2391317f6fa89ddeb42b91f2dba75e8343" address="unix:///run/containerd/s/604f30d5500ea97da73323276f3fde0f21d6cc0527a8e9361235c40e7e69416a" protocol=ttrpc version=3 May 27 03:15:51.942555 systemd[1]: Started cri-containerd-fb2cd4b4c76cd6163dd2b3cf90619f2391317f6fa89ddeb42b91f2dba75e8343.scope - libcontainer container fb2cd4b4c76cd6163dd2b3cf90619f2391317f6fa89ddeb42b91f2dba75e8343. May 27 03:15:52.034464 containerd[1545]: time="2025-05-27T03:15:52.034348931Z" level=info msg="StartContainer for \"fb2cd4b4c76cd6163dd2b3cf90619f2391317f6fa89ddeb42b91f2dba75e8343\" returns successfully" May 27 03:15:52.048269 systemd[1]: cri-containerd-fb2cd4b4c76cd6163dd2b3cf90619f2391317f6fa89ddeb42b91f2dba75e8343.scope: Deactivated successfully. May 27 03:15:52.055866 containerd[1545]: time="2025-05-27T03:15:52.055801516Z" level=info msg="TaskExit event in podsandbox handler container_id:\"fb2cd4b4c76cd6163dd2b3cf90619f2391317f6fa89ddeb42b91f2dba75e8343\" id:\"fb2cd4b4c76cd6163dd2b3cf90619f2391317f6fa89ddeb42b91f2dba75e8343\" pid:3453 exited_at:{seconds:1748315752 nanos:54266983}" May 27 03:15:52.055866 containerd[1545]: time="2025-05-27T03:15:52.055862550Z" level=info msg="received exit event container_id:\"fb2cd4b4c76cd6163dd2b3cf90619f2391317f6fa89ddeb42b91f2dba75e8343\" id:\"fb2cd4b4c76cd6163dd2b3cf90619f2391317f6fa89ddeb42b91f2dba75e8343\" pid:3453 exited_at:{seconds:1748315752 nanos:54266983}" May 27 03:15:52.107833 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-fb2cd4b4c76cd6163dd2b3cf90619f2391317f6fa89ddeb42b91f2dba75e8343-rootfs.mount: Deactivated successfully. May 27 03:15:52.689667 containerd[1545]: time="2025-05-27T03:15:52.689623809Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.0\"" May 27 03:15:52.715399 kubelet[2688]: I0527 03:15:52.714610 2688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-7d9f759884-9q2bq" podStartSLOduration=3.3710212 podStartE2EDuration="5.714583622s" podCreationTimestamp="2025-05-27 03:15:47 +0000 UTC" firstStartedPulling="2025-05-27 03:15:48.072516071 +0000 UTC m=+21.774128380" lastFinishedPulling="2025-05-27 03:15:50.416078443 +0000 UTC m=+24.117690802" observedRunningTime="2025-05-27 03:15:50.710296283 +0000 UTC m=+24.411908603" watchObservedRunningTime="2025-05-27 03:15:52.714583622 +0000 UTC m=+26.416195937" May 27 03:15:53.226003 kubelet[2688]: I0527 03:15:53.225153 2688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 27 03:15:53.226003 kubelet[2688]: E0527 03:15:53.225706 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:53.502704 kubelet[2688]: E0527 03:15:53.502540 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-drsns" podUID="851b262b-7562-4011-b67e-5e2de773f0d3" May 27 03:15:53.690535 kubelet[2688]: E0527 03:15:53.690081 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:55.503001 kubelet[2688]: E0527 03:15:55.502896 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-drsns" podUID="851b262b-7562-4011-b67e-5e2de773f0d3" May 27 03:15:55.793162 containerd[1545]: time="2025-05-27T03:15:55.792222210Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:55.793162 containerd[1545]: time="2025-05-27T03:15:55.792802235Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.0: active requests=0, bytes read=70300568" May 27 03:15:55.793797 containerd[1545]: time="2025-05-27T03:15:55.793201368Z" level=info msg="ImageCreate event name:\"sha256:15f996c472622f23047ea38b2d72940e8c34d0996b8a2e12a1f255c1d7083185\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:55.795927 containerd[1545]: time="2025-05-27T03:15:55.795832842Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:3dd06656abdc03fbd51782d5f6fe4d70e6825a1c0c5bce2a165bbd2ff9e0f7df\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:15:55.796754 containerd[1545]: time="2025-05-27T03:15:55.796722547Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.0\" with image id \"sha256:15f996c472622f23047ea38b2d72940e8c34d0996b8a2e12a1f255c1d7083185\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.0\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:3dd06656abdc03fbd51782d5f6fe4d70e6825a1c0c5bce2a165bbd2ff9e0f7df\", size \"71793271\" in 3.107054145s" May 27 03:15:55.796925 containerd[1545]: time="2025-05-27T03:15:55.796906838Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.0\" returns image reference \"sha256:15f996c472622f23047ea38b2d72940e8c34d0996b8a2e12a1f255c1d7083185\"" May 27 03:15:55.802046 containerd[1545]: time="2025-05-27T03:15:55.802004464Z" level=info msg="CreateContainer within sandbox \"6c251db89611d929b5891e7f51c221935e5f85890e558310106fd72966d833af\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" May 27 03:15:55.809216 containerd[1545]: time="2025-05-27T03:15:55.808761502Z" level=info msg="Container ef0cd5ed6ced4b7b7e3cdd547cad94bf57e36be23b7064167d88e1794c015ddb: CDI devices from CRI Config.CDIDevices: []" May 27 03:15:55.879898 containerd[1545]: time="2025-05-27T03:15:55.837850203Z" level=info msg="CreateContainer within sandbox \"6c251db89611d929b5891e7f51c221935e5f85890e558310106fd72966d833af\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"ef0cd5ed6ced4b7b7e3cdd547cad94bf57e36be23b7064167d88e1794c015ddb\"" May 27 03:15:55.882174 containerd[1545]: time="2025-05-27T03:15:55.882000218Z" level=info msg="StartContainer for \"ef0cd5ed6ced4b7b7e3cdd547cad94bf57e36be23b7064167d88e1794c015ddb\"" May 27 03:15:55.884900 containerd[1545]: time="2025-05-27T03:15:55.884854002Z" level=info msg="connecting to shim ef0cd5ed6ced4b7b7e3cdd547cad94bf57e36be23b7064167d88e1794c015ddb" address="unix:///run/containerd/s/604f30d5500ea97da73323276f3fde0f21d6cc0527a8e9361235c40e7e69416a" protocol=ttrpc version=3 May 27 03:15:55.918492 systemd[1]: Started cri-containerd-ef0cd5ed6ced4b7b7e3cdd547cad94bf57e36be23b7064167d88e1794c015ddb.scope - libcontainer container ef0cd5ed6ced4b7b7e3cdd547cad94bf57e36be23b7064167d88e1794c015ddb. May 27 03:15:55.986462 containerd[1545]: time="2025-05-27T03:15:55.986388756Z" level=info msg="StartContainer for \"ef0cd5ed6ced4b7b7e3cdd547cad94bf57e36be23b7064167d88e1794c015ddb\" returns successfully" May 27 03:15:56.595687 systemd[1]: cri-containerd-ef0cd5ed6ced4b7b7e3cdd547cad94bf57e36be23b7064167d88e1794c015ddb.scope: Deactivated successfully. May 27 03:15:56.596206 systemd[1]: cri-containerd-ef0cd5ed6ced4b7b7e3cdd547cad94bf57e36be23b7064167d88e1794c015ddb.scope: Consumed 640ms CPU time, 171.9M memory peak, 16M read from disk, 170.9M written to disk. May 27 03:15:56.601957 containerd[1545]: time="2025-05-27T03:15:56.601849163Z" level=info msg="received exit event container_id:\"ef0cd5ed6ced4b7b7e3cdd547cad94bf57e36be23b7064167d88e1794c015ddb\" id:\"ef0cd5ed6ced4b7b7e3cdd547cad94bf57e36be23b7064167d88e1794c015ddb\" pid:3515 exited_at:{seconds:1748315756 nanos:600889817}" May 27 03:15:56.604303 containerd[1545]: time="2025-05-27T03:15:56.604254022Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ef0cd5ed6ced4b7b7e3cdd547cad94bf57e36be23b7064167d88e1794c015ddb\" id:\"ef0cd5ed6ced4b7b7e3cdd547cad94bf57e36be23b7064167d88e1794c015ddb\" pid:3515 exited_at:{seconds:1748315756 nanos:600889817}" May 27 03:15:56.683533 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ef0cd5ed6ced4b7b7e3cdd547cad94bf57e36be23b7064167d88e1794c015ddb-rootfs.mount: Deactivated successfully. May 27 03:15:56.714110 kubelet[2688]: I0527 03:15:56.713658 2688 kubelet_node_status.go:488] "Fast updating node status as it just became ready" May 27 03:15:56.722451 containerd[1545]: time="2025-05-27T03:15:56.722396557Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.0\"" May 27 03:15:56.809485 systemd[1]: Created slice kubepods-burstable-pod9e7b1075_ea3d_4037_941f_f559a61e62a6.slice - libcontainer container kubepods-burstable-pod9e7b1075_ea3d_4037_941f_f559a61e62a6.slice. May 27 03:15:56.836927 systemd[1]: Created slice kubepods-besteffort-pod8c1deb33_d2c3_4f94_b199_70a79bc92bde.slice - libcontainer container kubepods-besteffort-pod8c1deb33_d2c3_4f94_b199_70a79bc92bde.slice. May 27 03:15:56.845419 systemd[1]: Created slice kubepods-burstable-pod5ef430cf_43c4_4f8a_96de_9c8721fea30a.slice - libcontainer container kubepods-burstable-pod5ef430cf_43c4_4f8a_96de_9c8721fea30a.slice. May 27 03:15:56.854781 systemd[1]: Created slice kubepods-besteffort-pod166fd789_e64d_4d0b_940a_5cec66cd52be.slice - libcontainer container kubepods-besteffort-pod166fd789_e64d_4d0b_940a_5cec66cd52be.slice. May 27 03:15:56.876525 systemd[1]: Created slice kubepods-besteffort-pod00c40ec8_4084_40ee_b388_e582df6866d9.slice - libcontainer container kubepods-besteffort-pod00c40ec8_4084_40ee_b388_e582df6866d9.slice. May 27 03:15:56.895523 systemd[1]: Created slice kubepods-besteffort-pod34798e8e_55b5_4e78_832b_798f5586248d.slice - libcontainer container kubepods-besteffort-pod34798e8e_55b5_4e78_832b_798f5586248d.slice. May 27 03:15:56.909557 systemd[1]: Created slice kubepods-besteffort-pode3d83951_7aca_4e07_9693_6aa4b63d6615.slice - libcontainer container kubepods-besteffort-pode3d83951_7aca_4e07_9693_6aa4b63d6615.slice. May 27 03:15:56.932988 kubelet[2688]: I0527 03:15:56.932922 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbcvj\" (UniqueName: \"kubernetes.io/projected/00c40ec8-4084-40ee-b388-e582df6866d9-kube-api-access-sbcvj\") pod \"calico-apiserver-77c54fb4c6-j9z9b\" (UID: \"00c40ec8-4084-40ee-b388-e582df6866d9\") " pod="calico-apiserver/calico-apiserver-77c54fb4c6-j9z9b" May 27 03:15:56.932988 kubelet[2688]: I0527 03:15:56.932988 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/166fd789-e64d-4d0b-940a-5cec66cd52be-calico-apiserver-certs\") pod \"calico-apiserver-77c54fb4c6-tgcsm\" (UID: \"166fd789-e64d-4d0b-940a-5cec66cd52be\") " pod="calico-apiserver/calico-apiserver-77c54fb4c6-tgcsm" May 27 03:15:56.933768 kubelet[2688]: I0527 03:15:56.933018 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c1deb33-d2c3-4f94-b199-70a79bc92bde-tigera-ca-bundle\") pod \"calico-kube-controllers-5d8fb586b5-2dbbs\" (UID: \"8c1deb33-d2c3-4f94-b199-70a79bc92bde\") " pod="calico-system/calico-kube-controllers-5d8fb586b5-2dbbs" May 27 03:15:56.933904 kubelet[2688]: I0527 03:15:56.933803 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk74v\" (UniqueName: \"kubernetes.io/projected/e3d83951-7aca-4e07-9693-6aa4b63d6615-kube-api-access-hk74v\") pod \"whisker-659d55984c-dwtcg\" (UID: \"e3d83951-7aca-4e07-9693-6aa4b63d6615\") " pod="calico-system/whisker-659d55984c-dwtcg" May 27 03:15:56.934005 kubelet[2688]: I0527 03:15:56.933980 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34798e8e-55b5-4e78-832b-798f5586248d-config\") pod \"goldmane-8f77d7b6c-zcnhc\" (UID: \"34798e8e-55b5-4e78-832b-798f5586248d\") " pod="calico-system/goldmane-8f77d7b6c-zcnhc" May 27 03:15:56.934275 kubelet[2688]: I0527 03:15:56.934080 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34798e8e-55b5-4e78-832b-798f5586248d-goldmane-ca-bundle\") pod \"goldmane-8f77d7b6c-zcnhc\" (UID: \"34798e8e-55b5-4e78-832b-798f5586248d\") " pod="calico-system/goldmane-8f77d7b6c-zcnhc" May 27 03:15:56.934275 kubelet[2688]: I0527 03:15:56.934125 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65z6c\" (UniqueName: \"kubernetes.io/projected/8c1deb33-d2c3-4f94-b199-70a79bc92bde-kube-api-access-65z6c\") pod \"calico-kube-controllers-5d8fb586b5-2dbbs\" (UID: \"8c1deb33-d2c3-4f94-b199-70a79bc92bde\") " pod="calico-system/calico-kube-controllers-5d8fb586b5-2dbbs" May 27 03:15:56.934275 kubelet[2688]: I0527 03:15:56.934153 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/00c40ec8-4084-40ee-b388-e582df6866d9-calico-apiserver-certs\") pod \"calico-apiserver-77c54fb4c6-j9z9b\" (UID: \"00c40ec8-4084-40ee-b388-e582df6866d9\") " pod="calico-apiserver/calico-apiserver-77c54fb4c6-j9z9b" May 27 03:15:56.934275 kubelet[2688]: I0527 03:15:56.934183 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rnl4\" (UniqueName: \"kubernetes.io/projected/5ef430cf-43c4-4f8a-96de-9c8721fea30a-kube-api-access-8rnl4\") pod \"coredns-7c65d6cfc9-bl2d5\" (UID: \"5ef430cf-43c4-4f8a-96de-9c8721fea30a\") " pod="kube-system/coredns-7c65d6cfc9-bl2d5" May 27 03:15:56.934275 kubelet[2688]: I0527 03:15:56.934214 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e7b1075-ea3d-4037-941f-f559a61e62a6-config-volume\") pod \"coredns-7c65d6cfc9-rqshl\" (UID: \"9e7b1075-ea3d-4037-941f-f559a61e62a6\") " pod="kube-system/coredns-7c65d6cfc9-rqshl" May 27 03:15:56.935586 kubelet[2688]: I0527 03:15:56.934370 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvj6r\" (UniqueName: \"kubernetes.io/projected/9e7b1075-ea3d-4037-941f-f559a61e62a6-kube-api-access-zvj6r\") pod \"coredns-7c65d6cfc9-rqshl\" (UID: \"9e7b1075-ea3d-4037-941f-f559a61e62a6\") " pod="kube-system/coredns-7c65d6cfc9-rqshl" May 27 03:15:56.935586 kubelet[2688]: I0527 03:15:56.934457 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ef430cf-43c4-4f8a-96de-9c8721fea30a-config-volume\") pod \"coredns-7c65d6cfc9-bl2d5\" (UID: \"5ef430cf-43c4-4f8a-96de-9c8721fea30a\") " pod="kube-system/coredns-7c65d6cfc9-bl2d5" May 27 03:15:56.935586 kubelet[2688]: I0527 03:15:56.934483 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqnph\" (UniqueName: \"kubernetes.io/projected/166fd789-e64d-4d0b-940a-5cec66cd52be-kube-api-access-jqnph\") pod \"calico-apiserver-77c54fb4c6-tgcsm\" (UID: \"166fd789-e64d-4d0b-940a-5cec66cd52be\") " pod="calico-apiserver/calico-apiserver-77c54fb4c6-tgcsm" May 27 03:15:56.935586 kubelet[2688]: I0527 03:15:56.934506 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e3d83951-7aca-4e07-9693-6aa4b63d6615-whisker-backend-key-pair\") pod \"whisker-659d55984c-dwtcg\" (UID: \"e3d83951-7aca-4e07-9693-6aa4b63d6615\") " pod="calico-system/whisker-659d55984c-dwtcg" May 27 03:15:56.935586 kubelet[2688]: I0527 03:15:56.934536 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3d83951-7aca-4e07-9693-6aa4b63d6615-whisker-ca-bundle\") pod \"whisker-659d55984c-dwtcg\" (UID: \"e3d83951-7aca-4e07-9693-6aa4b63d6615\") " pod="calico-system/whisker-659d55984c-dwtcg" May 27 03:15:56.937047 kubelet[2688]: I0527 03:15:56.934568 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/34798e8e-55b5-4e78-832b-798f5586248d-goldmane-key-pair\") pod \"goldmane-8f77d7b6c-zcnhc\" (UID: \"34798e8e-55b5-4e78-832b-798f5586248d\") " pod="calico-system/goldmane-8f77d7b6c-zcnhc" May 27 03:15:56.937047 kubelet[2688]: I0527 03:15:56.934596 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7gvq\" (UniqueName: \"kubernetes.io/projected/34798e8e-55b5-4e78-832b-798f5586248d-kube-api-access-t7gvq\") pod \"goldmane-8f77d7b6c-zcnhc\" (UID: \"34798e8e-55b5-4e78-832b-798f5586248d\") " pod="calico-system/goldmane-8f77d7b6c-zcnhc" May 27 03:15:57.158193 kubelet[2688]: E0527 03:15:57.153566 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:57.162540 containerd[1545]: time="2025-05-27T03:15:57.161803421Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-bl2d5,Uid:5ef430cf-43c4-4f8a-96de-9c8721fea30a,Namespace:kube-system,Attempt:0,}" May 27 03:15:57.179088 containerd[1545]: time="2025-05-27T03:15:57.176901250Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-77c54fb4c6-tgcsm,Uid:166fd789-e64d-4d0b-940a-5cec66cd52be,Namespace:calico-apiserver,Attempt:0,}" May 27 03:15:57.205776 containerd[1545]: time="2025-05-27T03:15:57.202519148Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-77c54fb4c6-j9z9b,Uid:00c40ec8-4084-40ee-b388-e582df6866d9,Namespace:calico-apiserver,Attempt:0,}" May 27 03:15:57.216952 containerd[1545]: time="2025-05-27T03:15:57.216892934Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-659d55984c-dwtcg,Uid:e3d83951-7aca-4e07-9693-6aa4b63d6615,Namespace:calico-system,Attempt:0,}" May 27 03:15:57.218240 containerd[1545]: time="2025-05-27T03:15:57.217894838Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-8f77d7b6c-zcnhc,Uid:34798e8e-55b5-4e78-832b-798f5586248d,Namespace:calico-system,Attempt:0,}" May 27 03:15:57.445439 containerd[1545]: time="2025-05-27T03:15:57.445375628Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5d8fb586b5-2dbbs,Uid:8c1deb33-d2c3-4f94-b199-70a79bc92bde,Namespace:calico-system,Attempt:0,}" May 27 03:15:57.447382 kubelet[2688]: E0527 03:15:57.447128 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:15:57.453015 containerd[1545]: time="2025-05-27T03:15:57.452953788Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-rqshl,Uid:9e7b1075-ea3d-4037-941f-f559a61e62a6,Namespace:kube-system,Attempt:0,}" May 27 03:15:57.518426 systemd[1]: Created slice kubepods-besteffort-pod851b262b_7562_4011_b67e_5e2de773f0d3.slice - libcontainer container kubepods-besteffort-pod851b262b_7562_4011_b67e_5e2de773f0d3.slice. May 27 03:15:57.525799 containerd[1545]: time="2025-05-27T03:15:57.525739443Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drsns,Uid:851b262b-7562-4011-b67e-5e2de773f0d3,Namespace:calico-system,Attempt:0,}" May 27 03:15:57.624600 containerd[1545]: time="2025-05-27T03:15:57.623408193Z" level=error msg="Failed to destroy network for sandbox \"e3c24da144829af172df354fb50a85d4a90dd079d032d1c4c1d9bdd6b9633692\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.631082 containerd[1545]: time="2025-05-27T03:15:57.630673523Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-77c54fb4c6-tgcsm,Uid:166fd789-e64d-4d0b-940a-5cec66cd52be,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3c24da144829af172df354fb50a85d4a90dd079d032d1c4c1d9bdd6b9633692\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.631969 kubelet[2688]: E0527 03:15:57.631902 2688 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3c24da144829af172df354fb50a85d4a90dd079d032d1c4c1d9bdd6b9633692\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.632311 kubelet[2688]: E0527 03:15:57.632006 2688 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3c24da144829af172df354fb50a85d4a90dd079d032d1c4c1d9bdd6b9633692\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-77c54fb4c6-tgcsm" May 27 03:15:57.632311 kubelet[2688]: E0527 03:15:57.632038 2688 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e3c24da144829af172df354fb50a85d4a90dd079d032d1c4c1d9bdd6b9633692\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-77c54fb4c6-tgcsm" May 27 03:15:57.634747 kubelet[2688]: E0527 03:15:57.632210 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-77c54fb4c6-tgcsm_calico-apiserver(166fd789-e64d-4d0b-940a-5cec66cd52be)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-77c54fb4c6-tgcsm_calico-apiserver(166fd789-e64d-4d0b-940a-5cec66cd52be)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e3c24da144829af172df354fb50a85d4a90dd079d032d1c4c1d9bdd6b9633692\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-77c54fb4c6-tgcsm" podUID="166fd789-e64d-4d0b-940a-5cec66cd52be" May 27 03:15:57.682978 containerd[1545]: time="2025-05-27T03:15:57.682776245Z" level=error msg="Failed to destroy network for sandbox \"7990c7de15c88e41ac0d957776d28613fae339a7d70ede307501f38fc7b71d2f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.686865 containerd[1545]: time="2025-05-27T03:15:57.686042844Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-77c54fb4c6-j9z9b,Uid:00c40ec8-4084-40ee-b388-e582df6866d9,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"7990c7de15c88e41ac0d957776d28613fae339a7d70ede307501f38fc7b71d2f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.687526 kubelet[2688]: E0527 03:15:57.687473 2688 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7990c7de15c88e41ac0d957776d28613fae339a7d70ede307501f38fc7b71d2f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.687659 kubelet[2688]: E0527 03:15:57.687560 2688 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7990c7de15c88e41ac0d957776d28613fae339a7d70ede307501f38fc7b71d2f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-77c54fb4c6-j9z9b" May 27 03:15:57.687659 kubelet[2688]: E0527 03:15:57.687595 2688 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7990c7de15c88e41ac0d957776d28613fae339a7d70ede307501f38fc7b71d2f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-77c54fb4c6-j9z9b" May 27 03:15:57.687780 kubelet[2688]: E0527 03:15:57.687653 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-77c54fb4c6-j9z9b_calico-apiserver(00c40ec8-4084-40ee-b388-e582df6866d9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-77c54fb4c6-j9z9b_calico-apiserver(00c40ec8-4084-40ee-b388-e582df6866d9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7990c7de15c88e41ac0d957776d28613fae339a7d70ede307501f38fc7b71d2f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-77c54fb4c6-j9z9b" podUID="00c40ec8-4084-40ee-b388-e582df6866d9" May 27 03:15:57.747917 containerd[1545]: time="2025-05-27T03:15:57.747319558Z" level=error msg="Failed to destroy network for sandbox \"a5e056443f9b3877f96c677478fbe8d0cae76d6c093d14375e5ffa98ff4a15f3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.758092 containerd[1545]: time="2025-05-27T03:15:57.757941583Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-659d55984c-dwtcg,Uid:e3d83951-7aca-4e07-9693-6aa4b63d6615,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"a5e056443f9b3877f96c677478fbe8d0cae76d6c093d14375e5ffa98ff4a15f3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.758931 kubelet[2688]: E0527 03:15:57.758872 2688 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a5e056443f9b3877f96c677478fbe8d0cae76d6c093d14375e5ffa98ff4a15f3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.760686 kubelet[2688]: E0527 03:15:57.759962 2688 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a5e056443f9b3877f96c677478fbe8d0cae76d6c093d14375e5ffa98ff4a15f3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-659d55984c-dwtcg" May 27 03:15:57.760686 kubelet[2688]: E0527 03:15:57.760032 2688 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a5e056443f9b3877f96c677478fbe8d0cae76d6c093d14375e5ffa98ff4a15f3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-659d55984c-dwtcg" May 27 03:15:57.760686 kubelet[2688]: E0527 03:15:57.760149 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-659d55984c-dwtcg_calico-system(e3d83951-7aca-4e07-9693-6aa4b63d6615)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-659d55984c-dwtcg_calico-system(e3d83951-7aca-4e07-9693-6aa4b63d6615)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a5e056443f9b3877f96c677478fbe8d0cae76d6c093d14375e5ffa98ff4a15f3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-659d55984c-dwtcg" podUID="e3d83951-7aca-4e07-9693-6aa4b63d6615" May 27 03:15:57.772664 containerd[1545]: time="2025-05-27T03:15:57.772451404Z" level=error msg="Failed to destroy network for sandbox \"7e30958b489a41721af456848c12e9f42dbbda4bd6d3522bbb09c8a8c7cc9063\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.774497 containerd[1545]: time="2025-05-27T03:15:57.774433069Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-bl2d5,Uid:5ef430cf-43c4-4f8a-96de-9c8721fea30a,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"7e30958b489a41721af456848c12e9f42dbbda4bd6d3522bbb09c8a8c7cc9063\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.774983 kubelet[2688]: E0527 03:15:57.774763 2688 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7e30958b489a41721af456848c12e9f42dbbda4bd6d3522bbb09c8a8c7cc9063\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.774983 kubelet[2688]: E0527 03:15:57.774857 2688 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7e30958b489a41721af456848c12e9f42dbbda4bd6d3522bbb09c8a8c7cc9063\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-bl2d5" May 27 03:15:57.774983 kubelet[2688]: E0527 03:15:57.774893 2688 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7e30958b489a41721af456848c12e9f42dbbda4bd6d3522bbb09c8a8c7cc9063\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-bl2d5" May 27 03:15:57.777777 kubelet[2688]: E0527 03:15:57.775202 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-bl2d5_kube-system(5ef430cf-43c4-4f8a-96de-9c8721fea30a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-bl2d5_kube-system(5ef430cf-43c4-4f8a-96de-9c8721fea30a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7e30958b489a41721af456848c12e9f42dbbda4bd6d3522bbb09c8a8c7cc9063\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-bl2d5" podUID="5ef430cf-43c4-4f8a-96de-9c8721fea30a" May 27 03:15:57.813178 containerd[1545]: time="2025-05-27T03:15:57.812919062Z" level=error msg="Failed to destroy network for sandbox \"b91c25503d5de88211ff29ebc22ad5b2a81353e6af7b9f97e615b916f6e2e5f1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.817608 containerd[1545]: time="2025-05-27T03:15:57.817532886Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-8f77d7b6c-zcnhc,Uid:34798e8e-55b5-4e78-832b-798f5586248d,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"b91c25503d5de88211ff29ebc22ad5b2a81353e6af7b9f97e615b916f6e2e5f1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.819917 kubelet[2688]: E0527 03:15:57.818336 2688 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b91c25503d5de88211ff29ebc22ad5b2a81353e6af7b9f97e615b916f6e2e5f1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.819917 kubelet[2688]: E0527 03:15:57.818427 2688 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b91c25503d5de88211ff29ebc22ad5b2a81353e6af7b9f97e615b916f6e2e5f1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-8f77d7b6c-zcnhc" May 27 03:15:57.819917 kubelet[2688]: E0527 03:15:57.818457 2688 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b91c25503d5de88211ff29ebc22ad5b2a81353e6af7b9f97e615b916f6e2e5f1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-8f77d7b6c-zcnhc" May 27 03:15:57.820926 kubelet[2688]: E0527 03:15:57.818710 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-8f77d7b6c-zcnhc_calico-system(34798e8e-55b5-4e78-832b-798f5586248d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-8f77d7b6c-zcnhc_calico-system(34798e8e-55b5-4e78-832b-798f5586248d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b91c25503d5de88211ff29ebc22ad5b2a81353e6af7b9f97e615b916f6e2e5f1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-8f77d7b6c-zcnhc" podUID="34798e8e-55b5-4e78-832b-798f5586248d" May 27 03:15:57.821078 containerd[1545]: time="2025-05-27T03:15:57.820518739Z" level=error msg="Failed to destroy network for sandbox \"89940c4cd298c5184924bd647f6c032376fad3f830cddcccd21d9e35fc21cc68\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.821904 containerd[1545]: time="2025-05-27T03:15:57.821626830Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-rqshl,Uid:9e7b1075-ea3d-4037-941f-f559a61e62a6,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"89940c4cd298c5184924bd647f6c032376fad3f830cddcccd21d9e35fc21cc68\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.822420 kubelet[2688]: E0527 03:15:57.822369 2688 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"89940c4cd298c5184924bd647f6c032376fad3f830cddcccd21d9e35fc21cc68\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.822537 kubelet[2688]: E0527 03:15:57.822455 2688 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"89940c4cd298c5184924bd647f6c032376fad3f830cddcccd21d9e35fc21cc68\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-rqshl" May 27 03:15:57.822537 kubelet[2688]: E0527 03:15:57.822494 2688 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"89940c4cd298c5184924bd647f6c032376fad3f830cddcccd21d9e35fc21cc68\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-rqshl" May 27 03:15:57.822642 kubelet[2688]: E0527 03:15:57.822559 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-rqshl_kube-system(9e7b1075-ea3d-4037-941f-f559a61e62a6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-rqshl_kube-system(9e7b1075-ea3d-4037-941f-f559a61e62a6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"89940c4cd298c5184924bd647f6c032376fad3f830cddcccd21d9e35fc21cc68\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-rqshl" podUID="9e7b1075-ea3d-4037-941f-f559a61e62a6" May 27 03:15:57.836382 containerd[1545]: time="2025-05-27T03:15:57.836145948Z" level=error msg="Failed to destroy network for sandbox \"4bc37a02fc9a96b72d67e46fb86fda7bca3de4a64db3b675a58876761ba4285c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.838024 containerd[1545]: time="2025-05-27T03:15:57.837947740Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5d8fb586b5-2dbbs,Uid:8c1deb33-d2c3-4f94-b199-70a79bc92bde,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"4bc37a02fc9a96b72d67e46fb86fda7bca3de4a64db3b675a58876761ba4285c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.839087 kubelet[2688]: E0527 03:15:57.838979 2688 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4bc37a02fc9a96b72d67e46fb86fda7bca3de4a64db3b675a58876761ba4285c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.840644 kubelet[2688]: E0527 03:15:57.839552 2688 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4bc37a02fc9a96b72d67e46fb86fda7bca3de4a64db3b675a58876761ba4285c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5d8fb586b5-2dbbs" May 27 03:15:57.840644 kubelet[2688]: E0527 03:15:57.839709 2688 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4bc37a02fc9a96b72d67e46fb86fda7bca3de4a64db3b675a58876761ba4285c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5d8fb586b5-2dbbs" May 27 03:15:57.840644 kubelet[2688]: E0527 03:15:57.839948 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5d8fb586b5-2dbbs_calico-system(8c1deb33-d2c3-4f94-b199-70a79bc92bde)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5d8fb586b5-2dbbs_calico-system(8c1deb33-d2c3-4f94-b199-70a79bc92bde)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4bc37a02fc9a96b72d67e46fb86fda7bca3de4a64db3b675a58876761ba4285c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5d8fb586b5-2dbbs" podUID="8c1deb33-d2c3-4f94-b199-70a79bc92bde" May 27 03:15:57.859170 containerd[1545]: time="2025-05-27T03:15:57.859080963Z" level=error msg="Failed to destroy network for sandbox \"93055a2b5173d19fa70960059acc6a3adceb8b2bb302739cfab9b2494a0dc694\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.860629 containerd[1545]: time="2025-05-27T03:15:57.860565584Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drsns,Uid:851b262b-7562-4011-b67e-5e2de773f0d3,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"93055a2b5173d19fa70960059acc6a3adceb8b2bb302739cfab9b2494a0dc694\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.862411 kubelet[2688]: E0527 03:15:57.861680 2688 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"93055a2b5173d19fa70960059acc6a3adceb8b2bb302739cfab9b2494a0dc694\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 27 03:15:57.862411 kubelet[2688]: E0527 03:15:57.861804 2688 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"93055a2b5173d19fa70960059acc6a3adceb8b2bb302739cfab9b2494a0dc694\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-drsns" May 27 03:15:57.862411 kubelet[2688]: E0527 03:15:57.861877 2688 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"93055a2b5173d19fa70960059acc6a3adceb8b2bb302739cfab9b2494a0dc694\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-drsns" May 27 03:15:57.862662 kubelet[2688]: E0527 03:15:57.862118 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-drsns_calico-system(851b262b-7562-4011-b67e-5e2de773f0d3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-drsns_calico-system(851b262b-7562-4011-b67e-5e2de773f0d3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"93055a2b5173d19fa70960059acc6a3adceb8b2bb302739cfab9b2494a0dc694\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-drsns" podUID="851b262b-7562-4011-b67e-5e2de773f0d3" May 27 03:15:58.075952 systemd[1]: run-netns-cni\x2d209e73ab\x2d793a\x2de004\x2d5c3b\x2d3fe5c5595114.mount: Deactivated successfully. May 27 03:16:03.671414 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2718887672.mount: Deactivated successfully. May 27 03:16:03.711511 containerd[1545]: time="2025-05-27T03:16:03.711305113Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:03.713449 containerd[1545]: time="2025-05-27T03:16:03.713333092Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.0: active requests=0, bytes read=156396372" May 27 03:16:03.714030 containerd[1545]: time="2025-05-27T03:16:03.713741332Z" level=info msg="ImageCreate event name:\"sha256:d12dae9bc0999225efe30fd5618bcf2195709d54ed2840234f5006aab5f7d721\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:03.732249 containerd[1545]: time="2025-05-27T03:16:03.731910264Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:7cb61ea47ca0a8e6d0526a42da4f1e399b37ccd13339d0776d272465cb7ee012\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:03.732897 containerd[1545]: time="2025-05-27T03:16:03.732850861Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.0\" with image id \"sha256:d12dae9bc0999225efe30fd5618bcf2195709d54ed2840234f5006aab5f7d721\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.0\", repo digest \"ghcr.io/flatcar/calico/node@sha256:7cb61ea47ca0a8e6d0526a42da4f1e399b37ccd13339d0776d272465cb7ee012\", size \"156396234\" in 7.010401561s" May 27 03:16:03.732897 containerd[1545]: time="2025-05-27T03:16:03.732898768Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.0\" returns image reference \"sha256:d12dae9bc0999225efe30fd5618bcf2195709d54ed2840234f5006aab5f7d721\"" May 27 03:16:03.773518 containerd[1545]: time="2025-05-27T03:16:03.773440274Z" level=info msg="CreateContainer within sandbox \"6c251db89611d929b5891e7f51c221935e5f85890e558310106fd72966d833af\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" May 27 03:16:03.793404 containerd[1545]: time="2025-05-27T03:16:03.793333627Z" level=info msg="Container 320d5852593cb9fdb7fcd6a5a8f42623b3c597874b90e987eaa194b5384f5175: CDI devices from CRI Config.CDIDevices: []" May 27 03:16:03.812453 containerd[1545]: time="2025-05-27T03:16:03.812246340Z" level=info msg="CreateContainer within sandbox \"6c251db89611d929b5891e7f51c221935e5f85890e558310106fd72966d833af\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"320d5852593cb9fdb7fcd6a5a8f42623b3c597874b90e987eaa194b5384f5175\"" May 27 03:16:03.813392 containerd[1545]: time="2025-05-27T03:16:03.813345417Z" level=info msg="StartContainer for \"320d5852593cb9fdb7fcd6a5a8f42623b3c597874b90e987eaa194b5384f5175\"" May 27 03:16:03.815568 containerd[1545]: time="2025-05-27T03:16:03.815524652Z" level=info msg="connecting to shim 320d5852593cb9fdb7fcd6a5a8f42623b3c597874b90e987eaa194b5384f5175" address="unix:///run/containerd/s/604f30d5500ea97da73323276f3fde0f21d6cc0527a8e9361235c40e7e69416a" protocol=ttrpc version=3 May 27 03:16:04.031654 systemd[1]: Started cri-containerd-320d5852593cb9fdb7fcd6a5a8f42623b3c597874b90e987eaa194b5384f5175.scope - libcontainer container 320d5852593cb9fdb7fcd6a5a8f42623b3c597874b90e987eaa194b5384f5175. May 27 03:16:04.106018 containerd[1545]: time="2025-05-27T03:16:04.105973717Z" level=info msg="StartContainer for \"320d5852593cb9fdb7fcd6a5a8f42623b3c597874b90e987eaa194b5384f5175\" returns successfully" May 27 03:16:04.227531 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. May 27 03:16:04.229845 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. May 27 03:16:04.510369 kubelet[2688]: I0527 03:16:04.510318 2688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e3d83951-7aca-4e07-9693-6aa4b63d6615-whisker-backend-key-pair\") pod \"e3d83951-7aca-4e07-9693-6aa4b63d6615\" (UID: \"e3d83951-7aca-4e07-9693-6aa4b63d6615\") " May 27 03:16:04.510806 kubelet[2688]: I0527 03:16:04.510456 2688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk74v\" (UniqueName: \"kubernetes.io/projected/e3d83951-7aca-4e07-9693-6aa4b63d6615-kube-api-access-hk74v\") pod \"e3d83951-7aca-4e07-9693-6aa4b63d6615\" (UID: \"e3d83951-7aca-4e07-9693-6aa4b63d6615\") " May 27 03:16:04.510806 kubelet[2688]: I0527 03:16:04.510483 2688 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3d83951-7aca-4e07-9693-6aa4b63d6615-whisker-ca-bundle\") pod \"e3d83951-7aca-4e07-9693-6aa4b63d6615\" (UID: \"e3d83951-7aca-4e07-9693-6aa4b63d6615\") " May 27 03:16:04.512511 kubelet[2688]: I0527 03:16:04.512361 2688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3d83951-7aca-4e07-9693-6aa4b63d6615-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "e3d83951-7aca-4e07-9693-6aa4b63d6615" (UID: "e3d83951-7aca-4e07-9693-6aa4b63d6615"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 27 03:16:04.516910 kubelet[2688]: I0527 03:16:04.516757 2688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3d83951-7aca-4e07-9693-6aa4b63d6615-kube-api-access-hk74v" (OuterVolumeSpecName: "kube-api-access-hk74v") pod "e3d83951-7aca-4e07-9693-6aa4b63d6615" (UID: "e3d83951-7aca-4e07-9693-6aa4b63d6615"). InnerVolumeSpecName "kube-api-access-hk74v". PluginName "kubernetes.io/projected", VolumeGidValue "" May 27 03:16:04.517694 kubelet[2688]: I0527 03:16:04.517631 2688 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d83951-7aca-4e07-9693-6aa4b63d6615-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "e3d83951-7aca-4e07-9693-6aa4b63d6615" (UID: "e3d83951-7aca-4e07-9693-6aa4b63d6615"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGidValue "" May 27 03:16:04.611311 kubelet[2688]: I0527 03:16:04.611216 2688 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk74v\" (UniqueName: \"kubernetes.io/projected/e3d83951-7aca-4e07-9693-6aa4b63d6615-kube-api-access-hk74v\") on node \"ci-4344.0.0-4-f2dd3e9233\" DevicePath \"\"" May 27 03:16:04.611311 kubelet[2688]: I0527 03:16:04.611260 2688 reconciler_common.go:293] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3d83951-7aca-4e07-9693-6aa4b63d6615-whisker-ca-bundle\") on node \"ci-4344.0.0-4-f2dd3e9233\" DevicePath \"\"" May 27 03:16:04.611311 kubelet[2688]: I0527 03:16:04.611276 2688 reconciler_common.go:293] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e3d83951-7aca-4e07-9693-6aa4b63d6615-whisker-backend-key-pair\") on node \"ci-4344.0.0-4-f2dd3e9233\" DevicePath \"\"" May 27 03:16:04.673264 systemd[1]: var-lib-kubelet-pods-e3d83951\x2d7aca\x2d4e07\x2d9693\x2d6aa4b63d6615-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dhk74v.mount: Deactivated successfully. May 27 03:16:04.673493 systemd[1]: var-lib-kubelet-pods-e3d83951\x2d7aca\x2d4e07\x2d9693\x2d6aa4b63d6615-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. May 27 03:16:04.782212 systemd[1]: Removed slice kubepods-besteffort-pode3d83951_7aca_4e07_9693_6aa4b63d6615.slice - libcontainer container kubepods-besteffort-pode3d83951_7aca_4e07_9693_6aa4b63d6615.slice. May 27 03:16:04.797097 kubelet[2688]: I0527 03:16:04.796916 2688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-2z8x7" podStartSLOduration=2.336136233 podStartE2EDuration="17.796877286s" podCreationTimestamp="2025-05-27 03:15:47 +0000 UTC" firstStartedPulling="2025-05-27 03:15:48.273444623 +0000 UTC m=+21.975056939" lastFinishedPulling="2025-05-27 03:16:03.734185685 +0000 UTC m=+37.435797992" observedRunningTime="2025-05-27 03:16:04.794407897 +0000 UTC m=+38.496020215" watchObservedRunningTime="2025-05-27 03:16:04.796877286 +0000 UTC m=+38.498489885" May 27 03:16:04.873941 systemd[1]: Created slice kubepods-besteffort-pod148d61b5_bb87_4239_99d0_544a4ea406a1.slice - libcontainer container kubepods-besteffort-pod148d61b5_bb87_4239_99d0_544a4ea406a1.slice. May 27 03:16:05.015026 kubelet[2688]: I0527 03:16:05.014820 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs6dk\" (UniqueName: \"kubernetes.io/projected/148d61b5-bb87-4239-99d0-544a4ea406a1-kube-api-access-bs6dk\") pod \"whisker-6464b44b6c-gfpc2\" (UID: \"148d61b5-bb87-4239-99d0-544a4ea406a1\") " pod="calico-system/whisker-6464b44b6c-gfpc2" May 27 03:16:05.015026 kubelet[2688]: I0527 03:16:05.014911 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/148d61b5-bb87-4239-99d0-544a4ea406a1-whisker-backend-key-pair\") pod \"whisker-6464b44b6c-gfpc2\" (UID: \"148d61b5-bb87-4239-99d0-544a4ea406a1\") " pod="calico-system/whisker-6464b44b6c-gfpc2" May 27 03:16:05.015026 kubelet[2688]: I0527 03:16:05.014930 2688 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/148d61b5-bb87-4239-99d0-544a4ea406a1-whisker-ca-bundle\") pod \"whisker-6464b44b6c-gfpc2\" (UID: \"148d61b5-bb87-4239-99d0-544a4ea406a1\") " pod="calico-system/whisker-6464b44b6c-gfpc2" May 27 03:16:05.182426 containerd[1545]: time="2025-05-27T03:16:05.182345598Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6464b44b6c-gfpc2,Uid:148d61b5-bb87-4239-99d0-544a4ea406a1,Namespace:calico-system,Attempt:0,}" May 27 03:16:05.536750 systemd-networkd[1448]: cali61953bffda8: Link UP May 27 03:16:05.537166 systemd-networkd[1448]: cali61953bffda8: Gained carrier May 27 03:16:05.560983 containerd[1545]: 2025-05-27 03:16:05.238 [INFO][3845] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist May 27 03:16:05.560983 containerd[1545]: 2025-05-27 03:16:05.283 [INFO][3845] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4344.0.0--4--f2dd3e9233-k8s-whisker--6464b44b6c--gfpc2-eth0 whisker-6464b44b6c- calico-system 148d61b5-bb87-4239-99d0-544a4ea406a1 899 0 2025-05-27 03:16:04 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:6464b44b6c projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4344.0.0-4-f2dd3e9233 whisker-6464b44b6c-gfpc2 eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali61953bffda8 [] [] }} ContainerID="78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" Namespace="calico-system" Pod="whisker-6464b44b6c-gfpc2" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-whisker--6464b44b6c--gfpc2-" May 27 03:16:05.560983 containerd[1545]: 2025-05-27 03:16:05.283 [INFO][3845] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" Namespace="calico-system" Pod="whisker-6464b44b6c-gfpc2" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-whisker--6464b44b6c--gfpc2-eth0" May 27 03:16:05.560983 containerd[1545]: 2025-05-27 03:16:05.450 [INFO][3856] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" HandleID="k8s-pod-network.78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-whisker--6464b44b6c--gfpc2-eth0" May 27 03:16:05.561562 containerd[1545]: 2025-05-27 03:16:05.454 [INFO][3856] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" HandleID="k8s-pod-network.78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-whisker--6464b44b6c--gfpc2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003a3270), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4344.0.0-4-f2dd3e9233", "pod":"whisker-6464b44b6c-gfpc2", "timestamp":"2025-05-27 03:16:05.450793273 +0000 UTC"}, Hostname:"ci-4344.0.0-4-f2dd3e9233", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 27 03:16:05.561562 containerd[1545]: 2025-05-27 03:16:05.454 [INFO][3856] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 27 03:16:05.561562 containerd[1545]: 2025-05-27 03:16:05.454 [INFO][3856] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 27 03:16:05.561562 containerd[1545]: 2025-05-27 03:16:05.454 [INFO][3856] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4344.0.0-4-f2dd3e9233' May 27 03:16:05.561562 containerd[1545]: 2025-05-27 03:16:05.474 [INFO][3856] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:05.561562 containerd[1545]: 2025-05-27 03:16:05.486 [INFO][3856] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:05.561562 containerd[1545]: 2025-05-27 03:16:05.494 [INFO][3856] ipam/ipam.go 511: Trying affinity for 192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:05.561562 containerd[1545]: 2025-05-27 03:16:05.496 [INFO][3856] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:05.561562 containerd[1545]: 2025-05-27 03:16:05.500 [INFO][3856] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:05.563193 containerd[1545]: 2025-05-27 03:16:05.500 [INFO][3856] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.192/26 handle="k8s-pod-network.78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:05.563193 containerd[1545]: 2025-05-27 03:16:05.503 [INFO][3856] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d May 27 03:16:05.563193 containerd[1545]: 2025-05-27 03:16:05.509 [INFO][3856] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.192/26 handle="k8s-pod-network.78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:05.563193 containerd[1545]: 2025-05-27 03:16:05.519 [INFO][3856] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.193/26] block=192.168.1.192/26 handle="k8s-pod-network.78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:05.563193 containerd[1545]: 2025-05-27 03:16:05.520 [INFO][3856] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.193/26] handle="k8s-pod-network.78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:05.563193 containerd[1545]: 2025-05-27 03:16:05.520 [INFO][3856] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 27 03:16:05.563193 containerd[1545]: 2025-05-27 03:16:05.520 [INFO][3856] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.193/26] IPv6=[] ContainerID="78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" HandleID="k8s-pod-network.78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-whisker--6464b44b6c--gfpc2-eth0" May 27 03:16:05.563436 containerd[1545]: 2025-05-27 03:16:05.523 [INFO][3845] cni-plugin/k8s.go 418: Populated endpoint ContainerID="78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" Namespace="calico-system" Pod="whisker-6464b44b6c-gfpc2" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-whisker--6464b44b6c--gfpc2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-whisker--6464b44b6c--gfpc2-eth0", GenerateName:"whisker-6464b44b6c-", Namespace:"calico-system", SelfLink:"", UID:"148d61b5-bb87-4239-99d0-544a4ea406a1", ResourceVersion:"899", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 16, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"6464b44b6c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"", Pod:"whisker-6464b44b6c-gfpc2", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.1.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali61953bffda8", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:05.563436 containerd[1545]: 2025-05-27 03:16:05.524 [INFO][3845] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.193/32] ContainerID="78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" Namespace="calico-system" Pod="whisker-6464b44b6c-gfpc2" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-whisker--6464b44b6c--gfpc2-eth0" May 27 03:16:05.563687 containerd[1545]: 2025-05-27 03:16:05.524 [INFO][3845] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali61953bffda8 ContainerID="78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" Namespace="calico-system" Pod="whisker-6464b44b6c-gfpc2" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-whisker--6464b44b6c--gfpc2-eth0" May 27 03:16:05.563687 containerd[1545]: 2025-05-27 03:16:05.539 [INFO][3845] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" Namespace="calico-system" Pod="whisker-6464b44b6c-gfpc2" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-whisker--6464b44b6c--gfpc2-eth0" May 27 03:16:05.564828 containerd[1545]: 2025-05-27 03:16:05.540 [INFO][3845] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" Namespace="calico-system" Pod="whisker-6464b44b6c-gfpc2" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-whisker--6464b44b6c--gfpc2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-whisker--6464b44b6c--gfpc2-eth0", GenerateName:"whisker-6464b44b6c-", Namespace:"calico-system", SelfLink:"", UID:"148d61b5-bb87-4239-99d0-544a4ea406a1", ResourceVersion:"899", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 16, 4, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"6464b44b6c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d", Pod:"whisker-6464b44b6c-gfpc2", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.1.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali61953bffda8", MAC:"26:7e:d2:9b:31:2b", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:05.564935 containerd[1545]: 2025-05-27 03:16:05.555 [INFO][3845] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" Namespace="calico-system" Pod="whisker-6464b44b6c-gfpc2" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-whisker--6464b44b6c--gfpc2-eth0" May 27 03:16:05.776091 kubelet[2688]: I0527 03:16:05.775222 2688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 27 03:16:05.785154 containerd[1545]: time="2025-05-27T03:16:05.785080921Z" level=info msg="connecting to shim 78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d" address="unix:///run/containerd/s/7bc1b0e51f4ec9ce8e831ce8ac1e39154a1b8dc45bffab8c1f77fae6b8f845de" namespace=k8s.io protocol=ttrpc version=3 May 27 03:16:05.836351 systemd[1]: Started cri-containerd-78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d.scope - libcontainer container 78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d. May 27 03:16:06.076177 containerd[1545]: time="2025-05-27T03:16:06.076127454Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-6464b44b6c-gfpc2,Uid:148d61b5-bb87-4239-99d0-544a4ea406a1,Namespace:calico-system,Attempt:0,} returns sandbox id \"78b65ae1762eded6ab101947cceeb9da45de6f7bfeefe1e99c1dae7c2cd1f98d\"" May 27 03:16:06.099079 containerd[1545]: time="2025-05-27T03:16:06.097795325Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.0\"" May 27 03:16:06.392724 containerd[1545]: time="2025-05-27T03:16:06.392207757Z" level=info msg="fetch failed" error="failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" host=ghcr.io May 27 03:16:06.394141 containerd[1545]: time="2025-05-27T03:16:06.394070430Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.0: active requests=0, bytes read=86" May 27 03:16:06.394813 containerd[1545]: time="2025-05-27T03:16:06.394044720Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.0\" failed" error="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" May 27 03:16:06.405232 kubelet[2688]: E0527 03:16:06.404935 2688 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/whisker:v3.30.0" May 27 03:16:06.406561 kubelet[2688]: E0527 03:16:06.405262 2688 kuberuntime_image.go:55] "Failed to pull image" err="failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/whisker:v3.30.0" May 27 03:16:06.408154 kubelet[2688]: E0527 03:16:06.408029 2688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.0,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.0,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:961d12259f7740ee97ccbf2715e8204e,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bs6dk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-6464b44b6c-gfpc2_calico-system(148d61b5-bb87-4239-99d0-544a4ea406a1): ErrImagePull: failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" logger="UnhandledError" May 27 03:16:06.411464 containerd[1545]: time="2025-05-27T03:16:06.411018630Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\"" May 27 03:16:06.524035 kubelet[2688]: I0527 03:16:06.523961 2688 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3d83951-7aca-4e07-9693-6aa4b63d6615" path="/var/lib/kubelet/pods/e3d83951-7aca-4e07-9693-6aa4b63d6615/volumes" May 27 03:16:06.637641 containerd[1545]: time="2025-05-27T03:16:06.637429261Z" level=info msg="fetch failed" error="failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" host=ghcr.io May 27 03:16:06.638483 containerd[1545]: time="2025-05-27T03:16:06.638393335Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\" failed" error="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" May 27 03:16:06.638986 containerd[1545]: time="2025-05-27T03:16:06.638441478Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.0: active requests=0, bytes read=86" May 27 03:16:06.639036 kubelet[2688]: E0527 03:16:06.638928 2688 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.0" May 27 03:16:06.639036 kubelet[2688]: E0527 03:16:06.638998 2688 kuberuntime_image.go:55] "Failed to pull image" err="failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.0" May 27 03:16:06.639411 kubelet[2688]: E0527 03:16:06.639336 2688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.0,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bs6dk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-6464b44b6c-gfpc2_calico-system(148d61b5-bb87-4239-99d0-544a4ea406a1): ErrImagePull: failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" logger="UnhandledError" May 27 03:16:06.641066 kubelet[2688]: E0527 03:16:06.640897 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.0\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.0\\\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\\\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden\"]" pod="calico-system/whisker-6464b44b6c-gfpc2" podUID="148d61b5-bb87-4239-99d0-544a4ea406a1" May 27 03:16:06.793718 kubelet[2688]: E0527 03:16:06.793671 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.0\\\"\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\\\"\"]" pod="calico-system/whisker-6464b44b6c-gfpc2" podUID="148d61b5-bb87-4239-99d0-544a4ea406a1" May 27 03:16:07.060453 systemd-networkd[1448]: vxlan.calico: Link UP May 27 03:16:07.060466 systemd-networkd[1448]: vxlan.calico: Gained carrier May 27 03:16:07.461379 systemd-networkd[1448]: cali61953bffda8: Gained IPv6LL May 27 03:16:07.795414 kubelet[2688]: E0527 03:16:07.794839 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.0\\\"\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\\\"\"]" pod="calico-system/whisker-6464b44b6c-gfpc2" podUID="148d61b5-bb87-4239-99d0-544a4ea406a1" May 27 03:16:08.504414 kubelet[2688]: E0527 03:16:08.503124 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:08.505368 containerd[1545]: time="2025-05-27T03:16:08.505306532Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-bl2d5,Uid:5ef430cf-43c4-4f8a-96de-9c8721fea30a,Namespace:kube-system,Attempt:0,}" May 27 03:16:08.506153 containerd[1545]: time="2025-05-27T03:16:08.505774427Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5d8fb586b5-2dbbs,Uid:8c1deb33-d2c3-4f94-b199-70a79bc92bde,Namespace:calico-system,Attempt:0,}" May 27 03:16:08.677828 systemd-networkd[1448]: vxlan.calico: Gained IPv6LL May 27 03:16:08.700749 systemd-networkd[1448]: calif382ef6ab7e: Link UP May 27 03:16:08.706784 systemd-networkd[1448]: calif382ef6ab7e: Gained carrier May 27 03:16:08.731920 containerd[1545]: 2025-05-27 03:16:08.580 [INFO][4113] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4344.0.0--4--f2dd3e9233-k8s-calico--kube--controllers--5d8fb586b5--2dbbs-eth0 calico-kube-controllers-5d8fb586b5- calico-system 8c1deb33-d2c3-4f94-b199-70a79bc92bde 833 0 2025-05-27 03:15:47 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:5d8fb586b5 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4344.0.0-4-f2dd3e9233 calico-kube-controllers-5d8fb586b5-2dbbs eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calif382ef6ab7e [] [] }} ContainerID="1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" Namespace="calico-system" Pod="calico-kube-controllers-5d8fb586b5-2dbbs" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--kube--controllers--5d8fb586b5--2dbbs-" May 27 03:16:08.731920 containerd[1545]: 2025-05-27 03:16:08.580 [INFO][4113] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" Namespace="calico-system" Pod="calico-kube-controllers-5d8fb586b5-2dbbs" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--kube--controllers--5d8fb586b5--2dbbs-eth0" May 27 03:16:08.731920 containerd[1545]: 2025-05-27 03:16:08.632 [INFO][4137] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" HandleID="k8s-pod-network.1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-calico--kube--controllers--5d8fb586b5--2dbbs-eth0" May 27 03:16:08.732809 containerd[1545]: 2025-05-27 03:16:08.632 [INFO][4137] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" HandleID="k8s-pod-network.1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-calico--kube--controllers--5d8fb586b5--2dbbs-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000233670), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4344.0.0-4-f2dd3e9233", "pod":"calico-kube-controllers-5d8fb586b5-2dbbs", "timestamp":"2025-05-27 03:16:08.632322252 +0000 UTC"}, Hostname:"ci-4344.0.0-4-f2dd3e9233", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 27 03:16:08.732809 containerd[1545]: 2025-05-27 03:16:08.632 [INFO][4137] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 27 03:16:08.732809 containerd[1545]: 2025-05-27 03:16:08.632 [INFO][4137] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 27 03:16:08.732809 containerd[1545]: 2025-05-27 03:16:08.632 [INFO][4137] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4344.0.0-4-f2dd3e9233' May 27 03:16:08.732809 containerd[1545]: 2025-05-27 03:16:08.641 [INFO][4137] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.732809 containerd[1545]: 2025-05-27 03:16:08.654 [INFO][4137] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.732809 containerd[1545]: 2025-05-27 03:16:08.661 [INFO][4137] ipam/ipam.go 511: Trying affinity for 192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.732809 containerd[1545]: 2025-05-27 03:16:08.665 [INFO][4137] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.732809 containerd[1545]: 2025-05-27 03:16:08.668 [INFO][4137] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.733512 containerd[1545]: 2025-05-27 03:16:08.668 [INFO][4137] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.192/26 handle="k8s-pod-network.1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.733512 containerd[1545]: 2025-05-27 03:16:08.671 [INFO][4137] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87 May 27 03:16:08.733512 containerd[1545]: 2025-05-27 03:16:08.676 [INFO][4137] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.192/26 handle="k8s-pod-network.1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.733512 containerd[1545]: 2025-05-27 03:16:08.690 [INFO][4137] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.194/26] block=192.168.1.192/26 handle="k8s-pod-network.1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.733512 containerd[1545]: 2025-05-27 03:16:08.690 [INFO][4137] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.194/26] handle="k8s-pod-network.1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.733512 containerd[1545]: 2025-05-27 03:16:08.690 [INFO][4137] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 27 03:16:08.733512 containerd[1545]: 2025-05-27 03:16:08.690 [INFO][4137] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.194/26] IPv6=[] ContainerID="1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" HandleID="k8s-pod-network.1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-calico--kube--controllers--5d8fb586b5--2dbbs-eth0" May 27 03:16:08.733680 containerd[1545]: 2025-05-27 03:16:08.695 [INFO][4113] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" Namespace="calico-system" Pod="calico-kube-controllers-5d8fb586b5-2dbbs" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--kube--controllers--5d8fb586b5--2dbbs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-calico--kube--controllers--5d8fb586b5--2dbbs-eth0", GenerateName:"calico-kube-controllers-5d8fb586b5-", Namespace:"calico-system", SelfLink:"", UID:"8c1deb33-d2c3-4f94-b199-70a79bc92bde", ResourceVersion:"833", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 15, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5d8fb586b5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"", Pod:"calico-kube-controllers-5d8fb586b5-2dbbs", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.1.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calif382ef6ab7e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:08.733750 containerd[1545]: 2025-05-27 03:16:08.695 [INFO][4113] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.194/32] ContainerID="1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" Namespace="calico-system" Pod="calico-kube-controllers-5d8fb586b5-2dbbs" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--kube--controllers--5d8fb586b5--2dbbs-eth0" May 27 03:16:08.733750 containerd[1545]: 2025-05-27 03:16:08.695 [INFO][4113] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif382ef6ab7e ContainerID="1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" Namespace="calico-system" Pod="calico-kube-controllers-5d8fb586b5-2dbbs" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--kube--controllers--5d8fb586b5--2dbbs-eth0" May 27 03:16:08.733750 containerd[1545]: 2025-05-27 03:16:08.708 [INFO][4113] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" Namespace="calico-system" Pod="calico-kube-controllers-5d8fb586b5-2dbbs" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--kube--controllers--5d8fb586b5--2dbbs-eth0" May 27 03:16:08.733822 containerd[1545]: 2025-05-27 03:16:08.709 [INFO][4113] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" Namespace="calico-system" Pod="calico-kube-controllers-5d8fb586b5-2dbbs" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--kube--controllers--5d8fb586b5--2dbbs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-calico--kube--controllers--5d8fb586b5--2dbbs-eth0", GenerateName:"calico-kube-controllers-5d8fb586b5-", Namespace:"calico-system", SelfLink:"", UID:"8c1deb33-d2c3-4f94-b199-70a79bc92bde", ResourceVersion:"833", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 15, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5d8fb586b5", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87", Pod:"calico-kube-controllers-5d8fb586b5-2dbbs", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.1.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calif382ef6ab7e", MAC:"72:8e:01:8d:ab:09", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:08.733890 containerd[1545]: 2025-05-27 03:16:08.725 [INFO][4113] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" Namespace="calico-system" Pod="calico-kube-controllers-5d8fb586b5-2dbbs" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--kube--controllers--5d8fb586b5--2dbbs-eth0" May 27 03:16:08.790848 containerd[1545]: time="2025-05-27T03:16:08.790417011Z" level=info msg="connecting to shim 1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87" address="unix:///run/containerd/s/9c83954df286f96f2e7acded33dba3afb4e5d81f42f9045a5e1f42ff56d846a2" namespace=k8s.io protocol=ttrpc version=3 May 27 03:16:08.857415 systemd-networkd[1448]: cali604729e09b5: Link UP May 27 03:16:08.858473 systemd[1]: Started cri-containerd-1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87.scope - libcontainer container 1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87. May 27 03:16:08.862969 systemd-networkd[1448]: cali604729e09b5: Gained carrier May 27 03:16:08.892099 containerd[1545]: 2025-05-27 03:16:08.597 [INFO][4119] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--bl2d5-eth0 coredns-7c65d6cfc9- kube-system 5ef430cf-43c4-4f8a-96de-9c8721fea30a 834 0 2025-05-27 03:15:31 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4344.0.0-4-f2dd3e9233 coredns-7c65d6cfc9-bl2d5 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali604729e09b5 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" Namespace="kube-system" Pod="coredns-7c65d6cfc9-bl2d5" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--bl2d5-" May 27 03:16:08.892099 containerd[1545]: 2025-05-27 03:16:08.597 [INFO][4119] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" Namespace="kube-system" Pod="coredns-7c65d6cfc9-bl2d5" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--bl2d5-eth0" May 27 03:16:08.892099 containerd[1545]: 2025-05-27 03:16:08.645 [INFO][4142] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" HandleID="k8s-pod-network.7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--bl2d5-eth0" May 27 03:16:08.892622 containerd[1545]: 2025-05-27 03:16:08.645 [INFO][4142] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" HandleID="k8s-pod-network.7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--bl2d5-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000233020), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4344.0.0-4-f2dd3e9233", "pod":"coredns-7c65d6cfc9-bl2d5", "timestamp":"2025-05-27 03:16:08.645264109 +0000 UTC"}, Hostname:"ci-4344.0.0-4-f2dd3e9233", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 27 03:16:08.892622 containerd[1545]: 2025-05-27 03:16:08.645 [INFO][4142] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 27 03:16:08.892622 containerd[1545]: 2025-05-27 03:16:08.691 [INFO][4142] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 27 03:16:08.892622 containerd[1545]: 2025-05-27 03:16:08.691 [INFO][4142] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4344.0.0-4-f2dd3e9233' May 27 03:16:08.892622 containerd[1545]: 2025-05-27 03:16:08.744 [INFO][4142] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.892622 containerd[1545]: 2025-05-27 03:16:08.756 [INFO][4142] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.892622 containerd[1545]: 2025-05-27 03:16:08.778 [INFO][4142] ipam/ipam.go 511: Trying affinity for 192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.892622 containerd[1545]: 2025-05-27 03:16:08.787 [INFO][4142] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.892622 containerd[1545]: 2025-05-27 03:16:08.801 [INFO][4142] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.894281 containerd[1545]: 2025-05-27 03:16:08.804 [INFO][4142] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.192/26 handle="k8s-pod-network.7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.894281 containerd[1545]: 2025-05-27 03:16:08.813 [INFO][4142] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107 May 27 03:16:08.894281 containerd[1545]: 2025-05-27 03:16:08.828 [INFO][4142] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.192/26 handle="k8s-pod-network.7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.894281 containerd[1545]: 2025-05-27 03:16:08.842 [INFO][4142] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.195/26] block=192.168.1.192/26 handle="k8s-pod-network.7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.894281 containerd[1545]: 2025-05-27 03:16:08.842 [INFO][4142] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.195/26] handle="k8s-pod-network.7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:08.894281 containerd[1545]: 2025-05-27 03:16:08.842 [INFO][4142] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 27 03:16:08.894281 containerd[1545]: 2025-05-27 03:16:08.842 [INFO][4142] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.195/26] IPv6=[] ContainerID="7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" HandleID="k8s-pod-network.7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--bl2d5-eth0" May 27 03:16:08.894794 containerd[1545]: 2025-05-27 03:16:08.847 [INFO][4119] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" Namespace="kube-system" Pod="coredns-7c65d6cfc9-bl2d5" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--bl2d5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--bl2d5-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"5ef430cf-43c4-4f8a-96de-9c8721fea30a", ResourceVersion:"834", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 15, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"", Pod:"coredns-7c65d6cfc9-bl2d5", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.1.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali604729e09b5", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:08.894794 containerd[1545]: 2025-05-27 03:16:08.848 [INFO][4119] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.195/32] ContainerID="7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" Namespace="kube-system" Pod="coredns-7c65d6cfc9-bl2d5" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--bl2d5-eth0" May 27 03:16:08.894794 containerd[1545]: 2025-05-27 03:16:08.848 [INFO][4119] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali604729e09b5 ContainerID="7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" Namespace="kube-system" Pod="coredns-7c65d6cfc9-bl2d5" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--bl2d5-eth0" May 27 03:16:08.894794 containerd[1545]: 2025-05-27 03:16:08.865 [INFO][4119] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" Namespace="kube-system" Pod="coredns-7c65d6cfc9-bl2d5" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--bl2d5-eth0" May 27 03:16:08.894794 containerd[1545]: 2025-05-27 03:16:08.867 [INFO][4119] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" Namespace="kube-system" Pod="coredns-7c65d6cfc9-bl2d5" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--bl2d5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--bl2d5-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"5ef430cf-43c4-4f8a-96de-9c8721fea30a", ResourceVersion:"834", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 15, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107", Pod:"coredns-7c65d6cfc9-bl2d5", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.1.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali604729e09b5", MAC:"5a:de:96:f0:27:15", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:08.894794 containerd[1545]: 2025-05-27 03:16:08.885 [INFO][4119] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" Namespace="kube-system" Pod="coredns-7c65d6cfc9-bl2d5" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--bl2d5-eth0" May 27 03:16:08.923134 containerd[1545]: time="2025-05-27T03:16:08.923071015Z" level=info msg="connecting to shim 7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107" address="unix:///run/containerd/s/52e8936084d7f0c669869c8e054818bb478cee54db86167fff14b94cc7c9c5cb" namespace=k8s.io protocol=ttrpc version=3 May 27 03:16:08.968463 systemd[1]: Started cri-containerd-7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107.scope - libcontainer container 7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107. May 27 03:16:09.042861 containerd[1545]: time="2025-05-27T03:16:09.042630686Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5d8fb586b5-2dbbs,Uid:8c1deb33-d2c3-4f94-b199-70a79bc92bde,Namespace:calico-system,Attempt:0,} returns sandbox id \"1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87\"" May 27 03:16:09.051871 containerd[1545]: time="2025-05-27T03:16:09.051674572Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.0\"" May 27 03:16:09.070952 containerd[1545]: time="2025-05-27T03:16:09.070864430Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-bl2d5,Uid:5ef430cf-43c4-4f8a-96de-9c8721fea30a,Namespace:kube-system,Attempt:0,} returns sandbox id \"7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107\"" May 27 03:16:09.072834 kubelet[2688]: E0527 03:16:09.072732 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:09.080914 containerd[1545]: time="2025-05-27T03:16:09.080788714Z" level=info msg="CreateContainer within sandbox \"7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 27 03:16:09.105530 containerd[1545]: time="2025-05-27T03:16:09.105460750Z" level=info msg="Container ce4c0b9c36c443ed213606905936b78763552349018984cb651d84ca973501ba: CDI devices from CRI Config.CDIDevices: []" May 27 03:16:09.121581 containerd[1545]: time="2025-05-27T03:16:09.121488705Z" level=info msg="CreateContainer within sandbox \"7e0f6e244d3cfc5331f1615744c09249308dd1ed36e37a60233cc78387e87107\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ce4c0b9c36c443ed213606905936b78763552349018984cb651d84ca973501ba\"" May 27 03:16:09.123842 containerd[1545]: time="2025-05-27T03:16:09.122522066Z" level=info msg="StartContainer for \"ce4c0b9c36c443ed213606905936b78763552349018984cb651d84ca973501ba\"" May 27 03:16:09.124574 containerd[1545]: time="2025-05-27T03:16:09.124523458Z" level=info msg="connecting to shim ce4c0b9c36c443ed213606905936b78763552349018984cb651d84ca973501ba" address="unix:///run/containerd/s/52e8936084d7f0c669869c8e054818bb478cee54db86167fff14b94cc7c9c5cb" protocol=ttrpc version=3 May 27 03:16:09.168397 systemd[1]: Started cri-containerd-ce4c0b9c36c443ed213606905936b78763552349018984cb651d84ca973501ba.scope - libcontainer container ce4c0b9c36c443ed213606905936b78763552349018984cb651d84ca973501ba. May 27 03:16:09.221854 containerd[1545]: time="2025-05-27T03:16:09.221785363Z" level=info msg="StartContainer for \"ce4c0b9c36c443ed213606905936b78763552349018984cb651d84ca973501ba\" returns successfully" May 27 03:16:09.504135 containerd[1545]: time="2025-05-27T03:16:09.504011738Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drsns,Uid:851b262b-7562-4011-b67e-5e2de773f0d3,Namespace:calico-system,Attempt:0,}" May 27 03:16:09.537888 kubelet[2688]: I0527 03:16:09.537707 2688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 27 03:16:09.734505 systemd-networkd[1448]: cali0827c0b152e: Link UP May 27 03:16:09.738130 systemd-networkd[1448]: cali0827c0b152e: Gained carrier May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.581 [INFO][4296] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4344.0.0--4--f2dd3e9233-k8s-csi--node--driver--drsns-eth0 csi-node-driver- calico-system 851b262b-7562-4011-b67e-5e2de773f0d3 716 0 2025-05-27 03:15:47 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:68bf44dd5 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4344.0.0-4-f2dd3e9233 csi-node-driver-drsns eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali0827c0b152e [] [] }} ContainerID="fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" Namespace="calico-system" Pod="csi-node-driver-drsns" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-csi--node--driver--drsns-" May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.581 [INFO][4296] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" Namespace="calico-system" Pod="csi-node-driver-drsns" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-csi--node--driver--drsns-eth0" May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.648 [INFO][4308] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" HandleID="k8s-pod-network.fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-csi--node--driver--drsns-eth0" May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.649 [INFO][4308] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" HandleID="k8s-pod-network.fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-csi--node--driver--drsns-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d9270), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4344.0.0-4-f2dd3e9233", "pod":"csi-node-driver-drsns", "timestamp":"2025-05-27 03:16:09.648771772 +0000 UTC"}, Hostname:"ci-4344.0.0-4-f2dd3e9233", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.649 [INFO][4308] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.649 [INFO][4308] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.649 [INFO][4308] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4344.0.0-4-f2dd3e9233' May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.665 [INFO][4308] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.677 [INFO][4308] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.687 [INFO][4308] ipam/ipam.go 511: Trying affinity for 192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.692 [INFO][4308] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.698 [INFO][4308] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.698 [INFO][4308] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.192/26 handle="k8s-pod-network.fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.701 [INFO][4308] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521 May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.712 [INFO][4308] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.192/26 handle="k8s-pod-network.fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.721 [INFO][4308] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.196/26] block=192.168.1.192/26 handle="k8s-pod-network.fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.722 [INFO][4308] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.196/26] handle="k8s-pod-network.fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.722 [INFO][4308] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 27 03:16:09.774405 containerd[1545]: 2025-05-27 03:16:09.722 [INFO][4308] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.196/26] IPv6=[] ContainerID="fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" HandleID="k8s-pod-network.fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-csi--node--driver--drsns-eth0" May 27 03:16:09.776697 containerd[1545]: 2025-05-27 03:16:09.726 [INFO][4296] cni-plugin/k8s.go 418: Populated endpoint ContainerID="fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" Namespace="calico-system" Pod="csi-node-driver-drsns" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-csi--node--driver--drsns-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-csi--node--driver--drsns-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"851b262b-7562-4011-b67e-5e2de773f0d3", ResourceVersion:"716", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 15, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"68bf44dd5", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"", Pod:"csi-node-driver-drsns", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.1.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali0827c0b152e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:09.776697 containerd[1545]: 2025-05-27 03:16:09.726 [INFO][4296] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.196/32] ContainerID="fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" Namespace="calico-system" Pod="csi-node-driver-drsns" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-csi--node--driver--drsns-eth0" May 27 03:16:09.776697 containerd[1545]: 2025-05-27 03:16:09.726 [INFO][4296] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali0827c0b152e ContainerID="fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" Namespace="calico-system" Pod="csi-node-driver-drsns" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-csi--node--driver--drsns-eth0" May 27 03:16:09.776697 containerd[1545]: 2025-05-27 03:16:09.738 [INFO][4296] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" Namespace="calico-system" Pod="csi-node-driver-drsns" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-csi--node--driver--drsns-eth0" May 27 03:16:09.776697 containerd[1545]: 2025-05-27 03:16:09.741 [INFO][4296] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" Namespace="calico-system" Pod="csi-node-driver-drsns" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-csi--node--driver--drsns-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-csi--node--driver--drsns-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"851b262b-7562-4011-b67e-5e2de773f0d3", ResourceVersion:"716", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 15, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"68bf44dd5", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521", Pod:"csi-node-driver-drsns", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.1.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali0827c0b152e", MAC:"d6:af:c7:5e:63:62", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:09.776697 containerd[1545]: 2025-05-27 03:16:09.766 [INFO][4296] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" Namespace="calico-system" Pod="csi-node-driver-drsns" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-csi--node--driver--drsns-eth0" May 27 03:16:09.814169 kubelet[2688]: E0527 03:16:09.813461 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:09.824772 containerd[1545]: time="2025-05-27T03:16:09.824723826Z" level=info msg="connecting to shim fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521" address="unix:///run/containerd/s/00284308e71b6df3068cacd6548b2c432015b0dd67248b07309b83ef25209286" namespace=k8s.io protocol=ttrpc version=3 May 27 03:16:09.869075 kubelet[2688]: I0527 03:16:09.868750 2688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-bl2d5" podStartSLOduration=38.868728167 podStartE2EDuration="38.868728167s" podCreationTimestamp="2025-05-27 03:15:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-27 03:16:09.831213322 +0000 UTC m=+43.532825638" watchObservedRunningTime="2025-05-27 03:16:09.868728167 +0000 UTC m=+43.570340481" May 27 03:16:09.899683 systemd[1]: Started cri-containerd-fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521.scope - libcontainer container fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521. May 27 03:16:09.979541 containerd[1545]: time="2025-05-27T03:16:09.979466613Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drsns,Uid:851b262b-7562-4011-b67e-5e2de773f0d3,Namespace:calico-system,Attempt:0,} returns sandbox id \"fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521\"" May 27 03:16:10.031710 containerd[1545]: time="2025-05-27T03:16:10.031556640Z" level=info msg="TaskExit event in podsandbox handler container_id:\"320d5852593cb9fdb7fcd6a5a8f42623b3c597874b90e987eaa194b5384f5175\" id:\"64f3d1d7211a28ab64831a429ab25b5b7daab64993dfa6311cbc9af49910c2b5\" pid:4330 exited_at:{seconds:1748315770 nanos:26294968}" May 27 03:16:10.182784 containerd[1545]: time="2025-05-27T03:16:10.182729378Z" level=info msg="TaskExit event in podsandbox handler container_id:\"320d5852593cb9fdb7fcd6a5a8f42623b3c597874b90e987eaa194b5384f5175\" id:\"4e78480d2d123020eda037bd38e0eb3edf6221d510da1832202d0c1b721fd685\" pid:4413 exited_at:{seconds:1748315770 nanos:182213860}" May 27 03:16:10.213255 systemd-networkd[1448]: calif382ef6ab7e: Gained IPv6LL May 27 03:16:10.504668 containerd[1545]: time="2025-05-27T03:16:10.504034134Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-77c54fb4c6-tgcsm,Uid:166fd789-e64d-4d0b-940a-5cec66cd52be,Namespace:calico-apiserver,Attempt:0,}" May 27 03:16:10.518102 containerd[1545]: time="2025-05-27T03:16:10.517966813Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-77c54fb4c6-j9z9b,Uid:00c40ec8-4084-40ee-b388-e582df6866d9,Namespace:calico-apiserver,Attempt:0,}" May 27 03:16:10.776811 systemd-networkd[1448]: calif9a9329fcf4: Link UP May 27 03:16:10.777132 systemd-networkd[1448]: calif9a9329fcf4: Gained carrier May 27 03:16:10.840346 kubelet[2688]: E0527 03:16:10.840107 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.627 [INFO][4424] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--tgcsm-eth0 calico-apiserver-77c54fb4c6- calico-apiserver 166fd789-e64d-4d0b-940a-5cec66cd52be 829 0 2025-05-27 03:15:43 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:77c54fb4c6 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4344.0.0-4-f2dd3e9233 calico-apiserver-77c54fb4c6-tgcsm eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calif9a9329fcf4 [] [] }} ContainerID="5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-tgcsm" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--tgcsm-" May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.629 [INFO][4424] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-tgcsm" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--tgcsm-eth0" May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.680 [INFO][4448] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" HandleID="k8s-pod-network.5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--tgcsm-eth0" May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.680 [INFO][4448] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" HandleID="k8s-pod-network.5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--tgcsm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000332ff0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4344.0.0-4-f2dd3e9233", "pod":"calico-apiserver-77c54fb4c6-tgcsm", "timestamp":"2025-05-27 03:16:10.680290609 +0000 UTC"}, Hostname:"ci-4344.0.0-4-f2dd3e9233", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.680 [INFO][4448] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.681 [INFO][4448] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.681 [INFO][4448] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4344.0.0-4-f2dd3e9233' May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.693 [INFO][4448] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.703 [INFO][4448] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.719 [INFO][4448] ipam/ipam.go 511: Trying affinity for 192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.725 [INFO][4448] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.731 [INFO][4448] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.732 [INFO][4448] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.192/26 handle="k8s-pod-network.5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.737 [INFO][4448] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.747 [INFO][4448] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.192/26 handle="k8s-pod-network.5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.759 [INFO][4448] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.197/26] block=192.168.1.192/26 handle="k8s-pod-network.5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.759 [INFO][4448] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.197/26] handle="k8s-pod-network.5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.759 [INFO][4448] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 27 03:16:10.852089 containerd[1545]: 2025-05-27 03:16:10.759 [INFO][4448] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.197/26] IPv6=[] ContainerID="5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" HandleID="k8s-pod-network.5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--tgcsm-eth0" May 27 03:16:10.855557 containerd[1545]: 2025-05-27 03:16:10.770 [INFO][4424] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-tgcsm" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--tgcsm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--tgcsm-eth0", GenerateName:"calico-apiserver-77c54fb4c6-", Namespace:"calico-apiserver", SelfLink:"", UID:"166fd789-e64d-4d0b-940a-5cec66cd52be", ResourceVersion:"829", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 15, 43, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"77c54fb4c6", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"", Pod:"calico-apiserver-77c54fb4c6-tgcsm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.1.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calif9a9329fcf4", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:10.855557 containerd[1545]: 2025-05-27 03:16:10.770 [INFO][4424] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.197/32] ContainerID="5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-tgcsm" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--tgcsm-eth0" May 27 03:16:10.855557 containerd[1545]: 2025-05-27 03:16:10.771 [INFO][4424] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif9a9329fcf4 ContainerID="5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-tgcsm" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--tgcsm-eth0" May 27 03:16:10.855557 containerd[1545]: 2025-05-27 03:16:10.776 [INFO][4424] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-tgcsm" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--tgcsm-eth0" May 27 03:16:10.855557 containerd[1545]: 2025-05-27 03:16:10.780 [INFO][4424] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-tgcsm" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--tgcsm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--tgcsm-eth0", GenerateName:"calico-apiserver-77c54fb4c6-", Namespace:"calico-apiserver", SelfLink:"", UID:"166fd789-e64d-4d0b-940a-5cec66cd52be", ResourceVersion:"829", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 15, 43, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"77c54fb4c6", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a", Pod:"calico-apiserver-77c54fb4c6-tgcsm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.1.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calif9a9329fcf4", MAC:"42:70:c7:c8:41:16", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:10.855557 containerd[1545]: 2025-05-27 03:16:10.818 [INFO][4424] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-tgcsm" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--tgcsm-eth0" May 27 03:16:10.853300 systemd-networkd[1448]: cali604729e09b5: Gained IPv6LL May 27 03:16:10.918351 systemd-networkd[1448]: cali0827c0b152e: Gained IPv6LL May 27 03:16:10.974325 containerd[1545]: time="2025-05-27T03:16:10.974186117Z" level=info msg="connecting to shim 5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a" address="unix:///run/containerd/s/55f54969f014ddeb35fec70179d47832b8d46dfdd94d04d2851cc455d3a0bcfe" namespace=k8s.io protocol=ttrpc version=3 May 27 03:16:10.981105 systemd-networkd[1448]: caliadacd59e0c6: Link UP May 27 03:16:10.983506 systemd-networkd[1448]: caliadacd59e0c6: Gained carrier May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.633 [INFO][4434] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--j9z9b-eth0 calico-apiserver-77c54fb4c6- calico-apiserver 00c40ec8-4084-40ee-b388-e582df6866d9 831 0 2025-05-27 03:15:43 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:77c54fb4c6 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4344.0.0-4-f2dd3e9233 calico-apiserver-77c54fb4c6-j9z9b eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] caliadacd59e0c6 [] [] }} ContainerID="5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-j9z9b" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--j9z9b-" May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.634 [INFO][4434] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-j9z9b" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--j9z9b-eth0" May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.766 [INFO][4450] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" HandleID="k8s-pod-network.5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--j9z9b-eth0" May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.766 [INFO][4450] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" HandleID="k8s-pod-network.5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--j9z9b-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000233f10), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4344.0.0-4-f2dd3e9233", "pod":"calico-apiserver-77c54fb4c6-j9z9b", "timestamp":"2025-05-27 03:16:10.766351362 +0000 UTC"}, Hostname:"ci-4344.0.0-4-f2dd3e9233", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.767 [INFO][4450] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.767 [INFO][4450] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.767 [INFO][4450] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4344.0.0-4-f2dd3e9233' May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.797 [INFO][4450] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.830 [INFO][4450] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.858 [INFO][4450] ipam/ipam.go 511: Trying affinity for 192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.864 [INFO][4450] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.872 [INFO][4450] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.873 [INFO][4450] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.192/26 handle="k8s-pod-network.5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.879 [INFO][4450] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.899 [INFO][4450] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.192/26 handle="k8s-pod-network.5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.926 [INFO][4450] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.198/26] block=192.168.1.192/26 handle="k8s-pod-network.5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.926 [INFO][4450] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.198/26] handle="k8s-pod-network.5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.926 [INFO][4450] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 27 03:16:11.024214 containerd[1545]: 2025-05-27 03:16:10.926 [INFO][4450] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.198/26] IPv6=[] ContainerID="5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" HandleID="k8s-pod-network.5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--j9z9b-eth0" May 27 03:16:11.025508 containerd[1545]: 2025-05-27 03:16:10.938 [INFO][4434] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-j9z9b" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--j9z9b-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--j9z9b-eth0", GenerateName:"calico-apiserver-77c54fb4c6-", Namespace:"calico-apiserver", SelfLink:"", UID:"00c40ec8-4084-40ee-b388-e582df6866d9", ResourceVersion:"831", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 15, 43, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"77c54fb4c6", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"", Pod:"calico-apiserver-77c54fb4c6-j9z9b", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.1.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"caliadacd59e0c6", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:11.025508 containerd[1545]: 2025-05-27 03:16:10.942 [INFO][4434] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.198/32] ContainerID="5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-j9z9b" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--j9z9b-eth0" May 27 03:16:11.025508 containerd[1545]: 2025-05-27 03:16:10.942 [INFO][4434] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliadacd59e0c6 ContainerID="5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-j9z9b" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--j9z9b-eth0" May 27 03:16:11.025508 containerd[1545]: 2025-05-27 03:16:10.981 [INFO][4434] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-j9z9b" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--j9z9b-eth0" May 27 03:16:11.025508 containerd[1545]: 2025-05-27 03:16:10.985 [INFO][4434] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-j9z9b" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--j9z9b-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--j9z9b-eth0", GenerateName:"calico-apiserver-77c54fb4c6-", Namespace:"calico-apiserver", SelfLink:"", UID:"00c40ec8-4084-40ee-b388-e582df6866d9", ResourceVersion:"831", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 15, 43, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"77c54fb4c6", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd", Pod:"calico-apiserver-77c54fb4c6-j9z9b", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.1.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"caliadacd59e0c6", MAC:"b2:2a:46:ef:65:08", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:11.025508 containerd[1545]: 2025-05-27 03:16:11.009 [INFO][4434] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" Namespace="calico-apiserver" Pod="calico-apiserver-77c54fb4c6-j9z9b" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-calico--apiserver--77c54fb4c6--j9z9b-eth0" May 27 03:16:11.062649 systemd[1]: Started cri-containerd-5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a.scope - libcontainer container 5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a. May 27 03:16:11.117892 containerd[1545]: time="2025-05-27T03:16:11.117828826Z" level=info msg="connecting to shim 5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd" address="unix:///run/containerd/s/c3f02747d2b211ea79a2c9020afbec4c3f3a191a358c01464f16cecfbb9cdf0b" namespace=k8s.io protocol=ttrpc version=3 May 27 03:16:11.206315 systemd[1]: Started cri-containerd-5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd.scope - libcontainer container 5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd. May 27 03:16:11.234175 containerd[1545]: time="2025-05-27T03:16:11.234041114Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-77c54fb4c6-tgcsm,Uid:166fd789-e64d-4d0b-940a-5cec66cd52be,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a\"" May 27 03:16:11.303990 containerd[1545]: time="2025-05-27T03:16:11.303896583Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-77c54fb4c6-j9z9b,Uid:00c40ec8-4084-40ee-b388-e582df6866d9,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd\"" May 27 03:16:11.505469 containerd[1545]: time="2025-05-27T03:16:11.505413341Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-8f77d7b6c-zcnhc,Uid:34798e8e-55b5-4e78-832b-798f5586248d,Namespace:calico-system,Attempt:0,}" May 27 03:16:11.809281 systemd-networkd[1448]: cali31444f1909a: Link UP May 27 03:16:11.811185 systemd-networkd[1448]: cali31444f1909a: Gained carrier May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.608 [INFO][4578] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4344.0.0--4--f2dd3e9233-k8s-goldmane--8f77d7b6c--zcnhc-eth0 goldmane-8f77d7b6c- calico-system 34798e8e-55b5-4e78-832b-798f5586248d 835 0 2025-05-27 03:15:47 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:8f77d7b6c projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4344.0.0-4-f2dd3e9233 goldmane-8f77d7b6c-zcnhc eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali31444f1909a [] [] }} ContainerID="b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" Namespace="calico-system" Pod="goldmane-8f77d7b6c-zcnhc" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-goldmane--8f77d7b6c--zcnhc-" May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.609 [INFO][4578] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" Namespace="calico-system" Pod="goldmane-8f77d7b6c-zcnhc" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-goldmane--8f77d7b6c--zcnhc-eth0" May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.696 [INFO][4589] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" HandleID="k8s-pod-network.b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-goldmane--8f77d7b6c--zcnhc-eth0" May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.696 [INFO][4589] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" HandleID="k8s-pod-network.b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-goldmane--8f77d7b6c--zcnhc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002cfe50), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4344.0.0-4-f2dd3e9233", "pod":"goldmane-8f77d7b6c-zcnhc", "timestamp":"2025-05-27 03:16:11.696599191 +0000 UTC"}, Hostname:"ci-4344.0.0-4-f2dd3e9233", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.697 [INFO][4589] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.697 [INFO][4589] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.697 [INFO][4589] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4344.0.0-4-f2dd3e9233' May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.722 [INFO][4589] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.735 [INFO][4589] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.745 [INFO][4589] ipam/ipam.go 511: Trying affinity for 192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.750 [INFO][4589] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.757 [INFO][4589] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.758 [INFO][4589] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.192/26 handle="k8s-pod-network.b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.762 [INFO][4589] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4 May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.773 [INFO][4589] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.192/26 handle="k8s-pod-network.b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.792 [INFO][4589] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.199/26] block=192.168.1.192/26 handle="k8s-pod-network.b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.793 [INFO][4589] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.199/26] handle="k8s-pod-network.b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.793 [INFO][4589] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 27 03:16:11.846702 containerd[1545]: 2025-05-27 03:16:11.793 [INFO][4589] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.199/26] IPv6=[] ContainerID="b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" HandleID="k8s-pod-network.b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-goldmane--8f77d7b6c--zcnhc-eth0" May 27 03:16:11.847854 containerd[1545]: 2025-05-27 03:16:11.801 [INFO][4578] cni-plugin/k8s.go 418: Populated endpoint ContainerID="b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" Namespace="calico-system" Pod="goldmane-8f77d7b6c-zcnhc" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-goldmane--8f77d7b6c--zcnhc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-goldmane--8f77d7b6c--zcnhc-eth0", GenerateName:"goldmane-8f77d7b6c-", Namespace:"calico-system", SelfLink:"", UID:"34798e8e-55b5-4e78-832b-798f5586248d", ResourceVersion:"835", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 15, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"8f77d7b6c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"", Pod:"goldmane-8f77d7b6c-zcnhc", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.1.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali31444f1909a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:11.847854 containerd[1545]: 2025-05-27 03:16:11.801 [INFO][4578] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.199/32] ContainerID="b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" Namespace="calico-system" Pod="goldmane-8f77d7b6c-zcnhc" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-goldmane--8f77d7b6c--zcnhc-eth0" May 27 03:16:11.847854 containerd[1545]: 2025-05-27 03:16:11.801 [INFO][4578] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali31444f1909a ContainerID="b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" Namespace="calico-system" Pod="goldmane-8f77d7b6c-zcnhc" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-goldmane--8f77d7b6c--zcnhc-eth0" May 27 03:16:11.847854 containerd[1545]: 2025-05-27 03:16:11.810 [INFO][4578] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" Namespace="calico-system" Pod="goldmane-8f77d7b6c-zcnhc" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-goldmane--8f77d7b6c--zcnhc-eth0" May 27 03:16:11.847854 containerd[1545]: 2025-05-27 03:16:11.811 [INFO][4578] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" Namespace="calico-system" Pod="goldmane-8f77d7b6c-zcnhc" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-goldmane--8f77d7b6c--zcnhc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-goldmane--8f77d7b6c--zcnhc-eth0", GenerateName:"goldmane-8f77d7b6c-", Namespace:"calico-system", SelfLink:"", UID:"34798e8e-55b5-4e78-832b-798f5586248d", ResourceVersion:"835", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 15, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"8f77d7b6c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4", Pod:"goldmane-8f77d7b6c-zcnhc", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.1.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali31444f1909a", MAC:"6a:eb:1a:a5:95:da", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:11.847854 containerd[1545]: 2025-05-27 03:16:11.840 [INFO][4578] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" Namespace="calico-system" Pod="goldmane-8f77d7b6c-zcnhc" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-goldmane--8f77d7b6c--zcnhc-eth0" May 27 03:16:11.865678 kubelet[2688]: E0527 03:16:11.865572 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:11.948972 containerd[1545]: time="2025-05-27T03:16:11.948870423Z" level=info msg="connecting to shim b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4" address="unix:///run/containerd/s/29b344036461f0ecc1603fa2f8118da39ad4ea574b14704ab2dd7bfedfeda645" namespace=k8s.io protocol=ttrpc version=3 May 27 03:16:12.006393 systemd[1]: Started cri-containerd-b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4.scope - libcontainer container b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4. May 27 03:16:12.137378 containerd[1545]: time="2025-05-27T03:16:12.136497224Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-8f77d7b6c-zcnhc,Uid:34798e8e-55b5-4e78-832b-798f5586248d,Namespace:calico-system,Attempt:0,} returns sandbox id \"b9b5578d601a9cba623207f1bc33b73acb912fd5ddbcb4fa889d0bd231e120e4\"" May 27 03:16:12.262135 systemd-networkd[1448]: calif9a9329fcf4: Gained IPv6LL May 27 03:16:12.645664 systemd-networkd[1448]: caliadacd59e0c6: Gained IPv6LL May 27 03:16:12.784461 containerd[1545]: time="2025-05-27T03:16:12.784402833Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:12.785295 containerd[1545]: time="2025-05-27T03:16:12.785252867Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.0: active requests=0, bytes read=51178512" May 27 03:16:12.785816 containerd[1545]: time="2025-05-27T03:16:12.785776527Z" level=info msg="ImageCreate event name:\"sha256:094053209304a3d20e6561c18d37ac2dc4c7fbb68c1579d9864c303edebffa50\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:12.788137 containerd[1545]: time="2025-05-27T03:16:12.787683709Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:eb5bc5c9e7a71f1d8ea69bbcc8e54b84fb7ec1e32d919c8b148f80b770f20182\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:12.788835 containerd[1545]: time="2025-05-27T03:16:12.788797219Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.0\" with image id \"sha256:094053209304a3d20e6561c18d37ac2dc4c7fbb68c1579d9864c303edebffa50\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.0\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:eb5bc5c9e7a71f1d8ea69bbcc8e54b84fb7ec1e32d919c8b148f80b770f20182\", size \"52671183\" in 3.736966034s" May 27 03:16:12.788995 containerd[1545]: time="2025-05-27T03:16:12.788841189Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.0\" returns image reference \"sha256:094053209304a3d20e6561c18d37ac2dc4c7fbb68c1579d9864c303edebffa50\"" May 27 03:16:12.791090 containerd[1545]: time="2025-05-27T03:16:12.790470827Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.0\"" May 27 03:16:12.811207 containerd[1545]: time="2025-05-27T03:16:12.811139049Z" level=info msg="CreateContainer within sandbox \"1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" May 27 03:16:12.846275 containerd[1545]: time="2025-05-27T03:16:12.845392212Z" level=info msg="Container 5aa9593f0cbaaca77974f8548443b3ecf0d1b89be2ecc5c8fa483c1687a2f605: CDI devices from CRI Config.CDIDevices: []" May 27 03:16:12.855680 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount666850864.mount: Deactivated successfully. May 27 03:16:12.860614 containerd[1545]: time="2025-05-27T03:16:12.860566414Z" level=info msg="CreateContainer within sandbox \"1d24f0025aa01e89da2b4601b30b9e1c5dc15d72f3839b25066c09e008a0bb87\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"5aa9593f0cbaaca77974f8548443b3ecf0d1b89be2ecc5c8fa483c1687a2f605\"" May 27 03:16:12.862522 containerd[1545]: time="2025-05-27T03:16:12.862478718Z" level=info msg="StartContainer for \"5aa9593f0cbaaca77974f8548443b3ecf0d1b89be2ecc5c8fa483c1687a2f605\"" May 27 03:16:12.865932 containerd[1545]: time="2025-05-27T03:16:12.865706304Z" level=info msg="connecting to shim 5aa9593f0cbaaca77974f8548443b3ecf0d1b89be2ecc5c8fa483c1687a2f605" address="unix:///run/containerd/s/9c83954df286f96f2e7acded33dba3afb4e5d81f42f9045a5e1f42ff56d846a2" protocol=ttrpc version=3 May 27 03:16:12.906353 systemd[1]: Started cri-containerd-5aa9593f0cbaaca77974f8548443b3ecf0d1b89be2ecc5c8fa483c1687a2f605.scope - libcontainer container 5aa9593f0cbaaca77974f8548443b3ecf0d1b89be2ecc5c8fa483c1687a2f605. May 27 03:16:12.984274 containerd[1545]: time="2025-05-27T03:16:12.984209167Z" level=info msg="StartContainer for \"5aa9593f0cbaaca77974f8548443b3ecf0d1b89be2ecc5c8fa483c1687a2f605\" returns successfully" May 27 03:16:13.286257 systemd-networkd[1448]: cali31444f1909a: Gained IPv6LL May 27 03:16:13.503314 kubelet[2688]: E0527 03:16:13.503277 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:13.504576 containerd[1545]: time="2025-05-27T03:16:13.504039696Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-rqshl,Uid:9e7b1075-ea3d-4037-941f-f559a61e62a6,Namespace:kube-system,Attempt:0,}" May 27 03:16:13.737592 systemd-networkd[1448]: calif16bf5174c7: Link UP May 27 03:16:13.739498 systemd-networkd[1448]: calif16bf5174c7: Gained carrier May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.578 [INFO][4699] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--rqshl-eth0 coredns-7c65d6cfc9- kube-system 9e7b1075-ea3d-4037-941f-f559a61e62a6 824 0 2025-05-27 03:15:31 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4344.0.0-4-f2dd3e9233 coredns-7c65d6cfc9-rqshl eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calif16bf5174c7 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" Namespace="kube-system" Pod="coredns-7c65d6cfc9-rqshl" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--rqshl-" May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.578 [INFO][4699] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" Namespace="kube-system" Pod="coredns-7c65d6cfc9-rqshl" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--rqshl-eth0" May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.661 [INFO][4711] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" HandleID="k8s-pod-network.0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--rqshl-eth0" May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.661 [INFO][4711] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" HandleID="k8s-pod-network.0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--rqshl-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d9d20), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4344.0.0-4-f2dd3e9233", "pod":"coredns-7c65d6cfc9-rqshl", "timestamp":"2025-05-27 03:16:13.661478525 +0000 UTC"}, Hostname:"ci-4344.0.0-4-f2dd3e9233", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.664 [INFO][4711] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.664 [INFO][4711] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.664 [INFO][4711] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4344.0.0-4-f2dd3e9233' May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.681 [INFO][4711] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.689 [INFO][4711] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.696 [INFO][4711] ipam/ipam.go 511: Trying affinity for 192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.700 [INFO][4711] ipam/ipam.go 158: Attempting to load block cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.704 [INFO][4711] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.1.192/26 host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.705 [INFO][4711] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.1.192/26 handle="k8s-pod-network.0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.707 [INFO][4711] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7 May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.713 [INFO][4711] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.1.192/26 handle="k8s-pod-network.0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.723 [INFO][4711] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.1.200/26] block=192.168.1.192/26 handle="k8s-pod-network.0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.724 [INFO][4711] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.1.200/26] handle="k8s-pod-network.0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" host="ci-4344.0.0-4-f2dd3e9233" May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.724 [INFO][4711] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 27 03:16:13.772278 containerd[1545]: 2025-05-27 03:16:13.724 [INFO][4711] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.1.200/26] IPv6=[] ContainerID="0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" HandleID="k8s-pod-network.0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" Workload="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--rqshl-eth0" May 27 03:16:13.774900 containerd[1545]: 2025-05-27 03:16:13.728 [INFO][4699] cni-plugin/k8s.go 418: Populated endpoint ContainerID="0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" Namespace="kube-system" Pod="coredns-7c65d6cfc9-rqshl" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--rqshl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--rqshl-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"9e7b1075-ea3d-4037-941f-f559a61e62a6", ResourceVersion:"824", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 15, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"", Pod:"coredns-7c65d6cfc9-rqshl", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.1.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calif16bf5174c7", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:13.774900 containerd[1545]: 2025-05-27 03:16:13.729 [INFO][4699] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.1.200/32] ContainerID="0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" Namespace="kube-system" Pod="coredns-7c65d6cfc9-rqshl" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--rqshl-eth0" May 27 03:16:13.774900 containerd[1545]: 2025-05-27 03:16:13.729 [INFO][4699] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif16bf5174c7 ContainerID="0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" Namespace="kube-system" Pod="coredns-7c65d6cfc9-rqshl" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--rqshl-eth0" May 27 03:16:13.774900 containerd[1545]: 2025-05-27 03:16:13.740 [INFO][4699] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" Namespace="kube-system" Pod="coredns-7c65d6cfc9-rqshl" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--rqshl-eth0" May 27 03:16:13.774900 containerd[1545]: 2025-05-27 03:16:13.741 [INFO][4699] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" Namespace="kube-system" Pod="coredns-7c65d6cfc9-rqshl" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--rqshl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--rqshl-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"9e7b1075-ea3d-4037-941f-f559a61e62a6", ResourceVersion:"824", Generation:0, CreationTimestamp:time.Date(2025, time.May, 27, 3, 15, 31, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4344.0.0-4-f2dd3e9233", ContainerID:"0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7", Pod:"coredns-7c65d6cfc9-rqshl", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.1.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calif16bf5174c7", MAC:"de:7a:ae:36:29:5e", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} May 27 03:16:13.774900 containerd[1545]: 2025-05-27 03:16:13.764 [INFO][4699] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" Namespace="kube-system" Pod="coredns-7c65d6cfc9-rqshl" WorkloadEndpoint="ci--4344.0.0--4--f2dd3e9233-k8s-coredns--7c65d6cfc9--rqshl-eth0" May 27 03:16:13.822782 containerd[1545]: time="2025-05-27T03:16:13.822662686Z" level=info msg="connecting to shim 0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7" address="unix:///run/containerd/s/942b003d0cc69eb2188f1b7f40f67a4b4f7c86d6a5790cb057b06858a465861f" namespace=k8s.io protocol=ttrpc version=3 May 27 03:16:13.863391 systemd[1]: Started cri-containerd-0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7.scope - libcontainer container 0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7. May 27 03:16:14.015699 containerd[1545]: time="2025-05-27T03:16:14.015550752Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-rqshl,Uid:9e7b1075-ea3d-4037-941f-f559a61e62a6,Namespace:kube-system,Attempt:0,} returns sandbox id \"0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7\"" May 27 03:16:14.019042 kubelet[2688]: E0527 03:16:14.018988 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:14.027514 containerd[1545]: time="2025-05-27T03:16:14.027466544Z" level=info msg="CreateContainer within sandbox \"0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 27 03:16:14.057457 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3151655215.mount: Deactivated successfully. May 27 03:16:14.062769 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2858547251.mount: Deactivated successfully. May 27 03:16:14.067102 containerd[1545]: time="2025-05-27T03:16:14.066226386Z" level=info msg="Container 1093a6ce6db5d567c48cbcf7eff1d85080af4f5be16a90e04528b6913cf5406c: CDI devices from CRI Config.CDIDevices: []" May 27 03:16:14.102087 containerd[1545]: time="2025-05-27T03:16:14.100995563Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5aa9593f0cbaaca77974f8548443b3ecf0d1b89be2ecc5c8fa483c1687a2f605\" id:\"fd3baadbeb35b127800381cbddfa040a6a89ce81e280b57e4ab6bc8f4b25af80\" pid:4780 exited_at:{seconds:1748315774 nanos:79366865}" May 27 03:16:14.135703 containerd[1545]: time="2025-05-27T03:16:14.135517213Z" level=info msg="CreateContainer within sandbox \"0907c03a5ff0c0c93cbf8c107047ecf4dd78aa4873027ce58162aad2dc1330b7\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"1093a6ce6db5d567c48cbcf7eff1d85080af4f5be16a90e04528b6913cf5406c\"" May 27 03:16:14.140094 containerd[1545]: time="2025-05-27T03:16:14.138928985Z" level=info msg="StartContainer for \"1093a6ce6db5d567c48cbcf7eff1d85080af4f5be16a90e04528b6913cf5406c\"" May 27 03:16:14.143079 containerd[1545]: time="2025-05-27T03:16:14.141712788Z" level=info msg="connecting to shim 1093a6ce6db5d567c48cbcf7eff1d85080af4f5be16a90e04528b6913cf5406c" address="unix:///run/containerd/s/942b003d0cc69eb2188f1b7f40f67a4b4f7c86d6a5790cb057b06858a465861f" protocol=ttrpc version=3 May 27 03:16:14.144297 kubelet[2688]: I0527 03:16:14.144235 2688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-5d8fb586b5-2dbbs" podStartSLOduration=23.402509873 podStartE2EDuration="27.144209494s" podCreationTimestamp="2025-05-27 03:15:47 +0000 UTC" firstStartedPulling="2025-05-27 03:16:09.048667084 +0000 UTC m=+42.750279392" lastFinishedPulling="2025-05-27 03:16:12.790366706 +0000 UTC m=+46.491979013" observedRunningTime="2025-05-27 03:16:13.913927662 +0000 UTC m=+47.615539976" watchObservedRunningTime="2025-05-27 03:16:14.144209494 +0000 UTC m=+47.845821809" May 27 03:16:14.197298 systemd[1]: Started cri-containerd-1093a6ce6db5d567c48cbcf7eff1d85080af4f5be16a90e04528b6913cf5406c.scope - libcontainer container 1093a6ce6db5d567c48cbcf7eff1d85080af4f5be16a90e04528b6913cf5406c. May 27 03:16:14.345408 containerd[1545]: time="2025-05-27T03:16:14.345258888Z" level=info msg="StartContainer for \"1093a6ce6db5d567c48cbcf7eff1d85080af4f5be16a90e04528b6913cf5406c\" returns successfully" May 27 03:16:14.519970 containerd[1545]: time="2025-05-27T03:16:14.519912140Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:14.521318 containerd[1545]: time="2025-05-27T03:16:14.521071419Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.0: active requests=0, bytes read=8758390" May 27 03:16:14.522486 containerd[1545]: time="2025-05-27T03:16:14.522439597Z" level=info msg="ImageCreate event name:\"sha256:d5b08093b7928c0ac1122e59edf69b2e58c6d10ecc8b9e5cffeb809a956dc48e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:14.533048 containerd[1545]: time="2025-05-27T03:16:14.532636468Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:27883a4104876fe239311dd93ce6efd0c4a87de7163d57a4c8d96bd65a287ffd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:14.533905 containerd[1545]: time="2025-05-27T03:16:14.533863593Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.0\" with image id \"sha256:d5b08093b7928c0ac1122e59edf69b2e58c6d10ecc8b9e5cffeb809a956dc48e\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.0\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:27883a4104876fe239311dd93ce6efd0c4a87de7163d57a4c8d96bd65a287ffd\", size \"10251093\" in 1.74335256s" May 27 03:16:14.534016 containerd[1545]: time="2025-05-27T03:16:14.533909443Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.0\" returns image reference \"sha256:d5b08093b7928c0ac1122e59edf69b2e58c6d10ecc8b9e5cffeb809a956dc48e\"" May 27 03:16:14.537713 containerd[1545]: time="2025-05-27T03:16:14.537675104Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.0\"" May 27 03:16:14.539739 containerd[1545]: time="2025-05-27T03:16:14.539699711Z" level=info msg="CreateContainer within sandbox \"fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" May 27 03:16:14.573115 containerd[1545]: time="2025-05-27T03:16:14.571991150Z" level=info msg="Container 7b5c40d4485c652db5fa04b64c356bfeee217a16f3a7b57be30fa2efb66423eb: CDI devices from CRI Config.CDIDevices: []" May 27 03:16:14.592241 containerd[1545]: time="2025-05-27T03:16:14.592186842Z" level=info msg="CreateContainer within sandbox \"fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"7b5c40d4485c652db5fa04b64c356bfeee217a16f3a7b57be30fa2efb66423eb\"" May 27 03:16:14.593152 containerd[1545]: time="2025-05-27T03:16:14.593114499Z" level=info msg="StartContainer for \"7b5c40d4485c652db5fa04b64c356bfeee217a16f3a7b57be30fa2efb66423eb\"" May 27 03:16:14.594662 containerd[1545]: time="2025-05-27T03:16:14.594582219Z" level=info msg="connecting to shim 7b5c40d4485c652db5fa04b64c356bfeee217a16f3a7b57be30fa2efb66423eb" address="unix:///run/containerd/s/00284308e71b6df3068cacd6548b2c432015b0dd67248b07309b83ef25209286" protocol=ttrpc version=3 May 27 03:16:14.628292 systemd[1]: Started cri-containerd-7b5c40d4485c652db5fa04b64c356bfeee217a16f3a7b57be30fa2efb66423eb.scope - libcontainer container 7b5c40d4485c652db5fa04b64c356bfeee217a16f3a7b57be30fa2efb66423eb. May 27 03:16:14.850294 containerd[1545]: time="2025-05-27T03:16:14.850090566Z" level=info msg="StartContainer for \"7b5c40d4485c652db5fa04b64c356bfeee217a16f3a7b57be30fa2efb66423eb\" returns successfully" May 27 03:16:14.898463 kubelet[2688]: E0527 03:16:14.897995 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:14.925342 kubelet[2688]: I0527 03:16:14.925270 2688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-rqshl" podStartSLOduration=43.92525274 podStartE2EDuration="43.92525274s" podCreationTimestamp="2025-05-27 03:15:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-27 03:16:14.924453583 +0000 UTC m=+48.626065898" watchObservedRunningTime="2025-05-27 03:16:14.92525274 +0000 UTC m=+48.626865049" May 27 03:16:15.461235 systemd-networkd[1448]: calif16bf5174c7: Gained IPv6LL May 27 03:16:15.900942 kubelet[2688]: E0527 03:16:15.900421 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:16.904340 kubelet[2688]: E0527 03:16:16.903724 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:17.908296 kubelet[2688]: E0527 03:16:17.908242 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:18.166915 containerd[1545]: time="2025-05-27T03:16:18.166144498Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:18.167701 containerd[1545]: time="2025-05-27T03:16:18.167591471Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.0: active requests=0, bytes read=47252431" May 27 03:16:18.168428 containerd[1545]: time="2025-05-27T03:16:18.168156556Z" level=info msg="ImageCreate event name:\"sha256:5fa544b30bbe7e24458b21b80890f8834eebe8bcb99071f6caded1a39fc59082\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:18.171909 containerd[1545]: time="2025-05-27T03:16:18.171313030Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:ad7d2e76f15777636c5d91c108d7655659b38fe8970255050ffa51223eb96ff4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:18.175211 containerd[1545]: time="2025-05-27T03:16:18.175145866Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.0\" with image id \"sha256:5fa544b30bbe7e24458b21b80890f8834eebe8bcb99071f6caded1a39fc59082\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.0\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ad7d2e76f15777636c5d91c108d7655659b38fe8970255050ffa51223eb96ff4\", size \"48745150\" in 3.637425831s" May 27 03:16:18.175211 containerd[1545]: time="2025-05-27T03:16:18.175203054Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.0\" returns image reference \"sha256:5fa544b30bbe7e24458b21b80890f8834eebe8bcb99071f6caded1a39fc59082\"" May 27 03:16:18.177549 containerd[1545]: time="2025-05-27T03:16:18.177468828Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.0\"" May 27 03:16:18.184066 containerd[1545]: time="2025-05-27T03:16:18.183301282Z" level=info msg="CreateContainer within sandbox \"5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 27 03:16:18.196686 containerd[1545]: time="2025-05-27T03:16:18.196623916Z" level=info msg="Container 17bce58af8a0b7e6687ff27f76c069239910181bd72ef27af93856bf4d516914: CDI devices from CRI Config.CDIDevices: []" May 27 03:16:18.215669 containerd[1545]: time="2025-05-27T03:16:18.215599455Z" level=info msg="CreateContainer within sandbox \"5385f3db114d85bf0e9fb839ae5431480b6021a7ec0170d645af6c9d26c9e28a\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"17bce58af8a0b7e6687ff27f76c069239910181bd72ef27af93856bf4d516914\"" May 27 03:16:18.217998 containerd[1545]: time="2025-05-27T03:16:18.217884322Z" level=info msg="StartContainer for \"17bce58af8a0b7e6687ff27f76c069239910181bd72ef27af93856bf4d516914\"" May 27 03:16:18.222872 containerd[1545]: time="2025-05-27T03:16:18.222808425Z" level=info msg="connecting to shim 17bce58af8a0b7e6687ff27f76c069239910181bd72ef27af93856bf4d516914" address="unix:///run/containerd/s/55f54969f014ddeb35fec70179d47832b8d46dfdd94d04d2851cc455d3a0bcfe" protocol=ttrpc version=3 May 27 03:16:18.289551 systemd[1]: Started cri-containerd-17bce58af8a0b7e6687ff27f76c069239910181bd72ef27af93856bf4d516914.scope - libcontainer container 17bce58af8a0b7e6687ff27f76c069239910181bd72ef27af93856bf4d516914. May 27 03:16:18.410993 containerd[1545]: time="2025-05-27T03:16:18.410829749Z" level=info msg="StartContainer for \"17bce58af8a0b7e6687ff27f76c069239910181bd72ef27af93856bf4d516914\" returns successfully" May 27 03:16:18.597121 containerd[1545]: time="2025-05-27T03:16:18.593985668Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:18.604719 containerd[1545]: time="2025-05-27T03:16:18.604638666Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.0: active requests=0, bytes read=77" May 27 03:16:18.611739 containerd[1545]: time="2025-05-27T03:16:18.611105863Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.0\" with image id \"sha256:5fa544b30bbe7e24458b21b80890f8834eebe8bcb99071f6caded1a39fc59082\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.0\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ad7d2e76f15777636c5d91c108d7655659b38fe8970255050ffa51223eb96ff4\", size \"48745150\" in 433.591012ms" May 27 03:16:18.611739 containerd[1545]: time="2025-05-27T03:16:18.611171935Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.0\" returns image reference \"sha256:5fa544b30bbe7e24458b21b80890f8834eebe8bcb99071f6caded1a39fc59082\"" May 27 03:16:18.616717 containerd[1545]: time="2025-05-27T03:16:18.616670092Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.0\"" May 27 03:16:18.619355 containerd[1545]: time="2025-05-27T03:16:18.619192890Z" level=info msg="CreateContainer within sandbox \"5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 27 03:16:18.642621 containerd[1545]: time="2025-05-27T03:16:18.642507085Z" level=info msg="Container 363f621555f99d516a610990e7195824ae79eb4242db357d9d2fe0955665d8cb: CDI devices from CRI Config.CDIDevices: []" May 27 03:16:18.658899 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1019804393.mount: Deactivated successfully. May 27 03:16:18.696441 containerd[1545]: time="2025-05-27T03:16:18.696372098Z" level=info msg="CreateContainer within sandbox \"5a4233a9288141e67d944ee7b0a98a979041df4cb37e86879ea40485874bfdcd\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"363f621555f99d516a610990e7195824ae79eb4242db357d9d2fe0955665d8cb\"" May 27 03:16:18.701113 containerd[1545]: time="2025-05-27T03:16:18.697469507Z" level=info msg="StartContainer for \"363f621555f99d516a610990e7195824ae79eb4242db357d9d2fe0955665d8cb\"" May 27 03:16:18.701372 containerd[1545]: time="2025-05-27T03:16:18.699035028Z" level=info msg="connecting to shim 363f621555f99d516a610990e7195824ae79eb4242db357d9d2fe0955665d8cb" address="unix:///run/containerd/s/c3f02747d2b211ea79a2c9020afbec4c3f3a191a358c01464f16cecfbb9cdf0b" protocol=ttrpc version=3 May 27 03:16:18.803372 systemd[1]: Started cri-containerd-363f621555f99d516a610990e7195824ae79eb4242db357d9d2fe0955665d8cb.scope - libcontainer container 363f621555f99d516a610990e7195824ae79eb4242db357d9d2fe0955665d8cb. May 27 03:16:18.860183 containerd[1545]: time="2025-05-27T03:16:18.859625985Z" level=info msg="fetch failed" error="failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" host=ghcr.io May 27 03:16:18.863017 containerd[1545]: time="2025-05-27T03:16:18.862958146Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.0: active requests=0, bytes read=86" May 27 03:16:18.869089 containerd[1545]: time="2025-05-27T03:16:18.868264952Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.0\" failed" error="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" May 27 03:16:18.907963 kubelet[2688]: E0527 03:16:18.898616 2688 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/goldmane:v3.30.0" May 27 03:16:18.910813 kubelet[2688]: E0527 03:16:18.910496 2688 kuberuntime_image.go:55] "Failed to pull image" err="failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/goldmane:v3.30.0" May 27 03:16:18.913942 containerd[1545]: time="2025-05-27T03:16:18.913782128Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.0\"" May 27 03:16:18.920081 kubelet[2688]: E0527 03:16:18.919565 2688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.0,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t7gvq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-8f77d7b6c-zcnhc_calico-system(34798e8e-55b5-4e78-832b-798f5586248d): ErrImagePull: failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" logger="UnhandledError" May 27 03:16:18.939875 kubelet[2688]: E0527 03:16:18.939755 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.0\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.0\\\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden\"" pod="calico-system/goldmane-8f77d7b6c-zcnhc" podUID="34798e8e-55b5-4e78-832b-798f5586248d" May 27 03:16:18.986388 kubelet[2688]: I0527 03:16:18.986281 2688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-77c54fb4c6-tgcsm" podStartSLOduration=29.049065001 podStartE2EDuration="35.986236299s" podCreationTimestamp="2025-05-27 03:15:43 +0000 UTC" firstStartedPulling="2025-05-27 03:16:11.240025622 +0000 UTC m=+44.941637916" lastFinishedPulling="2025-05-27 03:16:18.177196902 +0000 UTC m=+51.878809214" observedRunningTime="2025-05-27 03:16:18.985197183 +0000 UTC m=+52.686809502" watchObservedRunningTime="2025-05-27 03:16:18.986236299 +0000 UTC m=+52.687848615" May 27 03:16:19.103444 containerd[1545]: time="2025-05-27T03:16:19.103381586Z" level=info msg="StartContainer for \"363f621555f99d516a610990e7195824ae79eb4242db357d9d2fe0955665d8cb\" returns successfully" May 27 03:16:19.952059 kubelet[2688]: I0527 03:16:19.951379 2688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 27 03:16:19.953496 kubelet[2688]: E0527 03:16:19.953450 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.0\\\"\"" pod="calico-system/goldmane-8f77d7b6c-zcnhc" podUID="34798e8e-55b5-4e78-832b-798f5586248d" May 27 03:16:20.061102 kubelet[2688]: I0527 03:16:20.060271 2688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-77c54fb4c6-j9z9b" podStartSLOduration=29.773612708 podStartE2EDuration="37.060245762s" podCreationTimestamp="2025-05-27 03:15:43 +0000 UTC" firstStartedPulling="2025-05-27 03:16:11.327102283 +0000 UTC m=+45.028714593" lastFinishedPulling="2025-05-27 03:16:18.61373533 +0000 UTC m=+52.315347647" observedRunningTime="2025-05-27 03:16:19.982475426 +0000 UTC m=+53.684087741" watchObservedRunningTime="2025-05-27 03:16:20.060245762 +0000 UTC m=+53.761858077" May 27 03:16:20.585108 containerd[1545]: time="2025-05-27T03:16:20.583192933Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:20.634257 containerd[1545]: time="2025-05-27T03:16:20.588131323Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.0: active requests=0, bytes read=14705639" May 27 03:16:20.634425 containerd[1545]: time="2025-05-27T03:16:20.592427708Z" level=info msg="ImageCreate event name:\"sha256:45c8692ffc029387ee93ba83da8ad26da9749cf2ba6ed03981f8f9933ed5a5b0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:20.634917 containerd[1545]: time="2025-05-27T03:16:20.613428527Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.0\" with image id \"sha256:45c8692ffc029387ee93ba83da8ad26da9749cf2ba6ed03981f8f9933ed5a5b0\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.0\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:dca5c16181edde2e860463615523ce457cd9dcfca85b7cfdcd6f3ea7de6f2ac8\", size \"16198294\" in 1.699592282s" May 27 03:16:20.634917 containerd[1545]: time="2025-05-27T03:16:20.634532998Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.0\" returns image reference \"sha256:45c8692ffc029387ee93ba83da8ad26da9749cf2ba6ed03981f8f9933ed5a5b0\"" May 27 03:16:20.638721 containerd[1545]: time="2025-05-27T03:16:20.638523656Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:dca5c16181edde2e860463615523ce457cd9dcfca85b7cfdcd6f3ea7de6f2ac8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:16:20.640563 containerd[1545]: time="2025-05-27T03:16:20.640443857Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.0\"" May 27 03:16:20.643158 containerd[1545]: time="2025-05-27T03:16:20.643100487Z" level=info msg="CreateContainer within sandbox \"fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" May 27 03:16:20.674573 containerd[1545]: time="2025-05-27T03:16:20.674507315Z" level=info msg="Container 9d28770a8840c92871a54f44cbecc85dd02b678c06206b6192430b50f5ebbd3c: CDI devices from CRI Config.CDIDevices: []" May 27 03:16:20.730476 containerd[1545]: time="2025-05-27T03:16:20.730349459Z" level=info msg="CreateContainer within sandbox \"fae5afd9a9a9843cd862e6e092ed31d201263bf3f930298340ab811a3ad8f521\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"9d28770a8840c92871a54f44cbecc85dd02b678c06206b6192430b50f5ebbd3c\"" May 27 03:16:20.731721 containerd[1545]: time="2025-05-27T03:16:20.731680190Z" level=info msg="StartContainer for \"9d28770a8840c92871a54f44cbecc85dd02b678c06206b6192430b50f5ebbd3c\"" May 27 03:16:20.733805 containerd[1545]: time="2025-05-27T03:16:20.733713094Z" level=info msg="connecting to shim 9d28770a8840c92871a54f44cbecc85dd02b678c06206b6192430b50f5ebbd3c" address="unix:///run/containerd/s/00284308e71b6df3068cacd6548b2c432015b0dd67248b07309b83ef25209286" protocol=ttrpc version=3 May 27 03:16:20.791768 systemd[1]: Started cri-containerd-9d28770a8840c92871a54f44cbecc85dd02b678c06206b6192430b50f5ebbd3c.scope - libcontainer container 9d28770a8840c92871a54f44cbecc85dd02b678c06206b6192430b50f5ebbd3c. May 27 03:16:20.869004 containerd[1545]: time="2025-05-27T03:16:20.868832248Z" level=info msg="fetch failed" error="failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" host=ghcr.io May 27 03:16:20.870096 containerd[1545]: time="2025-05-27T03:16:20.870010891Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.0\" failed" error="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" May 27 03:16:20.870518 containerd[1545]: time="2025-05-27T03:16:20.870288888Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.0: active requests=0, bytes read=86" May 27 03:16:20.871017 kubelet[2688]: E0527 03:16:20.870974 2688 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/whisker:v3.30.0" May 27 03:16:20.871319 kubelet[2688]: E0527 03:16:20.871283 2688 kuberuntime_image.go:55] "Failed to pull image" err="failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/whisker:v3.30.0" May 27 03:16:20.871647 kubelet[2688]: E0527 03:16:20.871612 2688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.0,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.0,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:961d12259f7740ee97ccbf2715e8204e,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bs6dk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-6464b44b6c-gfpc2_calico-system(148d61b5-bb87-4239-99d0-544a4ea406a1): ErrImagePull: failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" logger="UnhandledError" May 27 03:16:20.874508 containerd[1545]: time="2025-05-27T03:16:20.874478614Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\"" May 27 03:16:20.960166 kubelet[2688]: I0527 03:16:20.959485 2688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 27 03:16:21.126105 containerd[1545]: time="2025-05-27T03:16:21.125338951Z" level=info msg="StartContainer for \"9d28770a8840c92871a54f44cbecc85dd02b678c06206b6192430b50f5ebbd3c\" returns successfully" May 27 03:16:21.321577 containerd[1545]: time="2025-05-27T03:16:21.321320832Z" level=info msg="fetch failed" error="failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" host=ghcr.io May 27 03:16:21.322532 containerd[1545]: time="2025-05-27T03:16:21.322348320Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\" failed" error="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" May 27 03:16:21.322532 containerd[1545]: time="2025-05-27T03:16:21.322489259Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.0: active requests=0, bytes read=86" May 27 03:16:21.323255 kubelet[2688]: E0527 03:16:21.323017 2688 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.0" May 27 03:16:21.323483 kubelet[2688]: E0527 03:16:21.323415 2688 kuberuntime_image.go:55] "Failed to pull image" err="failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.0" May 27 03:16:21.324211 kubelet[2688]: E0527 03:16:21.324143 2688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.0,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bs6dk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-6464b44b6c-gfpc2_calico-system(148d61b5-bb87-4239-99d0-544a4ea406a1): ErrImagePull: failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" logger="UnhandledError" May 27 03:16:21.325667 kubelet[2688]: E0527 03:16:21.325576 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.0\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.0\\\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\\\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden\"]" pod="calico-system/whisker-6464b44b6c-gfpc2" podUID="148d61b5-bb87-4239-99d0-544a4ea406a1" May 27 03:16:21.865149 kubelet[2688]: I0527 03:16:21.865084 2688 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 May 27 03:16:21.866176 kubelet[2688]: I0527 03:16:21.865189 2688 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock May 27 03:16:24.053354 systemd[1]: Started sshd@7-209.38.65.15:22-139.178.68.195:47334.service - OpenSSH per-connection server daemon (139.178.68.195:47334). May 27 03:16:24.244002 sshd[5011]: Accepted publickey for core from 139.178.68.195 port 47334 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:16:24.249787 sshd-session[5011]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:16:24.260229 systemd-logind[1522]: New session 8 of user core. May 27 03:16:24.267719 systemd[1]: Started session-8.scope - Session 8 of User core. May 27 03:16:25.082266 sshd[5013]: Connection closed by 139.178.68.195 port 47334 May 27 03:16:25.083223 sshd-session[5011]: pam_unix(sshd:session): session closed for user core May 27 03:16:25.095777 systemd[1]: sshd@7-209.38.65.15:22-139.178.68.195:47334.service: Deactivated successfully. May 27 03:16:25.104102 systemd[1]: session-8.scope: Deactivated successfully. May 27 03:16:25.106739 systemd-logind[1522]: Session 8 logged out. Waiting for processes to exit. May 27 03:16:25.108963 systemd-logind[1522]: Removed session 8. May 27 03:16:27.668110 containerd[1545]: time="2025-05-27T03:16:27.667997248Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5aa9593f0cbaaca77974f8548443b3ecf0d1b89be2ecc5c8fa483c1687a2f605\" id:\"fa05a7e794bdcb13d08a7724cf1d8994de23b68ed4610809708a27df8e5f85d3\" pid:5042 exited_at:{seconds:1748315787 nanos:591847406}" May 27 03:16:30.103367 systemd[1]: Started sshd@8-209.38.65.15:22-139.178.68.195:47338.service - OpenSSH per-connection server daemon (139.178.68.195:47338). May 27 03:16:30.198770 sshd[5058]: Accepted publickey for core from 139.178.68.195 port 47338 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:16:30.200601 sshd-session[5058]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:16:30.211896 systemd-logind[1522]: New session 9 of user core. May 27 03:16:30.220326 systemd[1]: Started session-9.scope - Session 9 of User core. May 27 03:16:30.480208 sshd[5060]: Connection closed by 139.178.68.195 port 47338 May 27 03:16:30.481419 sshd-session[5058]: pam_unix(sshd:session): session closed for user core May 27 03:16:30.488440 systemd[1]: sshd@8-209.38.65.15:22-139.178.68.195:47338.service: Deactivated successfully. May 27 03:16:30.491826 systemd[1]: session-9.scope: Deactivated successfully. May 27 03:16:30.495419 systemd-logind[1522]: Session 9 logged out. Waiting for processes to exit. May 27 03:16:30.498102 systemd-logind[1522]: Removed session 9. May 27 03:16:30.868749 kubelet[2688]: I0527 03:16:30.868351 2688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 27 03:16:30.935092 kubelet[2688]: I0527 03:16:30.934887 2688 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-drsns" podStartSLOduration=33.280023407 podStartE2EDuration="43.93486228s" podCreationTimestamp="2025-05-27 03:15:47 +0000 UTC" firstStartedPulling="2025-05-27 03:16:09.984307368 +0000 UTC m=+43.685919662" lastFinishedPulling="2025-05-27 03:16:20.639146241 +0000 UTC m=+54.340758535" observedRunningTime="2025-05-27 03:16:22.002932347 +0000 UTC m=+55.704544662" watchObservedRunningTime="2025-05-27 03:16:30.93486228 +0000 UTC m=+64.636474597" May 27 03:16:34.507738 kubelet[2688]: E0527 03:16:34.507602 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.0\\\"\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\\\"\"]" pod="calico-system/whisker-6464b44b6c-gfpc2" podUID="148d61b5-bb87-4239-99d0-544a4ea406a1" May 27 03:16:34.511890 containerd[1545]: time="2025-05-27T03:16:34.508204630Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.0\"" May 27 03:16:34.724520 containerd[1545]: time="2025-05-27T03:16:34.724448177Z" level=info msg="fetch failed" error="failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" host=ghcr.io May 27 03:16:34.737917 containerd[1545]: time="2025-05-27T03:16:34.737860441Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.0\" failed" error="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" May 27 03:16:34.738886 containerd[1545]: time="2025-05-27T03:16:34.738155150Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.0: active requests=0, bytes read=86" May 27 03:16:34.739513 kubelet[2688]: E0527 03:16:34.739451 2688 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/goldmane:v3.30.0" May 27 03:16:34.739643 kubelet[2688]: E0527 03:16:34.739546 2688 kuberuntime_image.go:55] "Failed to pull image" err="failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/goldmane:v3.30.0" May 27 03:16:34.740079 kubelet[2688]: E0527 03:16:34.739728 2688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.0,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t7gvq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-8f77d7b6c-zcnhc_calico-system(34798e8e-55b5-4e78-832b-798f5586248d): ErrImagePull: failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" logger="UnhandledError" May 27 03:16:34.741559 kubelet[2688]: E0527 03:16:34.741469 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.0\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.0\\\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden\"" pod="calico-system/goldmane-8f77d7b6c-zcnhc" podUID="34798e8e-55b5-4e78-832b-798f5586248d" May 27 03:16:35.500413 systemd[1]: Started sshd@9-209.38.65.15:22-139.178.68.195:60928.service - OpenSSH per-connection server daemon (139.178.68.195:60928). May 27 03:16:35.721867 sshd[5080]: Accepted publickey for core from 139.178.68.195 port 60928 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:16:35.727519 sshd-session[5080]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:16:35.742399 systemd-logind[1522]: New session 10 of user core. May 27 03:16:35.752654 systemd[1]: Started session-10.scope - Session 10 of User core. May 27 03:16:36.037697 sshd[5082]: Connection closed by 139.178.68.195 port 60928 May 27 03:16:36.038315 sshd-session[5080]: pam_unix(sshd:session): session closed for user core May 27 03:16:36.051613 systemd[1]: sshd@9-209.38.65.15:22-139.178.68.195:60928.service: Deactivated successfully. May 27 03:16:36.056832 systemd[1]: session-10.scope: Deactivated successfully. May 27 03:16:36.058398 systemd-logind[1522]: Session 10 logged out. Waiting for processes to exit. May 27 03:16:36.065179 systemd[1]: Started sshd@10-209.38.65.15:22-139.178.68.195:60930.service - OpenSSH per-connection server daemon (139.178.68.195:60930). May 27 03:16:36.067557 systemd-logind[1522]: Removed session 10. May 27 03:16:36.148110 sshd[5095]: Accepted publickey for core from 139.178.68.195 port 60930 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:16:36.151933 sshd-session[5095]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:16:36.164004 systemd-logind[1522]: New session 11 of user core. May 27 03:16:36.167562 systemd[1]: Started session-11.scope - Session 11 of User core. May 27 03:16:36.530681 sshd[5097]: Connection closed by 139.178.68.195 port 60930 May 27 03:16:36.535162 sshd-session[5095]: pam_unix(sshd:session): session closed for user core May 27 03:16:36.554427 systemd[1]: sshd@10-209.38.65.15:22-139.178.68.195:60930.service: Deactivated successfully. May 27 03:16:36.562396 systemd[1]: session-11.scope: Deactivated successfully. May 27 03:16:36.566939 systemd-logind[1522]: Session 11 logged out. Waiting for processes to exit. May 27 03:16:36.575205 systemd[1]: Started sshd@11-209.38.65.15:22-139.178.68.195:60946.service - OpenSSH per-connection server daemon (139.178.68.195:60946). May 27 03:16:36.577558 systemd-logind[1522]: Removed session 11. May 27 03:16:36.684163 sshd[5107]: Accepted publickey for core from 139.178.68.195 port 60946 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:16:36.689849 sshd-session[5107]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:16:36.714909 systemd-logind[1522]: New session 12 of user core. May 27 03:16:36.722401 systemd[1]: Started session-12.scope - Session 12 of User core. May 27 03:16:36.952297 sshd[5109]: Connection closed by 139.178.68.195 port 60946 May 27 03:16:36.952854 sshd-session[5107]: pam_unix(sshd:session): session closed for user core May 27 03:16:36.963862 systemd[1]: sshd@11-209.38.65.15:22-139.178.68.195:60946.service: Deactivated successfully. May 27 03:16:36.971549 systemd[1]: session-12.scope: Deactivated successfully. May 27 03:16:36.974518 systemd-logind[1522]: Session 12 logged out. Waiting for processes to exit. May 27 03:16:36.979302 systemd-logind[1522]: Removed session 12. May 27 03:16:38.503422 kubelet[2688]: E0527 03:16:38.503010 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:39.961985 containerd[1545]: time="2025-05-27T03:16:39.961898001Z" level=info msg="TaskExit event in podsandbox handler container_id:\"320d5852593cb9fdb7fcd6a5a8f42623b3c597874b90e987eaa194b5384f5175\" id:\"7b380f744c4e4f7cb956193a6bef9db6e4d1b07f1b6dad035182bea9ef82fe60\" pid:5133 exit_status:1 exited_at:{seconds:1748315799 nanos:958897445}" May 27 03:16:41.970265 systemd[1]: Started sshd@12-209.38.65.15:22-139.178.68.195:60962.service - OpenSSH per-connection server daemon (139.178.68.195:60962). May 27 03:16:42.079657 sshd[5149]: Accepted publickey for core from 139.178.68.195 port 60962 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:16:42.082290 sshd-session[5149]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:16:42.093610 systemd-logind[1522]: New session 13 of user core. May 27 03:16:42.100329 systemd[1]: Started session-13.scope - Session 13 of User core. May 27 03:16:42.311612 sshd[5151]: Connection closed by 139.178.68.195 port 60962 May 27 03:16:42.313614 sshd-session[5149]: pam_unix(sshd:session): session closed for user core May 27 03:16:42.320390 systemd-logind[1522]: Session 13 logged out. Waiting for processes to exit. May 27 03:16:42.321363 systemd[1]: sshd@12-209.38.65.15:22-139.178.68.195:60962.service: Deactivated successfully. May 27 03:16:42.327797 systemd[1]: session-13.scope: Deactivated successfully. May 27 03:16:42.333200 systemd-logind[1522]: Removed session 13. May 27 03:16:44.504339 kubelet[2688]: E0527 03:16:44.503322 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:45.506604 containerd[1545]: time="2025-05-27T03:16:45.506549702Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.0\"" May 27 03:16:45.741469 containerd[1545]: time="2025-05-27T03:16:45.741408353Z" level=info msg="fetch failed" error="failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" host=ghcr.io May 27 03:16:45.742307 containerd[1545]: time="2025-05-27T03:16:45.742209257Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.0\" failed" error="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" May 27 03:16:45.742451 containerd[1545]: time="2025-05-27T03:16:45.742301577Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.0: active requests=0, bytes read=86" May 27 03:16:45.742801 kubelet[2688]: E0527 03:16:45.742744 2688 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/whisker:v3.30.0" May 27 03:16:45.743390 kubelet[2688]: E0527 03:16:45.742812 2688 kuberuntime_image.go:55] "Failed to pull image" err="failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/whisker:v3.30.0" May 27 03:16:45.743390 kubelet[2688]: E0527 03:16:45.743080 2688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:whisker,Image:ghcr.io/flatcar/calico/whisker:v3.30.0,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:CALICO_VERSION,Value:v3.30.0,ValueFrom:nil,},EnvVar{Name:CLUSTER_ID,Value:961d12259f7740ee97ccbf2715e8204e,ValueFrom:nil,},EnvVar{Name:CLUSTER_TYPE,Value:typha,kdd,k8s,operator,bgp,kubeadm,ValueFrom:nil,},EnvVar{Name:NOTIFICATIONS,Value:Enabled,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bs6dk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-6464b44b6c-gfpc2_calico-system(148d61b5-bb87-4239-99d0-544a4ea406a1): ErrImagePull: failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden" logger="UnhandledError" May 27 03:16:45.745867 containerd[1545]: time="2025-05-27T03:16:45.745817541Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\"" May 27 03:16:45.966066 containerd[1545]: time="2025-05-27T03:16:45.965965290Z" level=info msg="fetch failed" error="failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" host=ghcr.io May 27 03:16:45.966949 containerd[1545]: time="2025-05-27T03:16:45.966876814Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\" failed" error="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" May 27 03:16:45.967139 containerd[1545]: time="2025-05-27T03:16:45.967022604Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.0: active requests=0, bytes read=86" May 27 03:16:45.967769 kubelet[2688]: E0527 03:16:45.967694 2688 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.0" May 27 03:16:45.967897 kubelet[2688]: E0527 03:16:45.967779 2688 kuberuntime_image.go:55] "Failed to pull image" err="failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/whisker-backend:v3.30.0" May 27 03:16:45.967997 kubelet[2688]: E0527 03:16:45.967943 2688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:whisker-backend,Image:ghcr.io/flatcar/calico/whisker-backend:v3.30.0,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:3002,ValueFrom:nil,},EnvVar{Name:GOLDMANE_HOST,Value:goldmane.calico-system.svc.cluster.local:7443,ValueFrom:nil,},EnvVar{Name:TLS_CERT_PATH,Value:/whisker-backend-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:TLS_KEY_PATH,Value:/whisker-backend-key-pair/tls.key,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:whisker-backend-key-pair,ReadOnly:true,MountPath:/whisker-backend-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:whisker-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bs6dk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod whisker-6464b44b6c-gfpc2_calico-system(148d61b5-bb87-4239-99d0-544a4ea406a1): ErrImagePull: failed to pull and unpack image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden" logger="UnhandledError" May 27 03:16:45.969343 kubelet[2688]: E0527 03:16:45.969255 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ErrImagePull: \"failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker:v3.30.0\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker:v3.30.0\\\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker%3Apull&service=ghcr.io: 403 Forbidden\", failed to \"StartContainer\" for \"whisker-backend\" with ErrImagePull: \"failed to pull and unpack image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\\\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fwhisker-backend%3Apull&service=ghcr.io: 403 Forbidden\"]" pod="calico-system/whisker-6464b44b6c-gfpc2" podUID="148d61b5-bb87-4239-99d0-544a4ea406a1" May 27 03:16:46.246428 kubelet[2688]: I0527 03:16:46.246299 2688 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 27 03:16:46.504190 kubelet[2688]: E0527 03:16:46.503283 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:46.506448 kubelet[2688]: E0527 03:16:46.506401 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.0\\\"\"" pod="calico-system/goldmane-8f77d7b6c-zcnhc" podUID="34798e8e-55b5-4e78-832b-798f5586248d" May 27 03:16:47.330278 systemd[1]: Started sshd@13-209.38.65.15:22-139.178.68.195:42634.service - OpenSSH per-connection server daemon (139.178.68.195:42634). May 27 03:16:47.462326 sshd[5164]: Accepted publickey for core from 139.178.68.195 port 42634 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:16:47.467098 sshd-session[5164]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:16:47.477776 systemd-logind[1522]: New session 14 of user core. May 27 03:16:47.484463 systemd[1]: Started session-14.scope - Session 14 of User core. May 27 03:16:47.748740 sshd[5166]: Connection closed by 139.178.68.195 port 42634 May 27 03:16:47.750935 sshd-session[5164]: pam_unix(sshd:session): session closed for user core May 27 03:16:47.757034 systemd[1]: sshd@13-209.38.65.15:22-139.178.68.195:42634.service: Deactivated successfully. May 27 03:16:47.757127 systemd-logind[1522]: Session 14 logged out. Waiting for processes to exit. May 27 03:16:47.763152 systemd[1]: session-14.scope: Deactivated successfully. May 27 03:16:47.772120 systemd-logind[1522]: Removed session 14. May 27 03:16:52.763535 systemd[1]: Started sshd@14-209.38.65.15:22-139.178.68.195:42650.service - OpenSSH per-connection server daemon (139.178.68.195:42650). May 27 03:16:52.835672 sshd[5184]: Accepted publickey for core from 139.178.68.195 port 42650 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:16:52.838546 sshd-session[5184]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:16:52.850500 systemd-logind[1522]: New session 15 of user core. May 27 03:16:52.855437 systemd[1]: Started session-15.scope - Session 15 of User core. May 27 03:16:53.134916 sshd[5186]: Connection closed by 139.178.68.195 port 42650 May 27 03:16:53.135979 sshd-session[5184]: pam_unix(sshd:session): session closed for user core May 27 03:16:53.147411 systemd[1]: sshd@14-209.38.65.15:22-139.178.68.195:42650.service: Deactivated successfully. May 27 03:16:53.148139 systemd-logind[1522]: Session 15 logged out. Waiting for processes to exit. May 27 03:16:53.152693 systemd[1]: session-15.scope: Deactivated successfully. May 27 03:16:53.157733 systemd-logind[1522]: Removed session 15. May 27 03:16:55.502897 kubelet[2688]: E0527 03:16:55.502834 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:16:57.599326 containerd[1545]: time="2025-05-27T03:16:57.599255464Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5aa9593f0cbaaca77974f8548443b3ecf0d1b89be2ecc5c8fa483c1687a2f605\" id:\"c9d20599d345b8064a2d734d18fc52ce014230003df564b4101c8e0a001b158c\" pid:5210 exited_at:{seconds:1748315817 nanos:598491616}" May 27 03:16:58.154569 systemd[1]: Started sshd@15-209.38.65.15:22-139.178.68.195:46176.service - OpenSSH per-connection server daemon (139.178.68.195:46176). May 27 03:16:58.329361 sshd[5221]: Accepted publickey for core from 139.178.68.195 port 46176 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:16:58.333738 sshd-session[5221]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:16:58.362120 systemd-logind[1522]: New session 16 of user core. May 27 03:16:58.373907 systemd[1]: Started session-16.scope - Session 16 of User core. May 27 03:16:58.627083 sshd[5223]: Connection closed by 139.178.68.195 port 46176 May 27 03:16:58.628073 sshd-session[5221]: pam_unix(sshd:session): session closed for user core May 27 03:16:58.638985 systemd[1]: sshd@15-209.38.65.15:22-139.178.68.195:46176.service: Deactivated successfully. May 27 03:16:58.645412 systemd[1]: session-16.scope: Deactivated successfully. May 27 03:16:58.649188 systemd-logind[1522]: Session 16 logged out. Waiting for processes to exit. May 27 03:16:58.654428 systemd[1]: Started sshd@16-209.38.65.15:22-139.178.68.195:46178.service - OpenSSH per-connection server daemon (139.178.68.195:46178). May 27 03:16:58.659420 systemd-logind[1522]: Removed session 16. May 27 03:16:58.741506 sshd[5235]: Accepted publickey for core from 139.178.68.195 port 46178 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:16:58.743434 sshd-session[5235]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:16:58.754667 systemd-logind[1522]: New session 17 of user core. May 27 03:16:58.758344 systemd[1]: Started session-17.scope - Session 17 of User core. May 27 03:16:59.192437 sshd[5237]: Connection closed by 139.178.68.195 port 46178 May 27 03:16:59.195014 sshd-session[5235]: pam_unix(sshd:session): session closed for user core May 27 03:16:59.209878 systemd[1]: sshd@16-209.38.65.15:22-139.178.68.195:46178.service: Deactivated successfully. May 27 03:16:59.216595 systemd[1]: session-17.scope: Deactivated successfully. May 27 03:16:59.233112 systemd-logind[1522]: Session 17 logged out. Waiting for processes to exit. May 27 03:16:59.239774 systemd[1]: Started sshd@17-209.38.65.15:22-139.178.68.195:46188.service - OpenSSH per-connection server daemon (139.178.68.195:46188). May 27 03:16:59.242673 systemd-logind[1522]: Removed session 17. May 27 03:16:59.448243 sshd[5247]: Accepted publickey for core from 139.178.68.195 port 46188 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:16:59.453257 sshd-session[5247]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:16:59.466137 systemd-logind[1522]: New session 18 of user core. May 27 03:16:59.471445 systemd[1]: Started session-18.scope - Session 18 of User core. May 27 03:16:59.509685 kubelet[2688]: E0527 03:16:59.509438 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.0\\\"\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\\\"\"]" pod="calico-system/whisker-6464b44b6c-gfpc2" podUID="148d61b5-bb87-4239-99d0-544a4ea406a1" May 27 03:17:00.503693 kubelet[2688]: E0527 03:17:00.503640 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:17:01.506124 containerd[1545]: time="2025-05-27T03:17:01.505435071Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.0\"" May 27 03:17:01.729359 containerd[1545]: time="2025-05-27T03:17:01.729294758Z" level=info msg="fetch failed" error="failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" host=ghcr.io May 27 03:17:01.732999 containerd[1545]: time="2025-05-27T03:17:01.732772625Z" level=error msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.0\" failed" error="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" May 27 03:17:01.732999 containerd[1545]: time="2025-05-27T03:17:01.732943742Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.0: active requests=0, bytes read=86" May 27 03:17:01.736088 kubelet[2688]: E0527 03:17:01.735732 2688 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/goldmane:v3.30.0" May 27 03:17:01.736088 kubelet[2688]: E0527 03:17:01.735820 2688 kuberuntime_image.go:55] "Failed to pull image" err="failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" image="ghcr.io/flatcar/calico/goldmane:v3.30.0" May 27 03:17:01.736088 kubelet[2688]: E0527 03:17:01.736005 2688 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:goldmane,Image:ghcr.io/flatcar/calico/goldmane:v3.30.0,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOG_LEVEL,Value:INFO,ValueFrom:nil,},EnvVar{Name:PORT,Value:7443,ValueFrom:nil,},EnvVar{Name:SERVER_CERT_PATH,Value:/goldmane-key-pair/tls.crt,ValueFrom:nil,},EnvVar{Name:SERVER_KEY_PATH,Value:/goldmane-key-pair/tls.key,ValueFrom:nil,},EnvVar{Name:CA_CERT_PATH,Value:/etc/pki/tls/certs/tigera-ca-bundle.crt,ValueFrom:nil,},EnvVar{Name:PUSH_URL,Value:https://guardian.calico-system.svc.cluster.local:443/api/v1/flows/bulk,ValueFrom:nil,},EnvVar{Name:FILE_CONFIG_PATH,Value:/config/config.json,ValueFrom:nil,},EnvVar{Name:HEALTH_ENABLED,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-ca-bundle,ReadOnly:true,MountPath:/etc/pki/tls/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:goldmane-key-pair,ReadOnly:true,MountPath:/goldmane-key-pair,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t7gvq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -live],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:60,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/health -ready],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*10001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*10001,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod goldmane-8f77d7b6c-zcnhc_calico-system(34798e8e-55b5-4e78-832b-798f5586248d): ErrImagePull: failed to pull and unpack image \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to resolve reference \"ghcr.io/flatcar/calico/goldmane:v3.30.0\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden" logger="UnhandledError" May 27 03:17:01.738547 kubelet[2688]: E0527 03:17:01.737983 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ErrImagePull: \"failed to pull and unpack image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.0\\\": failed to resolve reference \\\"ghcr.io/flatcar/calico/goldmane:v3.30.0\\\": failed to authorize: failed to fetch anonymous token: unexpected status from GET request to https://ghcr.io/token?scope=repository%3Aflatcar%2Fcalico%2Fgoldmane%3Apull&service=ghcr.io: 403 Forbidden\"" pod="calico-system/goldmane-8f77d7b6c-zcnhc" podUID="34798e8e-55b5-4e78-832b-798f5586248d" May 27 03:17:03.670271 sshd[5249]: Connection closed by 139.178.68.195 port 46188 May 27 03:17:03.684749 sshd-session[5247]: pam_unix(sshd:session): session closed for user core May 27 03:17:03.750757 systemd[1]: sshd@17-209.38.65.15:22-139.178.68.195:46188.service: Deactivated successfully. May 27 03:17:03.758169 systemd[1]: session-18.scope: Deactivated successfully. May 27 03:17:03.758768 systemd[1]: session-18.scope: Consumed 932ms CPU time, 76.8M memory peak. May 27 03:17:03.762514 systemd-logind[1522]: Session 18 logged out. Waiting for processes to exit. May 27 03:17:03.774187 systemd[1]: Started sshd@18-209.38.65.15:22-139.178.68.195:48722.service - OpenSSH per-connection server daemon (139.178.68.195:48722). May 27 03:17:03.784742 systemd-logind[1522]: Removed session 18. May 27 03:17:03.951700 sshd[5269]: Accepted publickey for core from 139.178.68.195 port 48722 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:17:03.959626 sshd-session[5269]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:17:03.993547 systemd-logind[1522]: New session 19 of user core. May 27 03:17:03.995846 systemd[1]: Started session-19.scope - Session 19 of User core. May 27 03:17:05.498109 sshd[5271]: Connection closed by 139.178.68.195 port 48722 May 27 03:17:05.520477 sshd-session[5269]: pam_unix(sshd:session): session closed for user core May 27 03:17:05.543239 systemd[1]: Started sshd@19-209.38.65.15:22-139.178.68.195:48724.service - OpenSSH per-connection server daemon (139.178.68.195:48724). May 27 03:17:05.544041 systemd[1]: sshd@18-209.38.65.15:22-139.178.68.195:48722.service: Deactivated successfully. May 27 03:17:05.562433 systemd[1]: session-19.scope: Deactivated successfully. May 27 03:17:05.574478 systemd-logind[1522]: Session 19 logged out. Waiting for processes to exit. May 27 03:17:05.581316 systemd-logind[1522]: Removed session 19. May 27 03:17:05.681490 sshd[5278]: Accepted publickey for core from 139.178.68.195 port 48724 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:17:05.684513 sshd-session[5278]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:17:05.696133 systemd-logind[1522]: New session 20 of user core. May 27 03:17:05.702353 systemd[1]: Started session-20.scope - Session 20 of User core. May 27 03:17:06.047132 sshd[5283]: Connection closed by 139.178.68.195 port 48724 May 27 03:17:06.047988 sshd-session[5278]: pam_unix(sshd:session): session closed for user core May 27 03:17:06.058844 systemd[1]: sshd@19-209.38.65.15:22-139.178.68.195:48724.service: Deactivated successfully. May 27 03:17:06.063354 systemd[1]: session-20.scope: Deactivated successfully. May 27 03:17:06.067748 systemd-logind[1522]: Session 20 logged out. Waiting for processes to exit. May 27 03:17:06.069932 systemd-logind[1522]: Removed session 20. May 27 03:17:10.101407 containerd[1545]: time="2025-05-27T03:17:10.101017915Z" level=info msg="TaskExit event in podsandbox handler container_id:\"320d5852593cb9fdb7fcd6a5a8f42623b3c597874b90e987eaa194b5384f5175\" id:\"0ff4cb6fbd654871692ef2ff7ca703735e08c1cb5832c7e2d9598db738901b16\" pid:5310 exited_at:{seconds:1748315830 nanos:100283104}" May 27 03:17:11.063914 systemd[1]: Started sshd@20-209.38.65.15:22-139.178.68.195:48740.service - OpenSSH per-connection server daemon (139.178.68.195:48740). May 27 03:17:11.199952 sshd[5322]: Accepted publickey for core from 139.178.68.195 port 48740 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:17:11.202323 sshd-session[5322]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:17:11.212681 systemd-logind[1522]: New session 21 of user core. May 27 03:17:11.222215 systemd[1]: Started session-21.scope - Session 21 of User core. May 27 03:17:11.441472 sshd[5324]: Connection closed by 139.178.68.195 port 48740 May 27 03:17:11.442487 sshd-session[5322]: pam_unix(sshd:session): session closed for user core May 27 03:17:11.447955 systemd[1]: sshd@20-209.38.65.15:22-139.178.68.195:48740.service: Deactivated successfully. May 27 03:17:11.451666 systemd[1]: session-21.scope: Deactivated successfully. May 27 03:17:11.455653 systemd-logind[1522]: Session 21 logged out. Waiting for processes to exit. May 27 03:17:11.458421 systemd-logind[1522]: Removed session 21. May 27 03:17:11.522350 kubelet[2688]: E0527 03:17:11.520531 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.0\\\"\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\\\"\"]" pod="calico-system/whisker-6464b44b6c-gfpc2" podUID="148d61b5-bb87-4239-99d0-544a4ea406a1" May 27 03:17:12.802807 containerd[1545]: time="2025-05-27T03:17:12.802723208Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5aa9593f0cbaaca77974f8548443b3ecf0d1b89be2ecc5c8fa483c1687a2f605\" id:\"7bb5783e0291801029b0956249f0b12b7aa134ac474162bf2242942c8878aaf2\" pid:5347 exited_at:{seconds:1748315832 nanos:796393230}" May 27 03:17:14.502981 kubelet[2688]: E0527 03:17:14.502930 2688 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 27 03:17:16.459101 systemd[1]: Started sshd@21-209.38.65.15:22-139.178.68.195:54596.service - OpenSSH per-connection server daemon (139.178.68.195:54596). May 27 03:17:16.509095 kubelet[2688]: E0527 03:17:16.507976 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"goldmane\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/goldmane:v3.30.0\\\"\"" pod="calico-system/goldmane-8f77d7b6c-zcnhc" podUID="34798e8e-55b5-4e78-832b-798f5586248d" May 27 03:17:16.579613 sshd[5358]: Accepted publickey for core from 139.178.68.195 port 54596 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:17:16.581970 sshd-session[5358]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:17:16.595871 systemd-logind[1522]: New session 22 of user core. May 27 03:17:16.600213 systemd[1]: Started session-22.scope - Session 22 of User core. May 27 03:17:16.795538 sshd[5360]: Connection closed by 139.178.68.195 port 54596 May 27 03:17:16.796999 sshd-session[5358]: pam_unix(sshd:session): session closed for user core May 27 03:17:16.806808 systemd[1]: sshd@21-209.38.65.15:22-139.178.68.195:54596.service: Deactivated successfully. May 27 03:17:16.816601 systemd[1]: session-22.scope: Deactivated successfully. May 27 03:17:16.821752 systemd-logind[1522]: Session 22 logged out. Waiting for processes to exit. May 27 03:17:16.826947 systemd-logind[1522]: Removed session 22. May 27 03:17:21.815734 systemd[1]: Started sshd@22-209.38.65.15:22-139.178.68.195:54608.service - OpenSSH per-connection server daemon (139.178.68.195:54608). May 27 03:17:21.893664 sshd[5372]: Accepted publickey for core from 139.178.68.195 port 54608 ssh2: RSA SHA256:OpqNG0xcwHOVaq8cQkq2R2eIMCkLtsAOIXX7FdpR448 May 27 03:17:21.896517 sshd-session[5372]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:17:21.905513 systemd-logind[1522]: New session 23 of user core. May 27 03:17:21.913407 systemd[1]: Started session-23.scope - Session 23 of User core. May 27 03:17:22.134188 sshd[5374]: Connection closed by 139.178.68.195 port 54608 May 27 03:17:22.135344 sshd-session[5372]: pam_unix(sshd:session): session closed for user core May 27 03:17:22.144186 systemd[1]: sshd@22-209.38.65.15:22-139.178.68.195:54608.service: Deactivated successfully. May 27 03:17:22.149874 systemd[1]: session-23.scope: Deactivated successfully. May 27 03:17:22.152195 systemd-logind[1522]: Session 23 logged out. Waiting for processes to exit. May 27 03:17:22.155840 systemd-logind[1522]: Removed session 23. May 27 03:17:24.507701 kubelet[2688]: E0527 03:17:24.507208 2688 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"whisker\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker:v3.30.0\\\"\", failed to \"StartContainer\" for \"whisker-backend\" with ImagePullBackOff: \"Back-off pulling image \\\"ghcr.io/flatcar/calico/whisker-backend:v3.30.0\\\"\"]" pod="calico-system/whisker-6464b44b6c-gfpc2" podUID="148d61b5-bb87-4239-99d0-544a4ea406a1"